From mdulko at redhat.com Wed Dec 1 08:55:52 2021 From: mdulko at redhat.com (=?UTF-8?Q?Micha=C5=82?= Dulko) Date: Wed, 01 Dec 2021 09:55:52 +0100 Subject: [kuryr] Propose to EOL stable branches (Stein, Rocky and Queens) for all the kuryr scope In-Reply-To: References: Message-ID: <3c2cbd0af92ff7dffb3fae4ce6c30015abd94b6b.camel@redhat.com> Yes, let's remove them. It has little sense to keep these versions and we're actually actively working to keep the compatibility with older OpenStack versions anyway. On Fri, 2021-11-26 at 23:50 +0100, Maysa De Macedo Souza wrote: > Hello again, > > We also have Pike as a stable branch[0], so that is the older one, > which would also be EOL. > > [0] > https://review.opendev.org/q/(project:%255Eopenstack/kuryr.*)AND((branch:stable/queens)OR(branch:stable/rocky)OR(branch:stable/stein)OR(branch:stable/pike)) > > Cheers, > Maysa. > > On Fri, Nov 26, 2021 at 11:42 PM Maysa De Macedo Souza > wrote: > > Hello, > > > > We have plenty of stable branches in Kuryr repositories still open > > with the oldest being stable/queens. > > We haven't seen many backports[0] proposed to Stein, Rocky and > > Queens, so I would like to propose retiring these branches. > > > > Let me know if anybody is interested in keeping any of those > > branches. > > > > [0] > > https://review.opendev.org/q/(project:%255Eopenstack/kuryr.*)AND((branch:stable/queens)OR(branch:stable/rocky)OR(branch:stable/stein)) > > > > Thank you, > > Maysa Macedo. From Iain.Stott at thehutgroup.com Wed Dec 1 11:17:03 2021 From: Iain.Stott at thehutgroup.com (Iain Stott) Date: Wed, 1 Dec 2021 11:17:03 +0000 Subject: [ceilometer][octavia][Victoria] No metrics from octavia loadbalancer Message-ID: Has anyone had any progress on this? We are currently in the same situation, we can see if loadblancers exists, but unable to see a way to configure the polling agent to poll endpoints like v2/lbaas/loadbalancers/[ID]/stats Cheers, Iain > OK. I have some progress. I created meters.d and pollster.d (in pollsters.d I?ve created octavia.yaml with sample type gauge and unit load balancer) and now I can see some measures, but only if load balancer exists or not. > Is there any way to force dynamic pollster to ask for url v2/lbaas/loadbalancers/[ID]/stats? > Best regards > Adam Iain Stott OpenStack Engineer The Hut Group Tel: Email: Iain.Stott at thehutgroup.com For the purposes of this email, the "company" means The Hut Group Limited, a company registered in England and Wales (company number 6539496) whose registered office is at Fifth Floor, Voyager House, Chicago Avenue, Manchester Airport, M90 3DQ and/or any of its respective subsidiaries. Confidentiality Notice This e-mail is confidential and intended for the use of the named recipient only. If you are not the intended recipient please notify us by telephone immediately on +44(0)1606 811888 or return it to us by e-mail. Please then delete it from your system and note that any use, dissemination, forwarding, printing or copying is strictly prohibited. Any views or opinions are solely those of the author and do not necessarily represent those of the company. Encryptions and Viruses Please note that this e-mail and any attachments have not been encrypted. They may therefore be liable to be compromised. Please also note that it is your responsibility to scan this e-mail and any attachments for viruses. We do not, to the extent permitted by law, accept any liability (whether in contract, negligence or otherwise) for any virus infection and/or external compromise of security and/or confidentiality in relation to transmissions sent by e-mail. Monitoring Activity and use of the company's systems is monitored to secure its effective use and operation and for other lawful business purposes. Communications using these systems will also be monitored and may be recorded to secure effective use and operation and for other lawful business purposes. hgvyjuv -------------- next part -------------- An HTML attachment was scrubbed... URL: From rosmaita.fossdev at gmail.com Wed Dec 1 13:08:59 2021 From: rosmaita.fossdev at gmail.com (Brian Rosmaita) Date: Wed, 1 Dec 2021 08:08:59 -0500 Subject: [cinder] reminder: yoga R-17 midcycle today at 1400 utc Message-ID: he Cinder Yoga R-17 virtual midcycle will be held: DATE: Today, Wednesday 1 December 2021 TIME: 1400-1600 UTC LOCATION: https://bluejeans.com/3228528973 The meeting will be recorded. The midcycle etherpad is here: https://etherpad.opendev.org/p/cinder-yoga-midcycles cheers, brian From geguileo at redhat.com Wed Dec 1 13:58:46 2021 From: geguileo at redhat.com (Gorka Eguileor) Date: Wed, 1 Dec 2021 14:58:46 +0100 Subject: [cinder] : SAN migration In-Reply-To: References: Message-ID: <20211201135846.7vdnzse6vrejstql@localhost> On 03/11, Gauurav Sabharwal1 wrote: > > > Hi Experts , > > I need some expert advise of one of the scenario, I have multiple isolated > OpenStack cluster running with train & rocky edition. Each OpenStack > cluster environment have it's own isolated infrastructure of SAN ( CISCO > fabric ) & Storage ( HP, EMC & IBM). > > Now company planning to refresh their SAN infrastructure. By procuring new > Brocade SAN switches. But there are some migration relevant challenges we > have. > > As we understand under one cinder instance only one typer of FC zone > manager is supported . Currently customer configured & managing CISCO . > Is it possible to configure two different vendor FC Zone manager under > one cinder instance. Hi, No, it's not currently supported. As you mention, Cinder only supports a single zone manager for all the backends from a single cinder-volume parent process. But there are things that can be done, like running a second cinder-volume service with the new Zone Manager configuration. > Migration of SAN zoning is supposedly going to be happen offline way > from OpenStack point of view. We will be migrating all ports of each > existing cisco fabric to Brocade with zone configuration using brocade > CLI. Our main concern is that after migration How CINDER DB update > new zone info & path via Brocade SAN. When you say that the update is going to happen offline, does it mean all instances will be shutdowwn and have the cinder volumes detached from the VMs? If cinder volumes are detached from instances then everything will be fine. If volumes are still attached (even if VM is shelved) then it's a different story, though it should still be doable with the right procedure. Though I definitely don't feel comfortable providing the procedure since I have never done it and would require appropriate on-site testing. Cheers, Gorka. > > > Regards > Gauurav Sabharwal > IBM India Pvt. Ltd. > IBM towers > Ground floor, Block -A , Plot number 26, > Sector 62, Noida > Gautam budhnagar UP-201307. > Email:gauurav.sabharwal at in.ibm.com > Mobile No.: +91-9910159277 From amy at demarco.com Wed Dec 1 14:52:34 2021 From: amy at demarco.com (Amy Marrich) Date: Wed, 1 Dec 2021 08:52:34 -0600 Subject: [RDO][Meeting] December Meeting Hiatus (12/22 and 12/29) Message-ID: In today's RDO meeting we decided to cancel the meetings which would have been held on 12/22 and 12/29. We will still be holding meetings the other weeks in December in the #RDO channel on OFTC and hope to see you there. Thanks, Amy (spotz) -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Wed Dec 1 14:53:16 2021 From: zigo at debian.org (Thomas Goirand) Date: Wed, 1 Dec 2021 15:53:16 +0100 Subject: [ceilometer][octavia][Victoria] No metrics from octavia loadbalancer In-Reply-To: References: Message-ID: <4acead5e-a812-f7ed-412b-4874af0a4186@debian.org> On 12/1/21 12:17 PM, Iain Stott wrote: > Has anyone had any progress on this? We are currently in the same situation, we can see if loadblancers exists, but unable to see a way to configure the polling agent to poll endpoints like > > v2/lbaas/loadbalancers/[ID]/stats > > > Cheers, > Iain Hi, We do have loadbalancer stats in our public cloud (ie: on the Infomaniak public cloud) going from Ceilometer to Gnocchi storage. Customers are even able to see how many requests are made to the LBs. Our dynamic pollster: # cat /etc/ceilometer/pollsters.d/ik_loadbalancer.yaml --- - name: "ik_loadbalancer" sample_type: "gauge" unit: "loadbalancer" endpoint_type: "load-balancer" url_path: "/loadbalance/v2.0/lbaas/loadbalancers" value_attribute: "provisioning_status" response_entries_key: "loadbalancers" project_id_attribute: "project_id" value_mapping: ACTIVE: "1" ERROR: "0" metadata_fields: - "name" - "description" - "vip_address" I hope this helps, Cheers, Thomas Goirand (zigo) From zbitter at redhat.com Wed Dec 1 15:13:02 2021 From: zbitter at redhat.com (Zane Bitter) Date: Wed, 1 Dec 2021 10:13:02 -0500 Subject: python_requires >= 3.8 during Yoga In-Reply-To: <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: On 26/11/21 10:29, Ghanshyam Mann wrote: > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur wrote ---- > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley wrote: > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > [...] > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > [...] > > > > Is this still true for CentOS Stream 9? The TC decision was to > > support that instead of CentOS Stream 8 in Yoga. > > > > No. But Stream 9 is pretty much beta, so it's not a replacement for us (and we don't have nodes in nodepool with it even yet?). > > I think here is the confusion. In TC, after checking with centos team impression was CentOS stream 9 is released and that is > what we should update In OpenStack testing. And then only we updated the centos stream 8 -> 9 and dropped py3.6 testing > > - https://review.opendev.org/c/openstack/governance/+/815851/3..6/reference/runtimes/yoga.rst The guidelines the TC have set are not that something exists, but that it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable LTS releases. It's not clear to me why CentOS Stream is the only distro being treated differently. The only difference is there is no plan for a CentOS-branded release to define a point in time where Stream 9 becomes an LTS release. However, other parties do have such plans, but pointedly have not done so: RHEL9 is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to release a version based on Stream 9. Presumably RDO folks were consulted about this decision and were OK with the time frame. However, there are other users out there, and from a Metal? perspective this is a giant PITA, requiring us to move from a LTS distro to a beta one, that was dropped on us in the middle of a release cycle in flagrant violation of the TC's own guidelines that the stable distibutions must be chosen from among those available at the *beginning* of the release cycle (which CentOS Stream 9 was not). cheers, Zane. From abraden at verisign.com Wed Dec 1 16:01:28 2021 From: abraden at verisign.com (Braden, Albert) Date: Wed, 1 Dec 2021 16:01:28 +0000 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: It appears that Centos is no longer a viable platform for production Openstack clusters. Of course we have to continue supporting those who are not yet able to move to another distro, but I think we should be clear-eyed about the fact that these are stopgap measures that only need to be implemented while people work on moving. Is anyone contemplating continuing to use Centos long-term? If so, I would be interested to hear your reasoning. -----Original Message----- From: Zane Bitter Sent: Wednesday, December 1, 2021 10:13 AM To: openstack-discuss at lists.openstack.org Subject: [EXTERNAL] Re: python_requires >= 3.8 during Yoga Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. On 26/11/21 10:29, Ghanshyam Mann wrote: > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur wrote ---- > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley wrote: > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > [...] > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > [...] > > > > Is this still true for CentOS Stream 9? The TC decision was to > > support that instead of CentOS Stream 8 in Yoga. > > > > No. But Stream 9 is pretty much beta, so it's not a replacement for us (and we don't have nodes in nodepool with it even yet?). > > I think here is the confusion. In TC, after checking with centos team impression was CentOS stream 9 is released and that is > what we should update In OpenStack testing. And then only we updated the centos stream 8 -> 9 and dropped py3.6 testing > > - https://secure-web.cisco.com/1l5YAwIy_Kf9i8nZRj01Av73trnFQMqoxRgz_2n5WHVL6dc2mfcz-we8VRFvRToxc9yvnpH8QhTDlo2oJoiPHPBheqzFvIaRh40w5Ib3WUxBlvdAfSSNFxyJmXgPOxrq_AwWW27UvaTeFD_ycxhRyngSr_hY7Hji2WkdMMsFl19QfRhk20MI9giWNQk6uMAKlsLKRl4Zuod2cfgERb8Fwm5qwZkfA8NkOU9gZr4IF-nbSgg5aLbgowl4imparhsKS/https%3A%2F%2Freview.opendev.org%2Fc%2Fopenstack%2Fgovernance%2F%2B%2F815851%2F3..6%2Freference%2Fruntimes%2Fyoga.rst The guidelines the TC have set are not that something exists, but that it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable LTS releases. It's not clear to me why CentOS Stream is the only distro being treated differently. The only difference is there is no plan for a CentOS-branded release to define a point in time where Stream 9 becomes an LTS release. However, other parties do have such plans, but pointedly have not done so: RHEL9 is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to release a version based on Stream 9. Presumably RDO folks were consulted about this decision and were OK with the time frame. However, there are other users out there, and from a Metal? perspective this is a giant PITA, requiring us to move from a LTS distro to a beta one, that was dropped on us in the middle of a release cycle in flagrant violation of the TC's own guidelines that the stable distibutions must be chosen from among those available at the *beginning* of the release cycle (which CentOS Stream 9 was not). cheers, Zane. From gmann at ghanshyammann.com Wed Dec 1 16:06:02 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Wed, 01 Dec 2021 10:06:02 -0600 Subject: [all][qa][triplo] Help need to maintain OpenStack health dashboard Message-ID: <17d76bee71c.cfa2c4c1178950.6715520613968754830@ghanshyammann.com> Hello Everyone, In the QA meeting[1], we discussed the help needed to maintain the OpenStack health dashboard which is broken currently and the QA team does not have the JS developer and bandwidth to fix it. While discussing it meeting, we found that the TripleO team might be working on a similar dashboard for their CI results. If that is the case, can we collaborate on maintaining the existing dashboard? OpenStack health dashboard: https://opendev.org/openstack/openstack-health Repo: http://status.openstack.org/openstack-health/#/ [1] https://meetings.opendev.org/meetings/qa/2021/qa.2021-11-16-15.00.log.html#l-71 -gmann From aschultz at redhat.com Wed Dec 1 16:13:07 2021 From: aschultz at redhat.com (Alex Schultz) Date: Wed, 1 Dec 2021 09:13:07 -0700 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: On Wed, Dec 1, 2021 at 9:04 AM Braden, Albert wrote: > > It appears that Centos is no longer a viable platform for production Openstack clusters. Of course we have to continue supporting those who are not yet able to move to another distro, but I think we should be clear-eyed about the fact that these are stopgap measures that only need to be implemented while people work on moving. > I'm not certain why this is how it's being interpreted. CentOS stream is perfectly viable but it is (as has always been) reliant on packaging. I believe Cloud SIG will still exist and RDO will be publishing packages. In fact CentOS stream is generally better than the previous release cadence that caused all sort of issues when point releases dropped with breaking changes. The difference in stream vs legacy method was the volume of changes over time. Stream you get smaller changes sooner (along with fixes) where as the legacy method you'd only really get updates at point release time and it was always terrible troubleshooting these failures. CentOS Stream 9 will have 3.9 as the default but we need to make sure that we can get it available to the various projects for testing. Right now Puppet OpenStack primarily relies on CentOS for testing (because Ubuntu jobs have been broken for a few years now with no one jumping in to fix). So bumping the minimum version without having a way to deploy for us would directly impact this project. IMHO the issue with this thread is timing on being able to bump a minimum python version requirement. I personally wouldn't want to bump to something that excludes 3.6 at this time without a really good reason. Are there features we require that will suddenly make developer/operator lives easier (e.g. switching to asyncio or something)? If not, what is the driving factor? I think pushing it off to Z would be best to allow for the initial standup required to have *all* the projects able to support the switch. > Is anyone contemplating continuing to use Centos long-term? If so, I would be interested to hear your reasoning. > > -----Original Message----- > From: Zane Bitter > Sent: Wednesday, December 1, 2021 10:13 AM > To: openstack-discuss at lists.openstack.org > Subject: [EXTERNAL] Re: python_requires >= 3.8 during Yoga > > Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. > > On 26/11/21 10:29, Ghanshyam Mann wrote: > > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur wrote ---- > > > > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley wrote: > > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > > [...] > > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > > [...] > > > > > > Is this still true for CentOS Stream 9? The TC decision was to > > > support that instead of CentOS Stream 8 in Yoga. > > > > > > No. But Stream 9 is pretty much beta, so it's not a replacement for us (and we don't have nodes in nodepool with it even yet?). > > > > I think here is the confusion. In TC, after checking with centos team impression was CentOS stream 9 is released and that is > > what we should update In OpenStack testing. And then only we updated the centos stream 8 -> 9 and dropped py3.6 testing > > > > - https://secure-web.cisco.com/1l5YAwIy_Kf9i8nZRj01Av73trnFQMqoxRgz_2n5WHVL6dc2mfcz-we8VRFvRToxc9yvnpH8QhTDlo2oJoiPHPBheqzFvIaRh40w5Ib3WUxBlvdAfSSNFxyJmXgPOxrq_AwWW27UvaTeFD_ycxhRyngSr_hY7Hji2WkdMMsFl19QfRhk20MI9giWNQk6uMAKlsLKRl4Zuod2cfgERb8Fwm5qwZkfA8NkOU9gZr4IF-nbSgg5aLbgowl4imparhsKS/https%3A%2F%2Freview.opendev.org%2Fc%2Fopenstack%2Fgovernance%2F%2B%2F815851%2F3..6%2Freference%2Fruntimes%2Fyoga.rst > > The guidelines the TC have set are not that something exists, but that > it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, > and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable > LTS releases. It's not clear to me why CentOS Stream is the only distro > being treated differently. > > The only difference is there is no plan for a CentOS-branded release to > define a point in time where Stream 9 becomes an LTS release. However, > other parties do have such plans, but pointedly have not done so: RHEL9 > is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to > release a version based on Stream 9. > > Presumably RDO folks were consulted about this decision and were OK with > the time frame. However, there are other users out there, and from a > Metal? perspective this is a giant PITA, requiring us to move from a LTS > distro to a beta one, that was dropped on us in the middle of a release > cycle in flagrant violation of the TC's own guidelines that the stable > distibutions must be chosen from among those available at the > *beginning* of the release cycle (which CentOS Stream 9 was not). > > cheers, > Zane. > > From gmann at ghanshyammann.com Wed Dec 1 16:28:15 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Wed, 01 Dec 2021 10:28:15 -0600 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: <17d76d33dc7.113f517f7180736.3792235396720560154@ghanshyammann.com> ---- On Wed, 01 Dec 2021 09:13:02 -0600 Zane Bitter wrote ---- > On 26/11/21 10:29, Ghanshyam Mann wrote: > > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur wrote ---- > > > > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley wrote: > > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > > [...] > > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > > [...] > > > > > > Is this still true for CentOS Stream 9? The TC decision was to > > > support that instead of CentOS Stream 8 in Yoga. > > > > > > No. But Stream 9 is pretty much beta, so it's not a replacement for us (and we don't have nodes in nodepool with it even yet?). > > > > I think here is the confusion. In TC, after checking with centos team impression was CentOS stream 9 is released and that is > > what we should update In OpenStack testing. And then only we updated the centos stream 8 -> 9 and dropped py3.6 testing > > > > - https://review.opendev.org/c/openstack/governance/+/815851/3..6/reference/runtimes/yoga.rst > > The guidelines the TC have set are not that something exists, but that > it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, > and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable > LTS releases. It's not clear to me why CentOS Stream is the only distro > being treated differently. > > The only difference is there is no plan for a CentOS-branded release to > define a point in time where Stream 9 becomes an LTS release. However, > other parties do have such plans, but pointedly have not done so: RHEL9 > is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to > release a version based on Stream 9. > > Presumably RDO folks were consulted about this decision and were OK with > the time frame. However, there are other users out there, and from a > Metal? perspective this is a giant PITA, requiring us to move from a LTS > distro to a beta one, that was dropped on us in the middle of a release > cycle in flagrant violation of the TC's own guidelines that the stable > distibutions must be chosen from among those available at the > *beginning* of the release cycle (which CentOS Stream 9 was not). Those are good points. We were discussing with the RDO team on the start of Yoga cycle itself whether to move to CentOS Stream9 or not. Yes, there is a clear confusion on what is stable or not in CentOS Stream distro. Like other distro has, is there any document we as TC can refer to in the future on checking if the CentOS Stream *X* version is stable/LTS or not? I cannot find it for CentOS Stream 8 also, is that stable or LTS? If we have such official documentation or announcement in RDO community then we can avoid such situations in future where we get to know the distro version stability well before trying it in OpenStack testing? - https://centos.org/download/ -gmann > > cheers, > Zane. > > > From cboylan at sapwetik.org Wed Dec 1 16:31:43 2021 From: cboylan at sapwetik.org (Clark Boylan) Date: Wed, 01 Dec 2021 08:31:43 -0800 Subject: [all][qa][triplo] Help need to maintain OpenStack health dashboard In-Reply-To: <17d76bee71c.cfa2c4c1178950.6715520613968754830@ghanshyammann.com> References: <17d76bee71c.cfa2c4c1178950.6715520613968754830@ghanshyammann.com> Message-ID: <64474420-8cc9-4db8-aec7-5a788c5cf762@www.fastmail.com> On Wed, Dec 1, 2021, at 8:06 AM, Ghanshyam Mann wrote: > Hello Everyone, > > In the QA meeting[1], we discussed the help needed to maintain the > OpenStack health dashboard which is > broken currently and the QA team does not have the JS developer and > bandwidth to fix it. While discussing > it meeting, we found that the TripleO team might be working on a > similar dashboard for their CI results. If that > is the case, can we collaborate on maintaining the existing dashboard? Note, this will also require operational assistance to deploy the subunit2sql processing pipeline, the subunit2sql database backend, and the health api service. These services are planned to be turned off at the end of the current cycle per prior discussion with the TC. > > OpenStack health dashboard: https://opendev.org/openstack/openstack-health > Repo: http://status.openstack.org/openstack-health/#/ > > [1] > https://meetings.opendev.org/meetings/qa/2021/qa.2021-11-16-15.00.log.html#l-71 > > -gmann From aschultz at redhat.com Wed Dec 1 16:43:10 2021 From: aschultz at redhat.com (Alex Schultz) Date: Wed, 1 Dec 2021 09:43:10 -0700 Subject: python_requires >= 3.8 during Yoga In-Reply-To: <17d76d33dc7.113f517f7180736.3792235396720560154@ghanshyammann.com> References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <17d76d33dc7.113f517f7180736.3792235396720560154@ghanshyammann.com> Message-ID: On Wed, Dec 1, 2021 at 9:30 AM Ghanshyam Mann wrote: > > ---- On Wed, 01 Dec 2021 09:13:02 -0600 Zane Bitter wrote ---- > > On 26/11/21 10:29, Ghanshyam Mann wrote: > > > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur wrote ---- > > > > > > > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley wrote: > > > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > > > [...] > > > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > > > [...] > > > > > > > > Is this still true for CentOS Stream 9? The TC decision was to > > > > support that instead of CentOS Stream 8 in Yoga. > > > > > > > > No. But Stream 9 is pretty much beta, so it's not a replacement for us (and we don't have nodes in nodepool with it even yet?). > > > > > > I think here is the confusion. In TC, after checking with centos team impression was CentOS stream 9 is released and that is > > > what we should update In OpenStack testing. And then only we updated the centos stream 8 -> 9 and dropped py3.6 testing > > > > > > - https://review.opendev.org/c/openstack/governance/+/815851/3..6/reference/runtimes/yoga.rst > > > > The guidelines the TC have set are not that something exists, but that > > it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, > > and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable > > LTS releases. It's not clear to me why CentOS Stream is the only distro > > being treated differently. > > > > The only difference is there is no plan for a CentOS-branded release to > > define a point in time where Stream 9 becomes an LTS release. However, > > other parties do have such plans, but pointedly have not done so: RHEL9 > > is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to > > release a version based on Stream 9. > > > > Presumably RDO folks were consulted about this decision and were OK with > > the time frame. However, there are other users out there, and from a > > Metal? perspective this is a giant PITA, requiring us to move from a LTS > > distro to a beta one, that was dropped on us in the middle of a release > > cycle in flagrant violation of the TC's own guidelines that the stable > > distibutions must be chosen from among those available at the > > *beginning* of the release cycle (which CentOS Stream 9 was not). > > Those are good points. We were discussing with the RDO team on the start > of Yoga cycle itself whether to move to CentOS Stream9 or not. > > Yes, there is a clear confusion on what is stable or not in CentOS Stream distro. > Like other distro has, is there any document we as TC can refer to in the future on > checking if the CentOS Stream *X* version is stable/LTS or not? I cannot find it > for CentOS Stream 8 also, is that stable or LTS? > I guess what's the definition of LTS? Support > 1 year? https://www.centos.org/cl-vs-cs/ CentOS Stream 8 is to be supported till 2024 and 9 is estimated until 2027. That seems pretty LTS to me. Additionally what is the definition of stable in this case? My understanding is that there still won't be major version bumps to core software similar to how CentOS used to operate. So it'd be a similar stability as previously existed in the non stream 7/8. What changes is that fixes will be available sooner, but it would have been the same that would have been released at a previous point release in the old method. > If we have such official documentation or announcement in RDO community > then we can avoid such situations in future where we get to know the distro > version stability well before trying it in OpenStack testing? > > - https://centos.org/download/ > > -gmann > > > > > cheers, > > Zane. > > > > > > > From gmann at ghanshyammann.com Wed Dec 1 16:43:46 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Wed, 01 Dec 2021 10:43:46 -0600 Subject: [all][qa][triplo] Help need to maintain OpenStack health dashboard In-Reply-To: <64474420-8cc9-4db8-aec7-5a788c5cf762@www.fastmail.com> References: <17d76bee71c.cfa2c4c1178950.6715520613968754830@ghanshyammann.com> <64474420-8cc9-4db8-aec7-5a788c5cf762@www.fastmail.com> Message-ID: <17d76e17422.103578670181902.6372441011959447878@ghanshyammann.com> ---- On Wed, 01 Dec 2021 10:31:43 -0600 Clark Boylan wrote ---- > On Wed, Dec 1, 2021, at 8:06 AM, Ghanshyam Mann wrote: > > Hello Everyone, > > > > In the QA meeting[1], we discussed the help needed to maintain the > > OpenStack health dashboard which is > > broken currently and the QA team does not have the JS developer and > > bandwidth to fix it. While discussing > > it meeting, we found that the TripleO team might be working on a > > similar dashboard for their CI results. If that > > is the case, can we collaborate on maintaining the existing dashboard? > > Note, this will also require operational assistance to deploy the subunit2sql processing pipeline, the subunit2sql database backend, and the health api service. These services are planned to be turned off at the end of the current cycle per prior discussion with the TC. +1, if the TripleO team working on a similar setup or so then we can reuse the existing dashboard or operational assistance. Otherwise, along with services shutdown what Clark mentioned above, the QA team will also start (have no choice) the repo retirement process also (which will impact anyone using it for downstream usage). -gmann > > > > > OpenStack health dashboard: https://opendev.org/openstack/openstack-health > > Repo: http://status.openstack.org/openstack-health/#/ > > > > [1] > > https://meetings.opendev.org/meetings/qa/2021/qa.2021-11-16-15.00.log.html#l-71 > > > > -gmann > > From rlandy at redhat.com Wed Dec 1 16:46:26 2021 From: rlandy at redhat.com (Ronelle Landy) Date: Wed, 1 Dec 2021 11:46:26 -0500 Subject: [all][qa][triplo] Help need to maintain OpenStack health dashboard In-Reply-To: <17d76bee71c.cfa2c4c1178950.6715520613968754830@ghanshyammann.com> References: <17d76bee71c.cfa2c4c1178950.6715520613968754830@ghanshyammann.com> Message-ID: On Wed, Dec 1, 2021 at 11:11 AM Ghanshyam Mann wrote: > Hello Everyone, > > In the QA meeting[1], we discussed the help needed to maintain the > OpenStack health dashboard which is > broken currently and the QA team does not have the JS developer and > bandwidth to fix it. While discussing > it meeting, we found that the TripleO team might be working on a similar > dashboard for their CI results. If that > is the case, can we collaborate on maintaining the existing dashboard? > We had a discussion with Martin Kopec about collaboration with QA in maintaining a TripleO focused dashboard. We have two dashboards in progress at the moment - one focused on jobs running in https://review.rdoproject.org/zuul/status - http://ci-health-rdo.tripleo.org/ and one where we are starting to track failures in select jobs running on https://zuul.openstack.org/ - http://ci-health.tripleo.org/ If you would like to collaborate on this work, please ping us on #oooq on (OFTC) join our community call on Tuesdays at 1:30pm UTC and we can discuss further. Thanks! > > OpenStack health dashboard: https://opendev.org/openstack/openstack-health > Repo: http://status.openstack.org/openstack-health/#/ > > [1] > https://meetings.opendev.org/meetings/qa/2021/qa.2021-11-16-15.00.log.html#l-71 > > -gmann > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From smooney at redhat.com Wed Dec 1 17:05:34 2021 From: smooney at redhat.com (Sean Mooney) Date: Wed, 01 Dec 2021 17:05:34 +0000 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: On Wed, 2021-12-01 at 09:13 -0700, Alex Schultz wrote: > On Wed, Dec 1, 2021 at 9:04 AM Braden, Albert wrote: > > > > It appears that Centos is no longer a viable platform for production Openstack clusters. Of course we have to continue supporting those who are not yet able to move to another distro, but I think we should be clear-eyed about the fact that these are stopgap measures that only need to be implemented while people work on moving. > > > > I'm not certain why this is how it's being interpreted. CentOS stream > is perfectly viable but it is (as has always been) reliant on > packaging. I believe Cloud SIG will still exist and RDO will be > publishing packages. In fact CentOS stream is generally better than > the previous release cadence that caused all sort of issues when point > releases dropped with breaking changes. The difference in stream vs > legacy method was the volume of changes over time. Stream you get > smaller changes sooner (along with fixes) where as the legacy method > you'd only really get updates at point release time and it was always > terrible troubleshooting these failures. CentOS Stream 9 will have > 3.9 as the default but we need to make sure that we can get it > available to the various projects for testing. Right now Puppet > OpenStack primarily relies on CentOS for testing (because Ubuntu jobs > have been broken for a few years now with no one jumping in to fix). > So bumping the minimum version without having a way to deploy for us > would directly impact this project. at least internally earlier this year i reached out about rhel9/centos9 supprot specifcly related to py 3.9 support and eventlets. my understandwing was there was already some basic rhel9/centos9 ci happening the puppet moduels since eventlet supprot was tested using a puppet module job. > > IMHO the issue with this thread is timing on being able to bump a > minimum python version requirement. I personally wouldn't want to > bump to something that excludes 3.6 at this time without a really good > reason. > there is/was an expecation that rdo woudl backport support for cento 9 and py39 to stable wababy at some point since our downstream osp 17 product will be based on rhel 9 on stable/walaby. im not directly involved in that packaging effort but i had assumed perhaps incorreectly that redhat would do that work upstream in rdo? is this not the case? at least form a downstram perspective unless we cahnge direction and decided to build OSP 17 using ubi8 images instead fo ubi9 we will need to do all this packaging work anyway so im surprised there is this much push back to moving to centos 9 and py39. > Are there features we require that will suddenly make > developer/operator lives easier (e.g. switching to asyncio or > something)? > actuly they are a few nice libary enhanchments but that is not the main driving factor. many do not want to support deploying the openstack project on python version we do not test. so either we keep 3.6 support and test it in every porject or we drop it and test with 3.8 and 3.9. > If not, what is the driving factor? > i have been pushing for 3.9 testing to become required for all project because centos/rhel 9 will only ship 3.9+ and it will be what our downstream product will be based on and i belive 3.9 is already used on debian 11 and 3.10 i belive is slated for ubuntu 22.04 which will be the next ubuntu lts and will supprot/ship openstack yoga. so haveing at least 3.9 support in yoga will be imporatn for multiple distros and to help use test that i and proably others suggested raising the minium python version to 3.8 so that we only have to test with 2 version in the gate instead of 3.(3.6 3.8 and 3.9) continuting to support py36 aslo puts a burden on developers not just on our ci infrastucre as it will becomre harder to test wtih going froward as it stops being shipped in distros and as it exits security support the end of life for python 3.6 is planned for this month https://www.python.org/dev/peps/pep-0494/#lifespan speficically (23 Dec 2021) based on https://endoflife.date/python i dont think we shoudl be releasing yoga on an end of life python version. if we raise the minium to 3.8 that will be supported until at least 14 Oct 2024 which is after the stabel branch will be retired. not support python 3.9 and 3.10 for yoga will make packaging openstack very hard for many ditros but since we cant realticaly test 4 verions properly we have to increase the minium verions at some point. > I think pushing it > off to Z would be best to allow for the initial standup required to > have *all* the projects able to support the switch. all project where ment to start adding 3.9 support already it has been a non vovting test requiremtn for 2? cycles i belvie since we knew it was gong to be used in debian 11 and centos 9 for well over a year. i do agree that its unfortunet that centos 9 missed it release target and did not release before the start of the yoga cycle but a better way to adress the upgrade issue i think woudl be to support xena/wallaby on centos 8 and 9 and make yoga 9 only. if we need to supprot yoga on centos 8 because that is not possibel then we coudl keep 3.6 support for one addtional release i guss but that wont be something that we just get for free and it will require use to reinstaet all 3.6 based ci and acnockolage that by the time yoga is released 3.6 will be EOL. > > > Is anyone contemplating continuing to use Centos long-term? If so, I would be interested to hear your reasoning. > > > > -----Original Message----- > > From: Zane Bitter > > Sent: Wednesday, December 1, 2021 10:13 AM > > To: openstack-discuss at lists.openstack.org > > Subject: [EXTERNAL] Re: python_requires >= 3.8 during Yoga > > > > Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. > > > > On 26/11/21 10:29, Ghanshyam Mann wrote: > > > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur wrote ---- > > > > > > > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley wrote: > > > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > > > [...] > > > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > > > [...] > > > > > > > > Is this still true for CentOS Stream 9? The TC decision was to > > > > support that instead of CentOS Stream 8 in Yoga. > > > > > > > > No. But Stream 9 is pretty much beta, so it's not a replacement for us (and we don't have nodes in nodepool with it even yet?). > > > > > > I think here is the confusion. In TC, after checking with centos team impression was CentOS stream 9 is released and that is > > > what we should update In OpenStack testing. And then only we updated the centos stream 8 -> 9 and dropped py3.6 testing > > > > > > - https://secure-web.cisco.com/1l5YAwIy_Kf9i8nZRj01Av73trnFQMqoxRgz_2n5WHVL6dc2mfcz-we8VRFvRToxc9yvnpH8QhTDlo2oJoiPHPBheqzFvIaRh40w5Ib3WUxBlvdAfSSNFxyJmXgPOxrq_AwWW27UvaTeFD_ycxhRyngSr_hY7Hji2WkdMMsFl19QfRhk20MI9giWNQk6uMAKlsLKRl4Zuod2cfgERb8Fwm5qwZkfA8NkOU9gZr4IF-nbSgg5aLbgowl4imparhsKS/https%3A%2F%2Freview.opendev.org%2Fc%2Fopenstack%2Fgovernance%2F%2B%2F815851%2F3..6%2Freference%2Fruntimes%2Fyoga.rst > > > > The guidelines the TC have set are not that something exists, but that > > it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, > > and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable > > LTS releases. It's not clear to me why CentOS Stream is the only distro > > being treated differently. > > > > The only difference is there is no plan for a CentOS-branded release to > > define a point in time where Stream 9 becomes an LTS release. However, > > other parties do have such plans, but pointedly have not done so: RHEL9 > > is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to > > release a version based on Stream 9. > > > > Presumably RDO folks were consulted about this decision and were OK with > > the time frame. However, there are other users out there, and from a > > Metal? perspective this is a giant PITA, requiring us to move from a LTS > > distro to a beta one, that was dropped on us in the middle of a release > > cycle in flagrant violation of the TC's own guidelines that the stable > > distibutions must be chosen from among those available at the > > *beginning* of the release cycle (which CentOS Stream 9 was not). > > > > cheers, > > Zane. > > > > > > From amoralej at redhat.com Wed Dec 1 17:30:01 2021 From: amoralej at redhat.com (Alfredo Moralejo Alonso) Date: Wed, 1 Dec 2021 18:30:01 +0100 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <17d76d33dc7.113f517f7180736.3792235396720560154@ghanshyammann.com> Message-ID: On Wed, Dec 1, 2021 at 5:48 PM Alex Schultz wrote: > On Wed, Dec 1, 2021 at 9:30 AM Ghanshyam Mann > wrote: > > > > ---- On Wed, 01 Dec 2021 09:13:02 -0600 Zane Bitter > wrote ---- > > > On 26/11/21 10:29, Ghanshyam Mann wrote: > > > > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur < > dtantsur at redhat.com> wrote ---- > > > > > > > > > > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley < > fungi at yuggoth.org> wrote: > > > > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > > > > [...] > > > > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > > > > [...] > > > > > > > > > > Is this still true for CentOS Stream 9? The TC decision was to > > > > > support that instead of CentOS Stream 8 in Yoga. > > > > > > > > > > No. But Stream 9 is pretty much beta, so it's not a replacement > for us (and we don't have nodes in nodepool with it even yet?). > > > > > > > > I think here is the confusion. In TC, after checking with centos > team impression was CentOS stream 9 is released and that is > > > > what we should update In OpenStack testing. And then only we > updated the centos stream 8 -> 9 and dropped py3.6 testing > > > > > > > > - > https://review.opendev.org/c/openstack/governance/+/815851/3..6/reference/runtimes/yoga.rst > > > > > > The guidelines the TC have set are not that something exists, but that > > > it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, > > > and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable > > > LTS releases. It's not clear to me why CentOS Stream is the only > distro > > > being treated differently. > > > > > > The only difference is there is no plan for a CentOS-branded release > to > > > define a point in time where Stream 9 becomes an LTS release. However, > > > other parties do have such plans, but pointedly have not done so: > RHEL9 > > > is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to > > > release a version based on Stream 9. > > > > > > Presumably RDO folks were consulted about this decision and were OK > with > > > the time frame. However, there are other users out there, and from a > > > Metal? perspective this is a giant PITA, requiring us to move from a > LTS > > > distro to a beta one, that was dropped on us in the middle of a > release > > > cycle in flagrant violation of the TC's own guidelines that the stable > > > distibutions must be chosen from among those available at the > > > *beginning* of the release cycle (which CentOS Stream 9 was not). > > > > Those are good points. We were discussing with the RDO team on the start > > of Yoga cycle itself whether to move to CentOS Stream9 or not. > > > > Yes, there is a clear confusion on what is stable or not in CentOS > Stream distro. > > Like other distro has, is there any document we as TC can refer to in > the future on > > checking if the CentOS Stream *X* version is stable/LTS or not? I cannot > find it > > for CentOS Stream 8 also, is that stable or LTS? > > > > I guess what's the definition of LTS? Support > 1 year? > > https://www.centos.org/cl-vs-cs/ > > CentOS Stream 8 is to be supported till 2024 and 9 is estimated until > 2027. That seems pretty LTS to me. > > Actually, there are no special LTS releases in CentOS. In Red Hat family distros, Fedora would be the non-LTS distro while CentOS is the LTS one. > Additionally what is the definition of stable in this case? My > understanding is that there still won't be major version bumps to core > software similar to how CentOS used to operate. So it'd be a similar > stability as previously existed in the non stream 7/8. What changes is > that fixes will be available sooner, but it would have been the same > that would have been released at a previous point release in the old > method. > > Exactly, so as mentioned several times in this thread the main difference between CentOS Stream and CentOS Linux is the fact that updates are not done in big batches (minor releases) but in more frequent and small package updates in a continuous delivery approach [1]. Stability should not be very different to what we had with CentOS Linux. [1] https://blog.centos.org/2020/12/centos-stream-is-continuous-delivery/ > > If we have such official documentation or announcement in RDO community > > then we can avoid such situations in future where we get to know the > distro > > version stability well before trying it in OpenStack testing? > > > > - https://centos.org/download/ > > > > -gmann > > > > > > > > cheers, > > > Zane. > > > > > > > > > > > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From aschultz at redhat.com Wed Dec 1 17:31:58 2021 From: aschultz at redhat.com (Alex Schultz) Date: Wed, 1 Dec 2021 10:31:58 -0700 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: On Wed, Dec 1, 2021 at 10:05 AM Sean Mooney wrote: > > On Wed, 2021-12-01 at 09:13 -0700, Alex Schultz wrote: > > On Wed, Dec 1, 2021 at 9:04 AM Braden, Albert wrote: > > > > > > It appears that Centos is no longer a viable platform for production Openstack clusters. Of course we have to continue supporting those who are not yet able to move to another distro, but I think we should be clear-eyed about the fact that these are stopgap measures that only need to be implemented while people work on moving. > > > > > > > I'm not certain why this is how it's being interpreted. CentOS stream > > is perfectly viable but it is (as has always been) reliant on > > packaging. I believe Cloud SIG will still exist and RDO will be > > publishing packages. In fact CentOS stream is generally better than > > the previous release cadence that caused all sort of issues when point > > releases dropped with breaking changes. The difference in stream vs > > legacy method was the volume of changes over time. Stream you get > > smaller changes sooner (along with fixes) where as the legacy method > > you'd only really get updates at point release time and it was always > > terrible troubleshooting these failures. CentOS Stream 9 will have > > 3.9 as the default but we need to make sure that we can get it > > available to the various projects for testing. Right now Puppet > > OpenStack primarily relies on CentOS for testing (because Ubuntu jobs > > have been broken for a few years now with no one jumping in to fix). > > So bumping the minimum version without having a way to deploy for us > > would directly impact this project. > > at least internally earlier this year i reached out about rhel9/centos9 supprot > specifcly related to py 3.9 support and eventlets. > my understandwing was there was already some basic rhel9/centos9 ci happening the puppet moduels > since eventlet supprot was tested using a puppet module job. > > > > > IMHO the issue with this thread is timing on being able to bump a > > minimum python version requirement. I personally wouldn't want to > > bump to something that excludes 3.6 at this time without a really good > > reason. > > > > there is/was an expecation that rdo woudl backport support for cento 9 and py39 to stable wababy at > some point since our downstream osp 17 product will be based on rhel 9 on stable/walaby. > > im not directly involved in that packaging effort but i had assumed perhaps incorreectly that redhat would > do that work upstream in rdo? is this not the case? > It is, however, because it's not done it is not OK to just do the switch. Really the issue is that you can't make a blanket switch until much of the bootstrapping work has been completed which is why doing the switch in Z makes more sense to allow for this. > at least form a downstram perspective unless we cahnge direction and decided to build OSP 17 using ubi8 images > instead fo ubi9 we will need to do all this packaging work anyway so im surprised there is this much push back > to moving to centos 9 and py39. > This is the same problem we had with the 7/8 switch where train was ultimately the thing where we had to backport support for py2.7/py3 problems. I would have really preferred to not do this again with 8/9 but here we are. IIRC, we managed to get Ussuri switched to centos8. Due to the delay in 9 availability we weren't able to get it ready in the Xena cycle. We do want to get it done in Yoga but because this thread is about making mandatory, we need to wait until the switch has been completed. > > Are there features we require that will suddenly make > > developer/operator lives easier (e.g. switching to asyncio or > > something)? > > > actuly they are a few nice libary enhanchments but that is not the main driving factor. > > many do not want to support deploying the openstack project on python version we do not test. > so either we keep 3.6 support and test it in every porject or we drop it and test with 3.8 and 3.9. > > > If not, what is the driving factor? > > > > i have been pushing for 3.9 testing to become required for all project because centos/rhel 9 will only ship 3.9+ > and it will be what our downstream product will be based on and i belive 3.9 is already used on debian 11 and 3.10 i belive > is slated for ubuntu 22.04 which will be the next ubuntu lts and will supprot/ship openstack yoga. > That's fine to push it as a goal, but I just think the goal needs to be delayed until Z. Is there anything that is broken in 3.9 > so haveing at least 3.9 support in yoga will be imporatn for multiple distros and to help use test that i and proably others > suggested raising the minium python version to 3.8 so that we only have to test with 2 version in the gate instead of 3.(3.6 3.8 and 3.9) > > continuting to support py36 aslo puts a burden on developers not just on our ci infrastucre as it will becomre harder to test wtih going froward > as it stops being shipped in distros and as it exits security support > the end of life for python 3.6 is planned for this month https://www.python.org/dev/peps/pep-0494/#lifespan > speficically (23 Dec 2021) based on https://endoflife.date/python > > i dont think we shoudl be releasing yoga on an end of life python version. > if we raise the minium to 3.8 that will be supported until at least 14 Oct 2024 > which is after the stabel branch will be retired. > So I guess the question is who do we use for language support EOLs? My understanding is that Ubuntu 18.04's default was 3.6 which has standard support until 2023[0] and CentOS Stream 8 will be supported until 2024[1]. Both of which should exceed Yoga's support[2]? With all these arguments about LTS distros, it seems that we should be using their support life cyles for the language versions and not python's[3]. [0] https://wiki.ubuntu.com/Releases [1] https://www.centos.org/cl-vs-cs/ [2] https://releases.openstack.org/ [3] https://devguide.python.org/#status-of-python-branches > not support python 3.9 and 3.10 for yoga will make packaging openstack very hard for many ditros but since we cant realticaly test 4 verions properly we > have to increase the minium verions at some point. > > I think pushing it > > off to Z would be best to allow for the initial standup required to > > have *all* the projects able to support the switch. > all project where ment to start adding 3.9 support already it has been a non vovting test requiremtn for 2? cycles i belvie > since we knew it was gong to be used in debian 11 and centos 9 for well over a year. > > i do agree that its unfortunet that centos 9 missed it release target and did not release before the start of the yoga cycle > but a better way to adress the upgrade issue i think woudl be to support xena/wallaby on centos 8 and 9 and make yoga 9 only. > if we need to supprot yoga on centos 8 because that is not possibel then we coudl keep 3.6 support for one addtional release i guss > but that wont be something that we just get for free and it will require use to reinstaet all 3.6 based ci and acnockolage that by the > time yoga is released 3.6 will be EOL. > > > > > > Is anyone contemplating continuing to use Centos long-term? If so, I would be interested to hear your reasoning. > > > > > > -----Original Message----- > > > From: Zane Bitter > > > Sent: Wednesday, December 1, 2021 10:13 AM > > > To: openstack-discuss at lists.openstack.org > > > Subject: [EXTERNAL] Re: python_requires >= 3.8 during Yoga > > > > > > Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. > > > > > > On 26/11/21 10:29, Ghanshyam Mann wrote: > > > > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur wrote ---- > > > > > > > > > > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley wrote: > > > > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > > > > [...] > > > > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > > > > [...] > > > > > > > > > > Is this still true for CentOS Stream 9? The TC decision was to > > > > > support that instead of CentOS Stream 8 in Yoga. > > > > > > > > > > No. But Stream 9 is pretty much beta, so it's not a replacement for us (and we don't have nodes in nodepool with it even yet?). > > > > > > > > I think here is the confusion. In TC, after checking with centos team impression was CentOS stream 9 is released and that is > > > > what we should update In OpenStack testing. And then only we updated the centos stream 8 -> 9 and dropped py3.6 testing > > > > > > > > - https://secure-web.cisco.com/1l5YAwIy_Kf9i8nZRj01Av73trnFQMqoxRgz_2n5WHVL6dc2mfcz-we8VRFvRToxc9yvnpH8QhTDlo2oJoiPHPBheqzFvIaRh40w5Ib3WUxBlvdAfSSNFxyJmXgPOxrq_AwWW27UvaTeFD_ycxhRyngSr_hY7Hji2WkdMMsFl19QfRhk20MI9giWNQk6uMAKlsLKRl4Zuod2cfgERb8Fwm5qwZkfA8NkOU9gZr4IF-nbSgg5aLbgowl4imparhsKS/https%3A%2F%2Freview.opendev.org%2Fc%2Fopenstack%2Fgovernance%2F%2B%2F815851%2F3..6%2Freference%2Fruntimes%2Fyoga.rst > > > > > > The guidelines the TC have set are not that something exists, but that > > > it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, > > > and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable > > > LTS releases. It's not clear to me why CentOS Stream is the only distro > > > being treated differently. > > > > > > The only difference is there is no plan for a CentOS-branded release to > > > define a point in time where Stream 9 becomes an LTS release. However, > > > other parties do have such plans, but pointedly have not done so: RHEL9 > > > is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to > > > release a version based on Stream 9. > > > > > > Presumably RDO folks were consulted about this decision and were OK with > > > the time frame. However, there are other users out there, and from a > > > Metal? perspective this is a giant PITA, requiring us to move from a LTS > > > distro to a beta one, that was dropped on us in the middle of a release > > > cycle in flagrant violation of the TC's own guidelines that the stable > > > distibutions must be chosen from among those available at the > > > *beginning* of the release cycle (which CentOS Stream 9 was not). > > > > > > cheers, > > > Zane. > > > > > > > > > > > From cboylan at sapwetik.org Wed Dec 1 18:06:16 2021 From: cboylan at sapwetik.org (Clark Boylan) Date: Wed, 01 Dec 2021 10:06:16 -0800 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: <50abb4f5-af13-435c-87c9-1c14d74dcfb2@www.fastmail.com> On Wed, Dec 1, 2021, at 9:31 AM, Alex Schultz wrote: > On Wed, Dec 1, 2021 at 10:05 AM Sean Mooney wrote: > Apologies if my reduction of thread content was too aggressive. > >> so haveing at least 3.9 support in yoga will be imporatn for multiple distros and to help use test that i and proably others >> suggested raising the minium python version to 3.8 so that we only have to test with 2 version in the gate instead of 3.(3.6 3.8 and 3.9) >> >> continuting to support py36 aslo puts a burden on developers not just on our ci infrastucre as it will becomre harder to test wtih going froward >> as it stops being shipped in distros and as it exits security support >> the end of life for python 3.6 is planned for this month https://www.python.org/dev/peps/pep-0494/#lifespan >> speficically (23 Dec 2021) based on https://endoflife.date/python >> >> i dont think we shoudl be releasing yoga on an end of life python version. >> if we raise the minium to 3.8 that will be supported until at least 14 Oct 2024 >> which is after the stabel branch will be retired. >> > > So I guess the question is who do we use for language support EOLs? > My understanding is that Ubuntu 18.04's default was 3.6 which has > standard support until 2023[0] and CentOS Stream 8 will be supported > until 2024[1]. Both of which should exceed Yoga's support[2]? With > all these arguments about LTS distros, it seems that we should be > using their support life cyles for the language versions and not > python's[3]. > > [0] https://wiki.ubuntu.com/Releases > [1] https://www.centos.org/cl-vs-cs/ > [2] https://releases.openstack.org/ > [3] https://devguide.python.org/#status-of-python-branches > I'd like to make a proposal for moving forward. But first let's summarize some facts so that we can evaluate the proposal against the state of the world. * Python 3.6 reaches EOL soon. * CentOS 9 Stream is released and makes no mention of being in a beta status (https://www.centos.org/centos-stream/) * We have the ability to run tests on CentOS 9 Stream today * It is useful to users to ensure we have an OpenStack release tested against old distro release with old python and new distro release with new python. This enables them to update underlying distro releases independently from upgrading OpenStack. * The various distros are not in sync over which python versions they support. Given this state it seems reasonable that we continue to test Yoga against python3.6 and newer python. It would also probably be a good idea to explicitly update the PTI and distro support policies to indicate that when adding new distro releases we continue to support the old distro release's python for one cycle. The level of actual testing done on these platforms will depend on who steps up to maintain that testing, but at the very least we'll accept bug fixes for old python and not prevent installation via setup.cfg python_requires settings. Yes, this means another cycle of python3.6 support which isn't ideal given it's soon EOL, but the intent is to support distros who will continue to support the older python on their platform. If we want to be bold and try an experiment, we could also reduce testing of the python versions to oldest and newest supported python versions. For example python3.6 and python3.9 during Yoga. Don't worry about python3.7 and python3.8 as it is likely they will continue to function as long as 3.6 and 3.9 are happy. This would reduce the size of the testing matrix. We can always reevaluate if problems for 3.7 and 3.8 occur. Finally, I think it is important to note that OpenStack is using CentOS as a stand in for RHEL. This worked a lot better when CentOS was an actual copy of RHEL. What we've got today is CentOS stream and it makes no claims about being in beta until RHEL releases (that I can find anyway), and I don't think we should make those assertions ourselves. If there is interest in using a different stand in (Alma or Rocky?) then we should replace CentOS stream with one of those alternatives. Another option may be to assert that Stream only functions as a standin once RHEL has released their corresponding update, but it is my understanding that Stream and RHEL will always have some delta. Clark From abraden at verisign.com Wed Dec 1 18:06:47 2021 From: abraden at verisign.com (Braden, Albert) Date: Wed, 1 Dec 2021 18:06:47 +0000 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: Obviously Redhat will continue to support and use Openstack on Centos, and Redhat employees are a substantial part of the Openstack community. Without the contributions from Redhat employees, Openstack would struggle to survive. However, I don?t think that use by a single company is enough to justify the effort required to support a distro, even when that single company is a major contributor. I hope and believe that Redhat would continue their contributions to the Openstack community if we continued to support Openstack on RHEL and dropped support for Centos. I'm curious whether anyone who does not work for Redhat is planning to continue using Openstack on Centos long-term, and if so, I'd like to hear your reasoning. -----Original Message----- From: Alex Schultz Sent: Wednesday, December 1, 2021 11:13 AM To: Braden, Albert Cc: openstack-discuss at lists.openstack.org Subject: [EXTERNAL] Re: Re: python_requires >= 3.8 during Yoga Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. On Wed, Dec 1, 2021 at 9:04 AM Braden, Albert wrote: > > It appears that Centos is no longer a viable platform for production Openstack clusters. Of course we have to continue supporting those who are not yet able to move to another distro, but I think we should be clear-eyed about the fact that these are stopgap measures that only need to be implemented while people work on moving. > I'm not certain why this is how it's being interpreted. CentOS stream is perfectly viable but it is (as has always been) reliant on packaging. I believe Cloud SIG will still exist and RDO will be publishing packages. In fact CentOS stream is generally better than the previous release cadence that caused all sort of issues when point releases dropped with breaking changes. The difference in stream vs legacy method was the volume of changes over time. Stream you get smaller changes sooner (along with fixes) where as the legacy method you'd only really get updates at point release time and it was always terrible troubleshooting these failures. CentOS Stream 9 will have 3.9 as the default but we need to make sure that we can get it available to the various projects for testing. Right now Puppet OpenStack primarily relies on CentOS for testing (because Ubuntu jobs have been broken for a few years now with no one jumping in to fix). So bumping the minimum version without having a way to deploy for us would directly impact this project. IMHO the issue with this thread is timing on being able to bump a minimum python version requirement. I personally wouldn't want to bump to something that excludes 3.6 at this time without a really good reason. Are there features we require that will suddenly make developer/operator lives easier (e.g. switching to asyncio or something)? If not, what is the driving factor? I think pushing it off to Z would be best to allow for the initial standup required to have *all* the projects able to support the switch. > Is anyone contemplating continuing to use Centos long-term? If so, I would be interested to hear your reasoning. > > -----Original Message----- > From: Zane Bitter > Sent: Wednesday, December 1, 2021 10:13 AM > To: openstack-discuss at lists.openstack.org > Subject: [EXTERNAL] Re: python_requires >= 3.8 during Yoga > > Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. > > On 26/11/21 10:29, Ghanshyam Mann wrote: > > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur wrote ---- > > > > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley wrote: > > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > > [...] > > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > > [...] > > > > > > Is this still true for CentOS Stream 9? The TC decision was to > > > support that instead of CentOS Stream 8 in Yoga. > > > > > > No. But Stream 9 is pretty much beta, so it's not a replacement for us (and we don't have nodes in nodepool with it even yet?). > > > > I think here is the confusion. In TC, after checking with centos team impression was CentOS stream 9 is released and that is > > what we should update In OpenStack testing. And then only we updated the centos stream 8 -> 9 and dropped py3.6 testing > > > > - https://secure-web.cisco.com/1l5YAwIy_Kf9i8nZRj01Av73trnFQMqoxRgz_2n5WHVL6dc2mfcz-we8VRFvRToxc9yvnpH8QhTDlo2oJoiPHPBheqzFvIaRh40w5Ib3WUxBlvdAfSSNFxyJmXgPOxrq_AwWW27UvaTeFD_ycxhRyngSr_hY7Hji2WkdMMsFl19QfRhk20MI9giWNQk6uMAKlsLKRl4Zuod2cfgERb8Fwm5qwZkfA8NkOU9gZr4IF-nbSgg5aLbgowl4imparhsKS/https%3A%2F%2Freview.opendev.org%2Fc%2Fopenstack%2Fgovernance%2F%2B%2F815851%2F3..6%2Freference%2Fruntimes%2Fyoga.rst > > The guidelines the TC have set are not that something exists, but that > it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, > and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable > LTS releases. It's not clear to me why CentOS Stream is the only distro > being treated differently. > > The only difference is there is no plan for a CentOS-branded release to > define a point in time where Stream 9 becomes an LTS release. However, > other parties do have such plans, but pointedly have not done so: RHEL9 > is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to > release a version based on Stream 9. > > Presumably RDO folks were consulted about this decision and were OK with > the time frame. However, there are other users out there, and from a > Metal? perspective this is a giant PITA, requiring us to move from a LTS > distro to a beta one, that was dropped on us in the middle of a release > cycle in flagrant violation of the TC's own guidelines that the stable > distibutions must be chosen from among those available at the > *beginning* of the release cycle (which CentOS Stream 9 was not). > > cheers, > Zane. > > From abraden at verisign.com Wed Dec 1 18:15:41 2021 From: abraden at verisign.com (Braden, Albert) Date: Wed, 1 Dec 2021 18:15:41 +0000 Subject: [ops] [kolla] RabbitMQ High Availability Message-ID: <7d613f7f8f71497db28b5e27ddcb8224@verisign.com> I read this with great interest because we are seeing this issue. Questions: 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? 2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? 3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? Does anyone have a sample set of RMQ config files that they can share? It looks like my Outlook has ruined the link; reposting: [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit -----Original Message----- From: Arnaud Morin Sent: Monday, November 29, 2021 2:04 PM To: Bogdan Dobrelya Cc: DHilsbos at performair.com; openstack-discuss at lists.openstack.org Subject: [EXTERNAL] Re: [ops]RabbitMQ High Availability Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. Hi, After a talk on this ml (which starts at [1]), we endup building a documentation with Large Scale group. The doc is accessible at [2]. Hope this will help. [1] http://secure-web.cisco.com/1gFccuTyEVGnFd9aBOZ-RTPG0hbVIPGAbuLBNnoXP4onSZGFG1umIn0EtkEpBJWko4mi6yOUZ8Vsm-5sDGmIVl8rC2sOHv3Z2I1s9lFIkVFyn16CXJgcJbQQ7SBU8wEz5I_TysLtIY6YrmiC3PkKdG4oVCZk6n_KqYPYjmYUmDn9BD6JcXKUbFujVfugbjewZDY4HDCBnTe43tPSqkIZRVarApPiwsFtHu5PQ5riSoSgTpupqZHZdPnnGz7sbVGzx/http%3A%2F%2Flists.openstack.org%2Fpipermail%2Fopenstack-discuss%2F2020-August%2F016362.html [2] https://secure-web.cisco.com/1OtQ3pcnPPBNwevAFxS8yOS2xFlkHo0tY4SmkFE-wpAU_YPYS-BxRX5omcjCPZ3cMOxefnaO0vc3qlVm_SvI3DpkhejUkQUrrRbBJ72ki_ly13bYzC_QKd0-VERmSnlx8SFUB_DWewMYIZ7JfaURBYN9QvJgwD0b0aG-hYgvxcN1ZCt7qHTDqneGTtpx-5gRUMvld2dFz5uXsPj7QzohumP5bAoTblw7xLJy3zXhlfvrg6aHhQIR4xw9_y8E5Lt7d/https%3A%2F%2Fwiki.openstack.org%2Fwiki%2FLarge_Scale_Configuration_Rabbit On 24.11.21 - 11:31, Bogdan Dobrelya wrote: > On 11/24/21 12:34 AM, DHilsbos at performair.com wrote: > > All; > > > > In the time I've been part of this mailing list, the subject of RabbitMQ high availability has come up several times, and each time specific recommendations for both Rabbit and Open Stack are provided. I remember it being an A or B kind of recommendation (i.e. configure Rabbit like A1, and Open Stack like A2, OR configure Rabbit like B1, and Open Stack like B2). > > There is no special recommendations for rabbitmq setup for openstack, > but probably a few, like instead of putting it behind a haproxy, or the > like, list the rabbit cluster nodes in the oslo messaging config > settings directly. Also, it seems that durable queues makes a very > little sense for highly ephemeral RPC calls, just by design. I would > also add that the raft quorum queues feature of rabbitmq >=3.18 does > neither fit well into the oslo messaging design for RPC calls. > > A discussable and highly opinionated thing is also configuring > ha/mirror queue policy params for queues used for RPC calls vs > broad-casted notifications. > > And my biased personal humble recommendation is: use the upstream OCF RA > [0][1], if configuring rabbitmq cluster by pacemaker. > > [0] https://secure-web.cisco.com/1N1wD9gW7NZho0LdTVNuiU2ZIB7NW-eJMfDgVzBH3D3E6URzGYPKa-uhcLHxy3tRvRXopjnLAd2CECD1urJyRpg8NBSxTOEUSPxOlS0cQyULtSQuDbVWr-W7Bl3ZRcdWPrF9EuX_b40IM7zTjqS40gImsEouTqtD1vlCuEoaFgpptDEuMuaNTqBJ0IAtiZHuWiW6E7ufTtgxmVbkGLjXCZw5ZNhibbu-kGVyA-7MQsxQ-RBgSq5peTcLBR2Vx-f9k/https%3A%2F%2Fwww.rabbitmq.com%2Fpacemaker.html%23auto-pacemaker > > [1] > https://secure-web.cisco.com/1iDK1NnL9JTkQqkpBda06xTQNrWY2W0pVOTDwUoadfQbSXn5r0g_GH8PB8wZC5-JmHW2-m1YWoj1Z86jFcmWT0m9W9Sax5fJE5G7MbvQN2JM0EbAVHJDCmiBkMZlrSLoTgmh30RGhvmF9ww7jAjVnas3_AYFmwc65P-YtpdcswFC8rYcg5HlE2d979gf2OQUeftP3lfClkVou7hnELIFanDq07MfOJc2exHIfBo2ZQyUXRqXWUqnTsj7df-jCySkz/https%3A%2F%2Fgithub.com%2FClusterLabs%2Fresource-agents%2Fblob%2Fmaster%2Fheartbeat%2Frabbitmq-server-ha > > > > > Unfortunately, I can't find the previous threads on this topic. > > > > Does anyone have this information, that they would care to share with me? > > > > Thank you, > > > > Dominic L. Hilsbos, MBA > > Vice President - Information Technology > > Perform Air International Inc. > > DHilsbos at PerformAir.com > > www.PerformAir.com > > > > > > > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando > > From fungi at yuggoth.org Wed Dec 1 18:24:31 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Wed, 1 Dec 2021 18:24:31 +0000 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: <20211201182430.5eh3wc6ezk5ittnm@yuggoth.org> On 2021-12-01 10:31:58 -0700 (-0700), Alex Schultz wrote: [...] > So I guess the question is who do we use for language support EOLs? > My understanding is that Ubuntu 18.04's default was 3.6 which has > standard support until 2023[0] and CentOS Stream 8 will be supported > until 2024[1]. Both of which should exceed Yoga's support[2]? With > all these arguments about LTS distros, it seems that we should be > using their support life cyles for the language versions and not > python's[3]. [...] The answer is that it's not clear. We use the packaged Python 3 interpreter and standard library from specific LTS distros, but we consume additional Python 3 library dependencies from PyPI rather than from packages in those distros contemporary with the interpreters they've packaged. The result is that we're stuck relying on when those dependencies decide to drop support for what they consider to be "old" interpreter versions. At one time we tried to rely on distro packaged dependencies, but discovered that breaks down as soon as projects want to use libraries which aren't yet packaged in those distributions. It's sort of Catch-22 in that the distros look to us to determine when they should be adding or updating dependencies, so if we look to them for which dependencies/versions are available then we can quickly get deadlocked. -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From fungi at yuggoth.org Wed Dec 1 18:34:17 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Wed, 1 Dec 2021 18:34:17 +0000 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> Message-ID: <20211201183416.bukigb2w4edsnftv@yuggoth.org> On 2021-12-01 18:06:47 +0000 (+0000), Braden, Albert wrote: [...] > I hope and believe that Redhat would continue their contributions > to the Openstack community if we continued to support Openstack on > RHEL and dropped support for Centos. [...] Just to clear up some misconceptions, what we mean when we say "support" (from the OpenStack community perspective) is that we test it upstream: https://governance.openstack.org/tc/resolutions/20170620-volunteer-support.html In that sense, OpenStack does not officially "support" RHEL because we lack the ability to test it due to licensing challenges. We've asked on multiple occasions, and been told that running RHEL in arbitrary public cloud environments driven by automation is not feasible to license (due to a combination of legalities and logistics), and is still outside the scope of the "free for developer use" RHEL license program available today. This is why we traditionally tested on CentOS as a stand-in for RHEL, rather than using actual RHEL. Informally we "support" RHEL in the sense that we'd like users to be able to run our software on the platform, but we rely on others to test that's possible since we can't guarantee it ourselves. -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From zbitter at redhat.com Wed Dec 1 19:21:01 2021 From: zbitter at redhat.com (Zane Bitter) Date: Wed, 1 Dec 2021 14:21:01 -0500 Subject: python_requires >= 3.8 during Yoga In-Reply-To: <17d76d33dc7.113f517f7180736.3792235396720560154@ghanshyammann.com> References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <17d76d33dc7.113f517f7180736.3792235396720560154@ghanshyammann.com> Message-ID: On 1/12/21 11:28, Ghanshyam Mann wrote: > ---- On Wed, 01 Dec 2021 09:13:02 -0600 Zane Bitter wrote ---- > > On 26/11/21 10:29, Ghanshyam Mann wrote: > > > ---- On Fri, 26 Nov 2021 09:20:39 -0600 Dmitry Tantsur wrote ---- > > > > > > > > > > > > On Fri, Nov 26, 2021 at 3:35 PM Jeremy Stanley wrote: > > > > On 2021-11-26 14:29:53 +0100 (+0100), Dmitry Tantsur wrote: > > > > [...] > > > > > CentOS/RHEL ships 3.6 and a limited version of 3.8 and 3.9. > > > > [...] > > > > > > > > Is this still true for CentOS Stream 9? The TC decision was to > > > > support that instead of CentOS Stream 8 in Yoga. > > > > > > > > No. But Stream 9 is pretty much beta, so it's not a replacement for us (and we don't have nodes in nodepool with it even yet?). > > > > > > I think here is the confusion. In TC, after checking with centos team impression was CentOS stream 9 is released and that is > > > what we should update In OpenStack testing. And then only we updated the centos stream 8 -> 9 and dropped py3.6 testing > > > > > > - https://review.opendev.org/c/openstack/governance/+/815851/3..6/reference/runtimes/yoga.rst > > > > The guidelines the TC have set are not that something exists, but that > > it is a stable LTS release. Debian sid, Ubuntu 20.10, Fedora rawhide, > > and OpenSUSE Tumbleweed all exist, but nobody mistakes them for stable > > LTS releases. It's not clear to me why CentOS Stream is the only distro > > being treated differently. > > > > The only difference is there is no plan for a CentOS-branded release to > > define a point in time where Stream 9 becomes an LTS release. However, > > other parties do have such plans, but pointedly have not done so: RHEL9 > > is in beta; Rocky Linux, Alma Linux, and Oracle Linux are all yet to > > release a version based on Stream 9. > > > > Presumably RDO folks were consulted about this decision and were OK with > > the time frame. However, there are other users out there, and from a > > Metal? perspective this is a giant PITA, requiring us to move from a LTS > > distro to a beta one, that was dropped on us in the middle of a release > > cycle in flagrant violation of the TC's own guidelines that the stable > > distibutions must be chosen from among those available at the > > *beginning* of the release cycle (which CentOS Stream 9 was not). > > Those are good points. We were discussing with the RDO team on the start > of Yoga cycle itself whether to move to CentOS Stream9 or not. As of today CentOS haven't announced the launch yet[1]. AIUI, it wasn't even possible for folks to start building packages against it before October 20.[2] The Yoga cycle had already started by at least October 11. So unless we're actively looking for ways to bend the rules to make users' lives harder, I think this is an easy call for Y. It's great if the RDO team are happy, but what I'm pointing out is that RDO is not the only consumer of OpenStack components in the EL ecosystem. > Yes, there is a clear confusion on what is stable or not in CentOS Stream distro. CentOS stream is the upstream for RHEL. That means that once RHEL is released, any patches to it must follow Red Hat's backward compatibility policies for RHEL. What policies apply to CentOS Stream commits when RHEL is still in beta? TBH I don't know. But I'd expect them to be... you know... beta policies. A beta is the opposite of stable long-term support. > Like other distro has, is there any document we as TC can refer to in the future on > checking if the CentOS Stream *X* version is stable/LTS or not? I cannot find it > for CentOS Stream 8 also, is that stable or LTS? When RHEL 9 is released you will definitely hear about it. If you don't want to tie it to a commercial distro, wait for Rocky Linux 9 (which is essentially equivalent to the old CentOS) to be released. I expect that will be later, not sooner, than RHEL though. > If we have such official documentation or announcement in RDO community > then we can avoid such situations in future where we get to know the distro > version stability well before trying it in OpenStack testing? I would have thought that waiting for a blog post from CentOS would be the absolute minimum, even if you don't agree that we should wait for *somebody* to build a long-term supported release from it before calling it an LTS release. cheers, Zane. [1] https://lists.centos.org/pipermail/centos-promo/2021-December/003599.html [2] https://lists.centos.org/pipermail/centos-devel/2021-October/077383.html From aschultz at redhat.com Wed Dec 1 19:46:08 2021 From: aschultz at redhat.com (Alex Schultz) Date: Wed, 1 Dec 2021 12:46:08 -0700 Subject: python_requires >= 3.8 during Yoga In-Reply-To: <50abb4f5-af13-435c-87c9-1c14d74dcfb2@www.fastmail.com> References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <50abb4f5-af13-435c-87c9-1c14d74dcfb2@www.fastmail.com> Message-ID: ))) On Wed, Dec 1, 2021 at 11:09 AM Clark Boylan wrote: > > On Wed, Dec 1, 2021, at 9:31 AM, Alex Schultz wrote: > > On Wed, Dec 1, 2021 at 10:05 AM Sean Mooney wrote: > > > > Apologies if my reduction of thread content was too aggressive. > > > > >> so haveing at least 3.9 support in yoga will be imporatn for multiple distros and to help use test that i and proably others > >> suggested raising the minium python version to 3.8 so that we only have to test with 2 version in the gate instead of 3.(3.6 3.8 and 3.9) > >> > >> continuting to support py36 aslo puts a burden on developers not just on our ci infrastucre as it will becomre harder to test wtih going froward > >> as it stops being shipped in distros and as it exits security support > >> the end of life for python 3.6 is planned for this month https://www.python.org/dev/peps/pep-0494/#lifespan > >> speficically (23 Dec 2021) based on https://endoflife.date/python > >> > >> i dont think we shoudl be releasing yoga on an end of life python version. > >> if we raise the minium to 3.8 that will be supported until at least 14 Oct 2024 > >> which is after the stabel branch will be retired. > >> > > > > So I guess the question is who do we use for language support EOLs? > > My understanding is that Ubuntu 18.04's default was 3.6 which has > > standard support until 2023[0] and CentOS Stream 8 will be supported > > until 2024[1]. Both of which should exceed Yoga's support[2]? With > > all these arguments about LTS distros, it seems that we should be > > using their support life cyles for the language versions and not > > python's[3]. > > > > [0] https://wiki.ubuntu.com/Releases > > [1] https://www.centos.org/cl-vs-cs/ > > [2] https://releases.openstack.org/ > > [3] https://devguide.python.org/#status-of-python-branches > > For this reply, "LTS Distro" means support/updates > 13 months This means regular Ubuntu and Fedora releases don't qualify. Ubuntu LTS, CentOS Stream * and Debian LTS would qualify. Can we all agree on this definition? https://ubuntu.com/about/release-cycle https://www.centos.org/cl-vs-cs/#end-of-life https://wiki.debian.org/LTS > > I'd like to make a proposal for moving forward. But first let's summarize some facts so that we can evaluate the proposal against the state of the world. > > * Python 3.6 reaches EOL soon. > * CentOS 9 Stream is released and makes no mention of being in a beta status (https://www.centos.org/centos-stream/) > * We have the ability to run tests on CentOS 9 Stream today Yes some OpenStack projects can run on CentOS 9 Stream today but not all can (Puppet OpenStack being one). The assumption that pip install is sufficient for *all* projects is not correct. At least some testing needs to be able to be run on all projects before projects can officially declare <= 3.6 dead. This hopefully will be completed during Yoga but it takes time. > * It is useful to users to ensure we have an OpenStack release tested against old distro release with old python and new distro release with new python. This enables them to update underlying distro releases independently from upgrading OpenStack. > * The various distros are not in sync over which python versions they support. > > Given this state it seems reasonable that we continue to test Yoga against python3.6 and newer python. It would also probably be a good idea to explicitly update the PTI and distro support policies to indicate that when adding new distro releases we continue to support the old distro release's python for one cycle. The level of actual testing done on these platforms will depend on who steps up to maintain that testing, but at the very least we'll accept bug fixes for old python and not prevent installation via setup.cfg python_requires settings. > > Yes, this means another cycle of python3.6 support which isn't ideal given it's soon EOL, but the intent is to support distros who will continue to support the older python on their platform. IMHO "another cycle" means continued support in Z not Yoga. I think the ask here was just delay the minimum version until Z so this would be the last cycle with 3.6 support. Jeremy said in a subsequent message that the interpreter comes from an LTS distro which IMHO means the EOL is the distro's EOL not Python's. Now this brings up issues if we can't get 3.6 fixes for specific versions of our dependencies that come in via pypi, but do we have data how often this is actually a problem? Isn't that already a problem for stable branches? Or are we worrying about something that doesn't happen? We need to all agree on where EOL dates come from for dependencies and expectations around them. LTS Distros are being referenced in this thread but people keep going back to EOL dates that are not aligned with these distros. For the full OpenStack project ecosystem, testing is occurring on Ubuntu, CentOS, and a bit of Debian. IMHO when a release (e.g. Yoga) is released, we should use what is available at the start of cycle and using the EOL dates of the tested distros in determining decisions around version support. If we use this for Yoga, realistically CentOS Stream 8 and Ubuntu 20.04 could be used which would leave us with 3.6/3.8. It seems Debian was the only one with 3.7 (https://wiki.debian.org/Python). Since CentOS Stream 9 was released during the Yoga cycle, we couldn't pick it up as a version until Z but we could then make the 3.8 cutoff then. If Ubuntu releases a new version (the website says they might) during Yoga with 3.9 then we could talk about raising it further in Z. This would allow for lower bounds to be pushed, but requires that one of our qualifying LTS distros in check/gate were already available at the start of a cycle. What I would like is that people stop pointing to Python's EOL date which is newer but does not actually align with the assumption of an LTS distro. It would only make sense to use that date if we were basing our testing on non-LTS distros which may be able to follow their EOL dates more closely. I keep hammering on the LTS distro because OpenStack is heavily reliant on packages and hardware supported by distros that it is very closely tied to what is available. If OpenStack had no tie to distro hardware support, venvs and pip installing all the things might be sufficient but there are many dependencies which end up not being satisfied this way. You can get away with it up to a point, but it's an important integration point. > > If we want to be bold and try an experiment, we could also reduce testing of the python versions to oldest and newest supported python versions. For example python3.6 and python3.9 during Yoga. Don't worry about python3.7 and python3.8 as it is likely they will continue to function as long as 3.6 and 3.9 are happy. This would reduce the size of the testing matrix. We can always reevaluate if problems for 3.7 and 3.8 occur. > The issue here I think is trying to align on a default LTS distro python. CS8,18.04 is 3.6, Debian Buster is 3.7, 20.04 is 3.8 and CS9 will be 3.9. Personally I think 3.6/3.9 is sufficient because for the projects I contribute to this covers my cases. But this would exclude the 3.8 on 20.04. Would this be a problem for Ubuntu's UCA? > Finally, I think it is important to note that OpenStack is using CentOS as a stand in for RHEL. This worked a lot better when CentOS was an actual copy of RHEL. What we've got today is CentOS stream and it makes no claims about being in beta until RHEL releases (that I can find anyway), and I don't think we should make those assertions ourselves. If there is interest in using a different stand in (Alma or Rocky?) then we should replace CentOS stream with one of those alternatives. Another option may be to assert that Stream only functions as a standin once RHEL has released their corresponding update, but it is my understanding that Stream and RHEL will always have some delta. There has always been a delta. The change now is the delta has been reversed such that we'll get fixes for things before RHEL but this is likely better for OpenStack's use cases where the previous meant long delays to get necessary fixes. I'm uncertain why the RHEL release schedule is being factored in to determine some indication of stability when testing is occurring (https://www.centos.org/cl-vs-cs/#testing). Thanks, -Alex > > Clark > From arnaud.morin at gmail.com Wed Dec 1 20:02:02 2021 From: arnaud.morin at gmail.com (Arnaud) Date: Wed, 01 Dec 2021 21:02:02 +0100 Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: <7d613f7f8f71497db28b5e27ddcb8224@verisign.com> References: <7d613f7f8f71497db28b5e27ddcb8224@verisign.com> Message-ID: <8C25B4D7-5D74-4E15-84C8-DBDA9E304855@gmail.com> Hi, I definitely recommend upgrading to 3.8. Also enable the durable queues. This helped us a lot in managing our clusters. We also applied the policy, which was originally taken from openstack ansible [1]. We also collect unroutable messages and send alerts based in that (but we had no issue recently thanks to all above). Finally we ping all our clients using oslo_ping_endpoint [2] every five minutes so we know when an agent is disconnected. Dunno about kolla, sorry. [1] https://github.com/openstack/openstack-ansible-rabbitmq_server/blob/fc27e735a68b64cb3c67dd8abeaf324803a9845b/defaults/main.yml#L172 [2] https://opendev.org/openstack/oslo.messaging/commit/82492442f3387a0e4f19623ccfda64f8b84d59c3 Le 1 d?cembre 2021 19:15:41 GMT+01:00, "Braden, Albert" a ?crit?: >I read this with great interest because we are seeing this issue. Questions: > >1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? >2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? >3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? > >Does anyone have a sample set of RMQ config files that they can share? > >It looks like my Outlook has ruined the link; reposting: >[2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > >-----Original Message----- >From: Arnaud Morin >Sent: Monday, November 29, 2021 2:04 PM >To: Bogdan Dobrelya >Cc: DHilsbos at performair.com; openstack-discuss at lists.openstack.org >Subject: [EXTERNAL] Re: [ops]RabbitMQ High Availability > >Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. > >Hi, > >After a talk on this ml (which starts at [1]), we endup building a >documentation with Large Scale group. >The doc is accessible at [2]. > >Hope this will help. > >[1] http://secure-web.cisco.com/1gFccuTyEVGnFd9aBOZ-RTPG0hbVIPGAbuLBNnoXP4onSZGFG1umIn0EtkEpBJWko4mi6yOUZ8Vsm-5sDGmIVl8rC2sOHv3Z2I1s9lFIkVFyn16CXJgcJbQQ7SBU8wEz5I_TysLtIY6YrmiC3PkKdG4oVCZk6n_KqYPYjmYUmDn9BD6JcXKUbFujVfugbjewZDY4HDCBnTe43tPSqkIZRVarApPiwsFtHu5PQ5riSoSgTpupqZHZdPnnGz7sbVGzx/http%3A%2F%2Flists.openstack.org%2Fpipermail%2Fopenstack-discuss%2F2020-August%2F016362.html >[2] https://secure-web.cisco.com/1OtQ3pcnPPBNwevAFxS8yOS2xFlkHo0tY4SmkFE-wpAU_YPYS-BxRX5omcjCPZ3cMOxefnaO0vc3qlVm_SvI3DpkhejUkQUrrRbBJ72ki_ly13bYzC_QKd0-VERmSnlx8SFUB_DWewMYIZ7JfaURBYN9QvJgwD0b0aG-hYgvxcN1ZCt7qHTDqneGTtpx-5gRUMvld2dFz5uXsPj7QzohumP5bAoTblw7xLJy3zXhlfvrg6aHhQIR4xw9_y8E5Lt7d/https%3A%2F%2Fwiki.openstack.org%2Fwiki%2FLarge_Scale_Configuration_Rabbit > >On 24.11.21 - 11:31, Bogdan Dobrelya wrote: >> On 11/24/21 12:34 AM, DHilsbos at performair.com wrote: >> > All; >> > >> > In the time I've been part of this mailing list, the subject of RabbitMQ high availability has come up several times, and each time specific recommendations for both Rabbit and Open Stack are provided. I remember it being an A or B kind of recommendation (i.e. configure Rabbit like A1, and Open Stack like A2, OR configure Rabbit like B1, and Open Stack like B2). >> >> There is no special recommendations for rabbitmq setup for openstack, >> but probably a few, like instead of putting it behind a haproxy, or the >> like, list the rabbit cluster nodes in the oslo messaging config >> settings directly. Also, it seems that durable queues makes a very >> little sense for highly ephemeral RPC calls, just by design. I would >> also add that the raft quorum queues feature of rabbitmq >=3.18 does >> neither fit well into the oslo messaging design for RPC calls. >> >> A discussable and highly opinionated thing is also configuring >> ha/mirror queue policy params for queues used for RPC calls vs >> broad-casted notifications. >> >> And my biased personal humble recommendation is: use the upstream OCF RA >> [0][1], if configuring rabbitmq cluster by pacemaker. >> >> [0] https://secure-web.cisco.com/1N1wD9gW7NZho0LdTVNuiU2ZIB7NW-eJMfDgVzBH3D3E6URzGYPKa-uhcLHxy3tRvRXopjnLAd2CECD1urJyRpg8NBSxTOEUSPxOlS0cQyULtSQuDbVWr-W7Bl3ZRcdWPrF9EuX_b40IM7zTjqS40gImsEouTqtD1vlCuEoaFgpptDEuMuaNTqBJ0IAtiZHuWiW6E7ufTtgxmVbkGLjXCZw5ZNhibbu-kGVyA-7MQsxQ-RBgSq5peTcLBR2Vx-f9k/https%3A%2F%2Fwww.rabbitmq.com%2Fpacemaker.html%23auto-pacemaker >> >> [1] >> https://secure-web.cisco.com/1iDK1NnL9JTkQqkpBda06xTQNrWY2W0pVOTDwUoadfQbSXn5r0g_GH8PB8wZC5-JmHW2-m1YWoj1Z86jFcmWT0m9W9Sax5fJE5G7MbvQN2JM0EbAVHJDCmiBkMZlrSLoTgmh30RGhvmF9ww7jAjVnas3_AYFmwc65P-YtpdcswFC8rYcg5HlE2d979gf2OQUeftP3lfClkVou7hnELIFanDq07MfOJc2exHIfBo2ZQyUXRqXWUqnTsj7df-jCySkz/https%3A%2F%2Fgithub.com%2FClusterLabs%2Fresource-agents%2Fblob%2Fmaster%2Fheartbeat%2Frabbitmq-server-ha >> >> > >> > Unfortunately, I can't find the previous threads on this topic. >> > >> > Does anyone have this information, that they would care to share with me? >> > >> > Thank you, >> > >> > Dominic L. Hilsbos, MBA >> > Vice President - Information Technology >> > Perform Air International Inc. >> > DHilsbos at PerformAir.com >> > www.PerformAir.com >> > >> > >> > >> >> >> -- >> Best regards, >> Bogdan Dobrelya, >> Irc #bogdando >> >> > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From fungi at yuggoth.org Wed Dec 1 20:16:27 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Wed, 1 Dec 2021 20:16:27 +0000 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <50abb4f5-af13-435c-87c9-1c14d74dcfb2@www.fastmail.com> Message-ID: <20211201201627.gkp4ik6g2qm5rdtf@yuggoth.org> On 2021-12-01 12:46:08 -0700 (-0700), Alex Schultz wrote: [...] > For this reply, "LTS Distro" means support/updates > 13 months This > means regular Ubuntu and Fedora releases don't qualify. Ubuntu LTS, > CentOS Stream * and Debian LTS would qualify. Can we all agree on > this definition? > > https://ubuntu.com/about/release-cycle > https://www.centos.org/cl-vs-cs/#end-of-life > https://wiki.debian.org/LTS [...] Technically, Debian on its own would meet this definition, as releases come every ~2.5 years and are generally supported for ~6 months after the next release, making it roughly 3 years of normal support. "LTS" in Debian is something which extends beyond the normal security support for the distribution, but is irrelevant here given the already long support timeframe for Debian releases. > Jeremy said in a subsequent message that the interpreter comes > from an LTS distro which IMHO means the EOL is the distro's EOL > not Python's. Now this brings up issues if we can't get 3.6 fixes > for specific versions of our dependencies that come in via pypi, > but do we have data how often this is actually a problem? You can get some idea by counting the number of times "python_version" appears in global-requirements.txt or upper-constraints.txt in openstack/requirements by the end of a release cycle. > Isn't that already a problem for stable branches? Not really, since we freeze openstack/requirements when branching, in a crude attempt to mimic testing with versions contemporary with what distros are probably packaging alongside that OpenStack release. > Or are we worrying about something that doesn't happen? We need > to all agree on where EOL dates come from for dependencies and > expectations around them. It's definitely a thing which happens. We're running Python 3.6 jobs on OpenStack project master branches now while pinning to older versions of more than a dozen dependencies, and when mainline 3.6 EOL comes that will almost certainly rise sharply as our dependencies start to aggressively drop support for it. Having this happen during a cycle definitely leads to a wider disparity in versions of dependencies used for various jobs, compared to having it happen after we freeze requirements for the release. > LTS Distros are being referenced in this thread but people keep going > back to EOL dates that are not aligned with these distros. For the > full OpenStack project ecosystem, testing is occurring on Ubuntu, > CentOS, and a bit of Debian. IMHO when a release (e.g. Yoga) is > released, we should use what is available at the start of cycle and > using the EOL dates of the tested distros in determining decisions > around version support. If we use this for Yoga, realistically CentOS > Stream 8 and Ubuntu 20.04 could be used which would leave us with > 3.6/3.8. It seems Debian was the only one with 3.7 > (https://wiki.debian.org/Python). Since CentOS Stream 9 was released > during the Yoga cycle, we couldn't pick it up as a version until Z but > we could then make the 3.8 cutoff then. If Ubuntu releases a new > version (the website says they might) during Yoga with 3.9 then we > could talk about raising it further in Z. This would allow for lower > bounds to be pushed, but requires that one of our qualifying LTS > distros in check/gate were already available at the start of a cycle. Actually, Debian's last stable release (11 a.k.a. bullseye) was available approximately 2 months before we started on OpenStack Yoga and uses Python 3.9 as its default python3, so testing 3.6 and 3.9 would be the preferred lower and upper bounds there. > What I would like is that people stop pointing to Python's EOL date > which is newer but does not actually align with the assumption of an > LTS distro. It would only make sense to use that date if we were > basing our testing on non-LTS distros which may be able to follow > their EOL dates more closely. [...] The underlying reason it keeps coming up is that, during development, we don't attempt to match the versions of dependencies we test with to the versions those distributions are packaging alongside the Python 3 interpreter. As a result, when our dependencies begin dropping upstream support for Python releases is what impacts us, at least under our current testing model. -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From tonyliu0592 at hotmail.com Wed Dec 1 20:26:12 2021 From: tonyliu0592 at hotmail.com (Tony Liu) Date: Wed, 1 Dec 2021 20:26:12 +0000 Subject: [kolla] Plan to deprecate binary and unify on single distrubition In-Reply-To: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> References: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> Message-ID: For OpenStack, I've been using centos-binary containers on CentOS. It's been working very well. I don't have much concern on supporting single-distro based container on different distro hosts. That's actually one of the beauties of container, which self-contains all packages and dependencies. We used to have a product based on single-distro container and support different distro hosts. Just need to be careful when container needs any resource from the host, like kernel, mounted host filesystem, networking, privilege, etc. Regarding to source container, what's the purpose of it? Is it allow user to get the package based on some specific source that is not provided by any existing package repo? In that case, I'd assume user should always build their own source container. Then what's the purpose of providing pre-build source container? Also, are those non-core containers, like MariaDB, HAProxy, RabbitMQ, etc. all built from source code? Thanks! Tony ________________________________________ From: Micha? Nasiadka Sent: November 18, 2021 04:33 AM To: openstack-discuss Subject: [kolla] Plan to deprecate binary and unify on single distrubition Hello Koalas, On the PTG we have discussed two topics: 1) Deprecate and drop binary type of Kolla images 2) Use a common base (single Linux distribution) for Kolla images This is a call for feedback - for people that have not been attending the PTG. What this essentially mean for consumers: 1) In Yoga cycle we will deprecate binary type of Kolla images, and in Z cycle those will be dropped. 2) We are not going to support CentOS Stream 9 (cs9) as a base operating system, and the source type build will rely on CentOS Stream 8 in Z release. 3) Beginning from A release Kolla will build only Debian source images - but Kolla-Ansible will still support deployment of those images on CentOS/Ubuntu/Debian Host operating systems (and Rocky Linux to be added in Yoga to that mix). Justification: The Kolla project team is limited in numbers, therefore supporting current broad mix of operating systems (especially with CentOS Stream 9 ,,on the way??) is a significant maintenance burden. By dropping binary type of images - users would be running more tested images (since Kolla/Kolla-Ansible CI runs source images jobs as voting). In Xena we?ve already changed the default image type Kolla-Ansible uses to source. We also feel that using a unified base OS for Kolla container images is a way to remove some of the maintenance burden (including CI cycles and Request for feedback: If any of those changes is a no go from your perspective - we?d like to hear your opinions. Best regards, Michal Nasiadka From pierre at stackhpc.com Wed Dec 1 21:22:52 2021 From: pierre at stackhpc.com (Pierre Riteau) Date: Wed, 1 Dec 2021 22:22:52 +0100 Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: <7d613f7f8f71497db28b5e27ddcb8224@verisign.com> References: <7d613f7f8f71497db28b5e27ddcb8224@verisign.com> Message-ID: Hello Albert, The amqp_durable_queues configuration option comes from oslo.messaging, which is a library used by many OpenStack projects. You can set this option in the [oslo_messaging_rabbit] section of each of these OpenStack projects (check their configuration reference for more details). As for how to configure it with Kolla Ansible: you can either set it directly for each service (e.g. in /etc/kolla/config/nova.conf to configure it for Nova) or for all projects at once using /etc/kolla/config/global.conf. Projects that don't support this option should just ignore it. Read this section of the documentation for more details: https://docs.openstack.org/kolla-ansible/latest/admin/advanced-configuration.html#openstack-service-configuration-in-kolla Best wishes, Pierre Riteau On Wed, 1 Dec 2021 at 19:20, Braden, Albert wrote: > > I read this with great interest because we are seeing this issue. Questions: > > 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? > 2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? > 3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? > > Does anyone have a sample set of RMQ config files that they can share? > > It looks like my Outlook has ruined the link; reposting: > [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > -----Original Message----- > From: Arnaud Morin > Sent: Monday, November 29, 2021 2:04 PM > To: Bogdan Dobrelya > Cc: DHilsbos at performair.com; openstack-discuss at lists.openstack.org > Subject: [EXTERNAL] Re: [ops]RabbitMQ High Availability > > Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. > > Hi, > > After a talk on this ml (which starts at [1]), we endup building a > documentation with Large Scale group. > The doc is accessible at [2]. > > Hope this will help. > > [1] http://secure-web.cisco.com/1gFccuTyEVGnFd9aBOZ-RTPG0hbVIPGAbuLBNnoXP4onSZGFG1umIn0EtkEpBJWko4mi6yOUZ8Vsm-5sDGmIVl8rC2sOHv3Z2I1s9lFIkVFyn16CXJgcJbQQ7SBU8wEz5I_TysLtIY6YrmiC3PkKdG4oVCZk6n_KqYPYjmYUmDn9BD6JcXKUbFujVfugbjewZDY4HDCBnTe43tPSqkIZRVarApPiwsFtHu5PQ5riSoSgTpupqZHZdPnnGz7sbVGzx/http%3A%2F%2Flists.openstack.org%2Fpipermail%2Fopenstack-discuss%2F2020-August%2F016362.html > [2] https://secure-web.cisco.com/1OtQ3pcnPPBNwevAFxS8yOS2xFlkHo0tY4SmkFE-wpAU_YPYS-BxRX5omcjCPZ3cMOxefnaO0vc3qlVm_SvI3DpkhejUkQUrrRbBJ72ki_ly13bYzC_QKd0-VERmSnlx8SFUB_DWewMYIZ7JfaURBYN9QvJgwD0b0aG-hYgvxcN1ZCt7qHTDqneGTtpx-5gRUMvld2dFz5uXsPj7QzohumP5bAoTblw7xLJy3zXhlfvrg6aHhQIR4xw9_y8E5Lt7d/https%3A%2F%2Fwiki.openstack.org%2Fwiki%2FLarge_Scale_Configuration_Rabbit > > On 24.11.21 - 11:31, Bogdan Dobrelya wrote: > > On 11/24/21 12:34 AM, DHilsbos at performair.com wrote: > > > All; > > > > > > In the time I've been part of this mailing list, the subject of RabbitMQ high availability has come up several times, and each time specific recommendations for both Rabbit and Open Stack are provided. I remember it being an A or B kind of recommendation (i.e. configure Rabbit like A1, and Open Stack like A2, OR configure Rabbit like B1, and Open Stack like B2). > > > > There is no special recommendations for rabbitmq setup for openstack, > > but probably a few, like instead of putting it behind a haproxy, or the > > like, list the rabbit cluster nodes in the oslo messaging config > > settings directly. Also, it seems that durable queues makes a very > > little sense for highly ephemeral RPC calls, just by design. I would > > also add that the raft quorum queues feature of rabbitmq >=3.18 does > > neither fit well into the oslo messaging design for RPC calls. > > > > A discussable and highly opinionated thing is also configuring > > ha/mirror queue policy params for queues used for RPC calls vs > > broad-casted notifications. > > > > And my biased personal humble recommendation is: use the upstream OCF RA > > [0][1], if configuring rabbitmq cluster by pacemaker. > > > > [0] https://secure-web.cisco.com/1N1wD9gW7NZho0LdTVNuiU2ZIB7NW-eJMfDgVzBH3D3E6URzGYPKa-uhcLHxy3tRvRXopjnLAd2CECD1urJyRpg8NBSxTOEUSPxOlS0cQyULtSQuDbVWr-W7Bl3ZRcdWPrF9EuX_b40IM7zTjqS40gImsEouTqtD1vlCuEoaFgpptDEuMuaNTqBJ0IAtiZHuWiW6E7ufTtgxmVbkGLjXCZw5ZNhibbu-kGVyA-7MQsxQ-RBgSq5peTcLBR2Vx-f9k/https%3A%2F%2Fwww.rabbitmq.com%2Fpacemaker.html%23auto-pacemaker > > > > [1] > > https://secure-web.cisco.com/1iDK1NnL9JTkQqkpBda06xTQNrWY2W0pVOTDwUoadfQbSXn5r0g_GH8PB8wZC5-JmHW2-m1YWoj1Z86jFcmWT0m9W9Sax5fJE5G7MbvQN2JM0EbAVHJDCmiBkMZlrSLoTgmh30RGhvmF9ww7jAjVnas3_AYFmwc65P-YtpdcswFC8rYcg5HlE2d979gf2OQUeftP3lfClkVou7hnELIFanDq07MfOJc2exHIfBo2ZQyUXRqXWUqnTsj7df-jCySkz/https%3A%2F%2Fgithub.com%2FClusterLabs%2Fresource-agents%2Fblob%2Fmaster%2Fheartbeat%2Frabbitmq-server-ha > > > > > > > > Unfortunately, I can't find the previous threads on this topic. > > > > > > Does anyone have this information, that they would care to share with me? > > > > > > Thank you, > > > > > > Dominic L. Hilsbos, MBA > > > Vice President - Information Technology > > > Perform Air International Inc. > > > DHilsbos at PerformAir.com > > > www.PerformAir.com > > > > > > > > > > > > > > > -- > > Best regards, > > Bogdan Dobrelya, > > Irc #bogdando > > > > > > > From adrian at fleio.com Wed Dec 1 21:30:48 2021 From: adrian at fleio.com (Adrian Andreias) Date: Wed, 1 Dec 2021 23:30:48 +0200 Subject: [kolla] Plan to deprecate binary and unify on single distrubition In-Reply-To: References: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> Message-ID: The user doesn't have to build the images that are source based. Ready built source-based images are available: https://quay.io/search?q=kolla Binary-based Kolla docker images are installing (at image build time) OpenStack projects via system packages, while source based images are symply adding projects' Python source code. Just as you said, the user shouldn't care much about distro inside images. But makes things a lot easier for developers to maintain less images and without working around different versions and dependencies provided by the system packages of various distros. Regards, Adrian Andreias On Wed, Dec 1, 2021 at 10:31 PM Tony Liu wrote: > For OpenStack, I've been using centos-binary containers on CentOS. It's > been working very well. > > I don't have much concern on supporting single-distro based container on > different distro hosts. > That's actually one of the beauties of container, which self-contains all > packages and dependencies. > We used to have a product based on single-distro container and support > different distro hosts. > Just need to be careful when container needs any resource from the host, > like kernel, mounted > host filesystem, networking, privilege, etc. > > Regarding to source container, what's the purpose of it? Is it allow user > to get the package based > on some specific source that is not provided by any existing package repo? > In that case, I'd assume > user should always build their own source container. Then what's the > purpose of providing pre-build > source container? Also, are those non-core containers, like MariaDB, > HAProxy, RabbitMQ, etc. all > built from source code? > > > Thanks! > Tony > ________________________________________ > From: Micha? Nasiadka > Sent: November 18, 2021 04:33 AM > To: openstack-discuss > Subject: [kolla] Plan to deprecate binary and unify on single distrubition > > Hello Koalas, > > On the PTG we have discussed two topics: > > 1) Deprecate and drop binary type of Kolla images > 2) Use a common base (single Linux distribution) for Kolla images > > This is a call for feedback - for people that have not been attending the > PTG. > > What this essentially mean for consumers: > > 1) In Yoga cycle we will deprecate binary type of Kolla images, and in Z > cycle those will be dropped. > 2) We are not going to support CentOS Stream 9 (cs9) as a base operating > system, and the source type build will rely on CentOS Stream 8 in Z release. > 3) Beginning from A release Kolla will build only Debian source images - > but Kolla-Ansible will still support deployment of those images on > CentOS/Ubuntu/Debian Host operating systems (and Rocky Linux to be added in > Yoga to that mix). > > Justification: > The Kolla project team is limited in numbers, therefore supporting current > broad mix of operating systems (especially with CentOS Stream 9 ,,on the > way??) is a significant maintenance burden. > By dropping binary type of images - users would be running more tested > images (since Kolla/Kolla-Ansible CI runs source images jobs as voting). > In Xena we?ve already changed the default image type Kolla-Ansible uses to > source. > We also feel that using a unified base OS for Kolla container images is a > way to remove some of the maintenance burden (including CI cycles and > > Request for feedback: > If any of those changes is a no go from your perspective - we?d like to > hear your opinions. > > Best regards, > Michal Nasiadka > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From cboylan at sapwetik.org Wed Dec 1 21:30:44 2021 From: cboylan at sapwetik.org (Clark Boylan) Date: Wed, 01 Dec 2021 13:30:44 -0800 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <50abb4f5-af13-435c-87c9-1c14d74dcfb2@www.fastmail.com> Message-ID: On Wed, Dec 1, 2021, at 11:46 AM, Alex Schultz wrote: > ))) > > On Wed, Dec 1, 2021 at 11:09 AM Clark Boylan wrote: >> >> On Wed, Dec 1, 2021, at 9:31 AM, Alex Schultz wrote: >> > On Wed, Dec 1, 2021 at 10:05 AM Sean Mooney wrote: >> > >> >> Apologies if my reduction of thread content was too aggressive. >> >> > >> >> so haveing at least 3.9 support in yoga will be imporatn for multiple distros and to help use test that i and proably others >> >> suggested raising the minium python version to 3.8 so that we only have to test with 2 version in the gate instead of 3.(3.6 3.8 and 3.9) >> >> >> >> continuting to support py36 aslo puts a burden on developers not just on our ci infrastucre as it will becomre harder to test wtih going froward >> >> as it stops being shipped in distros and as it exits security support >> >> the end of life for python 3.6 is planned for this month https://www.python.org/dev/peps/pep-0494/#lifespan >> >> speficically (23 Dec 2021) based on https://endoflife.date/python >> >> >> >> i dont think we shoudl be releasing yoga on an end of life python version. >> >> if we raise the minium to 3.8 that will be supported until at least 14 Oct 2024 >> >> which is after the stabel branch will be retired. >> >> >> > >> > So I guess the question is who do we use for language support EOLs? >> > My understanding is that Ubuntu 18.04's default was 3.6 which has >> > standard support until 2023[0] and CentOS Stream 8 will be supported >> > until 2024[1]. Both of which should exceed Yoga's support[2]? With >> > all these arguments about LTS distros, it seems that we should be >> > using their support life cyles for the language versions and not >> > python's[3]. >> > >> > [0] https://wiki.ubuntu.com/Releases >> > [1] https://www.centos.org/cl-vs-cs/ >> > [2] https://releases.openstack.org/ >> > [3] https://devguide.python.org/#status-of-python-branches >> > > > For this reply, "LTS Distro" means support/updates > 13 months This > means regular Ubuntu and Fedora releases don't qualify. Ubuntu LTS, > CentOS Stream * and Debian LTS would qualify. Can we all agree on > this definition? > > https://ubuntu.com/about/release-cycle > https://www.centos.org/cl-vs-cs/#end-of-life > https://wiki.debian.org/LTS > >> >> I'd like to make a proposal for moving forward. But first let's summarize some facts so that we can evaluate the proposal against the state of the world. >> >> * Python 3.6 reaches EOL soon. >> * CentOS 9 Stream is released and makes no mention of being in a beta status (https://www.centos.org/centos-stream/) >> * We have the ability to run tests on CentOS 9 Stream today > > Yes some OpenStack projects can run on CentOS 9 Stream today but not > all can (Puppet OpenStack being one). The assumption that pip install > is sufficient for *all* projects is not correct. At least some testing > needs to be able to be run on all projects before projects can > officially declare <= 3.6 dead. This hopefully will be completed > during Yoga but it takes time. > >> * It is useful to users to ensure we have an OpenStack release tested against old distro release with old python and new distro release with new python. This enables them to update underlying distro releases independently from upgrading OpenStack. >> * The various distros are not in sync over which python versions they support. >> >> Given this state it seems reasonable that we continue to test Yoga against python3.6 and newer python. It would also probably be a good idea to explicitly update the PTI and distro support policies to indicate that when adding new distro releases we continue to support the old distro release's python for one cycle. The level of actual testing done on these platforms will depend on who steps up to maintain that testing, but at the very least we'll accept bug fixes for old python and not prevent installation via setup.cfg python_requires settings. >> >> Yes, this means another cycle of python3.6 support which isn't ideal given it's soon EOL, but the intent is to support distros who will continue to support the older python on their platform. > > IMHO "another cycle" means continued support in Z not Yoga. I think > the ask here was just delay the minimum version until Z so this would > be the last cycle with 3.6 support. Jeremy said in a subsequent > message that the interpreter comes from an LTS distro which IMHO means > the EOL is the distro's EOL not Python's. Now this brings up issues if > we can't get 3.6 fixes for specific versions of our dependencies that > come in via pypi, but do we have data how often this is actually a > problem? Isn't that already a problem for stable branches? Or are we > worrying about something that doesn't happen? We need to all agree on > where EOL dates come from for dependencies and expectations around > them. The point I was trying to make was more that we should be explicit about when the overlap happens; write that down in our policy so that the next time this comes up the process goes more smoothly. I suppose if we have to delay for an additional cycle that may be the cost of not having considered this ahead of time. I think Jeremy covered why the Python EOL is important. Our dependencies tend to start dropping like flies when python versions go EOL. I think we can survive for a time, but we shouldn't try to keep them alive for the long term. > > LTS Distros are being referenced in this thread but people keep going > back to EOL dates that are not aligned with these distros. For the > full OpenStack project ecosystem, testing is occurring on Ubuntu, > CentOS, and a bit of Debian. IMHO when a release (e.g. Yoga) is > released, we should use what is available at the start of cycle and > using the EOL dates of the tested distros in determining decisions > around version support. If we use this for Yoga, realistically CentOS > Stream 8 and Ubuntu 20.04 could be used which would leave us with > 3.6/3.8. It seems Debian was the only one with 3.7 > (https://wiki.debian.org/Python). Since CentOS Stream 9 was released > during the Yoga cycle, we couldn't pick it up as a version until Z but > we could then make the 3.8 cutoff then. If Ubuntu releases a new > version (the website says they might) during Yoga with 3.9 then we > could talk about raising it further in Z. This would allow for lower > bounds to be pushed, but requires that one of our qualifying LTS > distros in check/gate were already available at the start of a cycle. > > What I would like is that people stop pointing to Python's EOL date > which is newer but does not actually align with the assumption of an > LTS distro. It would only make sense to use that date if we were > basing our testing on non-LTS distros which may be able to follow > their EOL dates more closely. I keep hammering on the LTS distro > because OpenStack is heavily reliant on packages and hardware > supported by distros that it is very closely tied to what is > available. If OpenStack had no tie to distro hardware support, venvs > and pip installing all the things might be sufficient but there are > many dependencies which end up not being satisfied this way. You can > get away with it up to a point, but it's an important integration > point. > >> >> If we want to be bold and try an experiment, we could also reduce testing of the python versions to oldest and newest supported python versions. For example python3.6 and python3.9 during Yoga. Don't worry about python3.7 and python3.8 as it is likely they will continue to function as long as 3.6 and 3.9 are happy. This would reduce the size of the testing matrix. We can always reevaluate if problems for 3.7 and 3.8 occur. >> > > The issue here I think is trying to align on a default LTS distro > python. CS8,18.04 is 3.6, Debian Buster is 3.7, 20.04 is 3.8 and CS9 > will be 3.9. Personally I think 3.6/3.9 is sufficient because for the > projects I contribute to this covers my cases. But this would exclude > the 3.8 on 20.04. Would this be a problem for Ubuntu's UCA? Ubuntu 20.04 includes a python 3.9. I don't know how that would impact the UCA or Ubuntu's packaging though. That said I have a strong suspicion that if OpenStack runs on 3.6 and 3.9 that it should just work on 3.8 as well. And as I mentioned if we prove that wrong through experience we can always fix 3.8 and go back to explicitly testing it. I don't think this particular change is super critical, but I know some people were concerned about the number of pythons that would need testing. > >> Finally, I think it is important to note that OpenStack is using CentOS as a stand in for RHEL. This worked a lot better when CentOS was an actual copy of RHEL. What we've got today is CentOS stream and it makes no claims about being in beta until RHEL releases (that I can find anyway), and I don't think we should make those assertions ourselves. If there is interest in using a different stand in (Alma or Rocky?) then we should replace CentOS stream with one of those alternatives. Another option may be to assert that Stream only functions as a standin once RHEL has released their corresponding update, but it is my understanding that Stream and RHEL will always have some delta. > > There has always been a delta. The change now is the delta has been > reversed such that we'll get fixes for things before RHEL but this is > likely better for OpenStack's use cases where the previous meant long > delays to get necessary fixes. I'm uncertain why the RHEL release > schedule is being factored in to determine some indication of > stability when testing is occurring > (https://www.centos.org/cl-vs-cs/#testing). I'm explicitly saying don't compare to RHEL for stability. CentOS 9 Stream is released as far as I can tell. It isn't labeled a beta release. You can download and run it today. We run CentOS testing because OpenStack asserts that it supports RHEL and historically CentOS was the closest we could get to RHEL. I'm saying we might consider if continuing to use CentOS for that goal is still appropriate. If it is then lets just use 9 Stream today as it is released and ready to go. If it isn't appropriate because it isn't close enough to RHEL, then we should consider an alternative like Alma or Rocky or perhaps they will converge more once RHEL has released. Basically I'm saying lets get away from thinking about this in terms of "is stream ready/beta/etc" and think in terms of "does using stream accomplish the goals of testing to ensure RHEL compatibility". > > Thanks, > -Alex > >> >> Clark >> From tonyliu0592 at hotmail.com Wed Dec 1 22:04:19 2021 From: tonyliu0592 at hotmail.com (Tony Liu) Date: Wed, 1 Dec 2021 22:04:19 +0000 Subject: [kolla] Plan to deprecate binary and unify on single distrubition In-Reply-To: References: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> Message-ID: Since I've never used source container and not able to find answers from doc, bear my dummy questions. BTW, I am not able to figure it out from Dockerfile.j2. Take centos-source-mariadb as an example, is it built by 1) installing packages from package repo or 2) download source code, compile and build packages and install self-build packages? In case of 2), how can I tell where and which release source code is downloaded and how it's compiled and built? In case of 1), what's the difference here from centos-binary-mariadb? Similarly, take centos-source-keystone as another example, is it built by 1) installing package from eg. centos/8-stream/cloud/x86_64/openstack-ussuri/ or 2) download source code from Github, build packages locally and install them? The same questions here, for 1) what's the difference from binary container, for 2) how can I tell which revision the source code is downloaded? Or, the source container is binary container with source code? Thanks! Tony ________________________________________ Fro: Adrian Andreias Sent: December 1, 2021 01:30 PM To: Tony Liu Cc: Micha? Nasiadka; openstack-discuss Subject: Re: [kolla] Plan to deprecate binary and unify on single distrubition The user doesn't have to build the images that are source based. Ready built source-based images are available: https://quay.io/search?q=kolla Binary-based Kolla docker images are installing (at image build time) OpenStack projects via system packages, while source based images are symply adding projects' Python source code. Just as you said, the user shouldn't care much about distro inside images. But makes things a lot easier for developers to maintain less images and without working around different versions and dependencies provided by the system packages of various distros. Regards, Adrian Andreias On Wed, Dec 1, 2021 at 10:31 PM Tony Liu > wrote: For OpenStack, I've been using centos-binary containers on CentOS. It's been working very well. I don't have much concern on supporting single-distro based container on different distro hosts. That's actually one of the beauties of container, which self-contains all packages and dependencies. We used to have a product based on single-distro container and support different distro hosts. Just need to be careful when container needs any resource from the host, like kernel, mounted host filesystem, networking, privilege, etc. Regarding to source container, what's the purpose of it? Is it allow user to get the package based on some specific source that is not provided by any existing package repo? In that case, I'd assume user should always build their own source container. Then what's the purpose of providing pre-build source container? Also, are those non-core containers, like MariaDB, HAProxy, RabbitMQ, etc. all built from source code? Thanks! Tony ________________________________________ From: Micha? Nasiadka > Sent: November 18, 2021 04:33 AM To: openstack-discuss Subject: [kolla] Plan to deprecate binary and unify on single distrubition Hello Koalas, On the PTG we have discussed two topics: 1) Deprecate and drop binary type of Kolla images 2) Use a common base (single Linux distribution) for Kolla images This is a call for feedback - for people that have not been attending the PTG. What this essentially mean for consumers: 1) In Yoga cycle we will deprecate binary type of Kolla images, and in Z cycle those will be dropped. 2) We are not going to support CentOS Stream 9 (cs9) as a base operating system, and the source type build will rely on CentOS Stream 8 in Z release. 3) Beginning from A release Kolla will build only Debian source images - but Kolla-Ansible will still support deployment of those images on CentOS/Ubuntu/Debian Host operating systems (and Rocky Linux to be added in Yoga to that mix). Justification: The Kolla project team is limited in numbers, therefore supporting current broad mix of operating systems (especially with CentOS Stream 9 ,,on the way??) is a significant maintenance burden. By dropping binary type of images - users would be running more tested images (since Kolla/Kolla-Ansible CI runs source images jobs as voting). In Xena we?ve already changed the default image type Kolla-Ansible uses to source. We also feel that using a unified base OS for Kolla container images is a way to remove some of the maintenance burden (including CI cycles and Request for feedback: If any of those changes is a no go from your perspective - we?d like to hear your opinions. Best regards, Michal Nasiadka From gmann at ghanshyammann.com Wed Dec 1 22:29:35 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Wed, 01 Dec 2021 16:29:35 -0600 Subject: [all][tc] Technical Committee next weekly meeting on Dec 2nd at 1500 UTC In-Reply-To: <17d6cf1413a.11da1006039481.6604682679623549161@ghanshyammann.com> References: <17d6cf1413a.11da1006039481.6604682679623549161@ghanshyammann.com> Message-ID: <17d781e0eab.b5fe932f194678.5688946776937692361@ghanshyammann.com> Hello Everyone, Below is the agenda for Tomorrow's TC video meeting schedule at 1500 UTC. https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting == Agenda for tomorrow's TC meeting == * Roll call * Follow up on past action items * Gate health check ** Fixing Zuul config error in OpenStack *** https://etherpad.opendev.org/p/zuul-config-error-openstack * Updates on community-wide goal ** Selecting RBAC goal *** https://review.opendev.org/c/openstack/governance/+/818817 ** Proposed community goal for FIPS compatibility and compliance *** https://review.opendev.org/c/openstack/governance/+/816587 * Yoga testing runtime: revert the drop of py3.6? ** http://lists.openstack.org/pipermail/openstack-discuss/2021-November/026024.html * Adjutant need PTLs and maintainers ** http://lists.openstack.org/pipermail/openstack-discuss/2021-October/025555.html * TC position on release cadence? ** http://lists.openstack.org/pipermail/openstack-discuss/2021-November/025684.html * Open Reviews ** https://review.opendev.org/q/projects:openstack/governance+is:open -gmann ---- On Mon, 29 Nov 2021 12:24:49 -0600 Ghanshyam Mann wrote ---- > Hello Everyone, > > Technical Committee's next weekly meeting is scheduled for Dec 2nd at 1500 UTC. > > If you would like to add topics for discussion, please add them to the below wiki page by > Wednesday, Dec 1st, at 2100 UTC. > > https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting > > -gmann > > From zigo at debian.org Wed Dec 1 22:33:54 2021 From: zigo at debian.org (Thomas Goirand) Date: Wed, 1 Dec 2021 23:33:54 +0100 Subject: [nova] Spawning instance that will do PXE booting Message-ID: Hi, In order to test OCI [1] within OpenStack itself, I would need my instances to do PXE booting. This would be with something like this on the qemu command line: qemu-system-x86_64 -boot n I believe this would also mean setting-up network cards that can do PXE booting, like the e1000, which Qemu supports. If that is possible, then I would setup DHCP + tftp-hpa on another instance, wired on the same private network. Your thoughts? Is this possible? Cheers, Thomas Goirand (zigo) [1] https://salsa.debian.org/openstack-team/debian/openstack-cluster-installer From openstack at a.spamming.party Wed Dec 1 22:43:32 2021 From: openstack at a.spamming.party (Jean-Philippe Evrard) Date: Wed, 01 Dec 2021 23:43:32 +0100 Subject: [all][tc] Relmgt team position on release cadence In-Reply-To: References: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> <8301ef05-4011-43e9-beec-66f6aeeb00f9@www.fastmail.com> <20211129130957.3poysdeqsufkp3pg@yuggoth.org> <17d72ae5ee2.b80bfd38119150.5945079300853303270@ghanshyammann.com> <20211130220913.ydk5zhyvkdl7g7zc@yuggoth.org> Message-ID: <6536dfae-cb75-4263-856d-f4dbdfa2db01@www.fastmail.com> Hello, On Tue, Nov 30, 2021, at 23:31, Julia Kreger wrote: >> > It feels like this is a giant "no win situation", > It feels like we have created a bunch of basically immovable, > insurmountable conflicting obstacles. Kind of like a self digging > holes. I'm worried not even hacker-Kirk can save us. Well, maybe his > answer might actually be to abolish the integrated release so he can > not only rescue the operators on the ship, but also beam them the > tools they need to move forward. Granted, that is change, and human > nature is a thing. :( Well, I feel completely differently. For me, people are using different words, and are in agreement in some points. Or maybe I am reading this wrong? Here is what I read: 1) Many want more releases, not less. I haven't seen a complaint about tagging more releases. 2) More than one person is proposing to abandon the integrated release, and nobody has complained about it. 3) Many people seem eager to carry "stable branches" for "critical patches", but no new definition of such criticality was done. 4) Many people want to make sure it's easy to upgrade, and with less steps for operations. I don't see any conflicts, just areas for improvement, for those who have been participating on this topic. Can someone clarify if I have tunnel vision/bias (as it seems exactly what I proposed in my first answer)? Thank you in advance. Regards, Jean-Philippe Evrard (evrardjp) From adrian at fleio.com Wed Dec 1 23:04:26 2021 From: adrian at fleio.com (Adrian Andreias) Date: Thu, 2 Dec 2021 01:04:26 +0200 Subject: [kolla] Plan to deprecate binary and unify on single distrubition In-Reply-To: References: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> Message-ID: "source" refers here to OpenStack projects source code: https://opendev.org/openstack/kolla/src/commit/1a9d5a1a42c86b447af371ab5e0864f2b452f6ad/docker/nova/nova-base/Dockerfile.j2#L52 There is no reason to build MariaDB from source. System package are always used (regardless of binary/source config): https://opendev.org/openstack/kolla/src/commit/1a9d5a1a42c86b447af371ab5e0864f2b452f6ad/docker/mariadb/mariadb-base/Dockerfile.j2#L14 https://opendev.org/openstack/kolla/src/commit/1a9d5a1a42c86b447af371ab5e0864f2b452f6ad/docker/mariadb/mariadb-server/Dockerfile.j2#L15 Image names like "centos-source-mariadb" are probably just kept for consistency. If you want to experiment and see what's happening, you can build images yourself: https://docs.openstack.org/kolla/latest/admin/image-building.html Anyhow, the thread is about any blockers on Micha?'s proposal. I think it makes sense: less ground to cover and effort will be focused on a few reliable images. The only possible issue I can imagine would be a project that already has massive image customizations that are based on binary images. Otherwise, minor customizations or standard images should not impact change from binary to source nor to single distro. Regards, Adrian Andreias On Thu, Dec 2, 2021 at 12:04 AM Tony Liu wrote: > Since I've never used source container and not able to find answers from > doc, > bear my dummy questions. BTW, I am not able to figure it out from > Dockerfile.j2. > > Take centos-source-mariadb as an example, is it built by 1) installing > packages > from package repo or 2) download source code, compile and build packages > and > install self-build packages? In case of 2), how can I tell where and which > release > source code is downloaded and how it's compiled and built? In case of 1), > what's > the difference here from centos-binary-mariadb? > > Similarly, take centos-source-keystone as another example, is it built by > 1) installing package from eg. > centos/8-stream/cloud/x86_64/openstack-ussuri/ > or 2) download source code from Github, build packages locally and install > them? > The same questions here, for 1) what's the difference from binary > container, > for 2) how can I tell which revision the source code is downloaded? > > Or, the source container is binary container with source code? > > Thanks! > Tony > ________________________________________ > Fro: Adrian Andreias > Sent: December 1, 2021 01:30 PM > To: Tony Liu > Cc: Micha? Nasiadka; openstack-discuss > Subject: Re: [kolla] Plan to deprecate binary and unify on single > distrubition > > The user doesn't have to build the images that are source based. Ready > built source-based images are available: https://quay.io/search?q=kolla > > Binary-based Kolla docker images are installing (at image build time) > OpenStack projects via system packages, while source based images are > symply adding projects' Python source code. > > Just as you said, the user shouldn't care much about distro inside images. > But makes things a lot easier for developers to maintain less images and > without working around different versions and dependencies provided by the > system packages of various distros. > > > Regards, > Adrian Andreias > > > > > On Wed, Dec 1, 2021 at 10:31 PM Tony Liu tonyliu0592 at hotmail.com>> wrote: > For OpenStack, I've been using centos-binary containers on CentOS. It's > been working very well. > > I don't have much concern on supporting single-distro based container on > different distro hosts. > That's actually one of the beauties of container, which self-contains all > packages and dependencies. > We used to have a product based on single-distro container and support > different distro hosts. > Just need to be careful when container needs any resource from the host, > like kernel, mounted > host filesystem, networking, privilege, etc. > > Regarding to source container, what's the purpose of it? Is it allow user > to get the package based > on some specific source that is not provided by any existing package repo? > In that case, I'd assume > user should always build their own source container. Then what's the > purpose of providing pre-build > source container? Also, are those non-core containers, like MariaDB, > HAProxy, RabbitMQ, etc. all > built from source code? > > > Thanks! > Tony > ________________________________________ > From: Micha? Nasiadka > > Sent: November 18, 2021 04:33 AM > To: openstack-discuss > Subject: [kolla] Plan to deprecate binary and unify on single distrubition > > Hello Koalas, > > On the PTG we have discussed two topics: > > 1) Deprecate and drop binary type of Kolla images > 2) Use a common base (single Linux distribution) for Kolla images > > This is a call for feedback - for people that have not been attending the > PTG. > > What this essentially mean for consumers: > > 1) In Yoga cycle we will deprecate binary type of Kolla images, and in Z > cycle those will be dropped. > 2) We are not going to support CentOS Stream 9 (cs9) as a base operating > system, and the source type build will rely on CentOS Stream 8 in Z release. > 3) Beginning from A release Kolla will build only Debian source images - > but Kolla-Ansible will still support deployment of those images on > CentOS/Ubuntu/Debian Host operating systems (and Rocky Linux to be added in > Yoga to that mix). > > Justification: > The Kolla project team is limited in numbers, therefore supporting current > broad mix of operating systems (especially with CentOS Stream 9 ,,on the > way??) is a significant maintenance burden. > By dropping binary type of images - users would be running more tested > images (since Kolla/Kolla-Ansible CI runs source images jobs as voting). > In Xena we?ve already changed the default image type Kolla-Ansible uses to > source. > We also feel that using a unified base OS for Kolla container images is a > way to remove some of the maintenance burden (including CI cycles and > > Request for feedback: > If any of those changes is a no go from your perspective - we?d like to > hear your opinions. > > Best regards, > Michal Nasiadka > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From tonyliu0592 at hotmail.com Wed Dec 1 23:14:55 2021 From: tonyliu0592 at hotmail.com (Tony Liu) Date: Wed, 1 Dec 2021 23:14:55 +0000 Subject: [kolla] Plan to deprecate binary and unify on single distrubition In-Reply-To: References: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> Message-ID: Thank you Adrian for clarifications! I am good with Micha?'s proposal, although it will add a bit overhead to rpm-operator when need to look into container. For example, you need to keep it in mind that the container is Debian based and run dpkg instead of rpm. Tony ________________________________________ From: Adrian Andreias Sent: December 1, 2021 03:04 PM To: Tony Liu Cc: Micha? Nasiadka; openstack-discuss Subject: Re: [kolla] Plan to deprecate binary and unify on single distrubition "source" refers here to OpenStack projects source code: https://opendev.org/openstack/kolla/src/commit/1a9d5a1a42c86b447af371ab5e0864f2b452f6ad/docker/nova/nova-base/Dockerfile.j2#L52 There is no reason to build MariaDB from source. System package are always used (regardless of binary/source config): https://opendev.org/openstack/kolla/src/commit/1a9d5a1a42c86b447af371ab5e0864f2b452f6ad/docker/mariadb/mariadb-base/Dockerfile.j2#L14 https://opendev.org/openstack/kolla/src/commit/1a9d5a1a42c86b447af371ab5e0864f2b452f6ad/docker/mariadb/mariadb-server/Dockerfile.j2#L15 Image names like "centos-source-mariadb" are probably just kept for consistency. If you want to experiment and see what's happening, you can build images yourself: https://docs.openstack.org/kolla/latest/admin/image-building.html Anyhow, the thread is about any blockers on Micha?'s proposal. I think it makes sense: less ground to cover and effort will be focused on a few reliable images. The only possible issue I can imagine would be a project that already has massive image customizations that are based on binary images. Otherwise, minor customizations or standard images should not impact change from binary to source nor to single distro. Regards, Adrian Andreias On Thu, Dec 2, 2021 at 12:04 AM Tony Liu > wrote: Since I've never used source container and not able to find answers from doc, bear my dummy questions. BTW, I am not able to figure it out from Dockerfile.j2. Take centos-source-mariadb as an example, is it built by 1) installing packages from package repo or 2) download source code, compile and build packages and install self-build packages? In case of 2), how can I tell where and which release source code is downloaded and how it's compiled and built? In case of 1), what's the difference here from centos-binary-mariadb? Similarly, take centos-source-keystone as another example, is it built by 1) installing package from eg. centos/8-stream/cloud/x86_64/openstack-ussuri/ or 2) download source code from Github, build packages locally and install them? The same questions here, for 1) what's the difference from binary container, for 2) how can I tell which revision the source code is downloaded? Or, the source container is binary container with source code? Thanks! Tony ________________________________________ Fro: Adrian Andreias > Sent: December 1, 2021 01:30 PM To: Tony Liu Cc: Micha? Nasiadka; openstack-discuss Subject: Re: [kolla] Plan to deprecate binary and unify on single distrubition The user doesn't have to build the images that are source based. Ready built source-based images are available: https://quay.io/search?q=kolla Binary-based Kolla docker images are installing (at image build time) OpenStack projects via system packages, while source based images are symply adding projects' Python source code. Just as you said, the user shouldn't care much about distro inside images. But makes things a lot easier for developers to maintain less images and without working around different versions and dependencies provided by the system packages of various distros. Regards, Adrian Andreias On Wed, Dec 1, 2021 at 10:31 PM Tony Liu >> wrote: For OpenStack, I've been using centos-binary containers on CentOS. It's been working very well. I don't have much concern on supporting single-distro based container on different distro hosts. That's actually one of the beauties of container, which self-contains all packages and dependencies. We used to have a product based on single-distro container and support different distro hosts. Just need to be careful when container needs any resource from the host, like kernel, mounted host filesystem, networking, privilege, etc. Regarding to source container, what's the purpose of it? Is it allow user to get the package based on some specific source that is not provided by any existing package repo? In that case, I'd assume user should always build their own source container. Then what's the purpose of providing pre-build source container? Also, are those non-core containers, like MariaDB, HAProxy, RabbitMQ, etc. all built from source code? Thanks! Tony ________________________________________ From: Micha? Nasiadka >> Sent: November 18, 2021 04:33 AM To: openstack-discuss Subject: [kolla] Plan to deprecate binary and unify on single distrubition Hello Koalas, On the PTG we have discussed two topics: 1) Deprecate and drop binary type of Kolla images 2) Use a common base (single Linux distribution) for Kolla images This is a call for feedback - for people that have not been attending the PTG. What this essentially mean for consumers: 1) In Yoga cycle we will deprecate binary type of Kolla images, and in Z cycle those will be dropped. 2) We are not going to support CentOS Stream 9 (cs9) as a base operating system, and the source type build will rely on CentOS Stream 8 in Z release. 3) Beginning from A release Kolla will build only Debian source images - but Kolla-Ansible will still support deployment of those images on CentOS/Ubuntu/Debian Host operating systems (and Rocky Linux to be added in Yoga to that mix). Justification: The Kolla project team is limited in numbers, therefore supporting current broad mix of operating systems (especially with CentOS Stream 9 ,,on the way??) is a significant maintenance burden. By dropping binary type of images - users would be running more tested images (since Kolla/Kolla-Ansible CI runs source images jobs as voting). In Xena we?ve already changed the default image type Kolla-Ansible uses to source. We also feel that using a unified base OS for Kolla container images is a way to remove some of the maintenance burden (including CI cycles and Request for feedback: If any of those changes is a no go from your perspective - we?d like to hear your opinions. Best regards, Michal Nasiadka From tonyliu0592 at hotmail.com Wed Dec 1 23:47:09 2021 From: tonyliu0592 at hotmail.com (Tony Liu) Date: Wed, 1 Dec 2021 23:47:09 +0000 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: Message-ID: Here is what I would do, not validated. * Create a network with DHCP disabled. * Launch a VM to support PXE boot. * Launch a VM on volume based on any image for PXE boot. * The default virtio nic works fine with PXE boot. * "virsh edit " to enable PXE boot on the compute node. * "soft reboot" VM should bring VM to PXE boot. I prefer to do this with virsh/libvirt directly, which is much easier. Tony ________________________________________ From: Thomas Goirand Sent: December 1, 2021 02:33 PM To: openstack-discuss at lists.openstack.org Subject: [nova] Spawning instance that will do PXE booting Hi, In order to test OCI [1] within OpenStack itself, I would need my instances to do PXE booting. This would be with something like this on the qemu command line: qemu-system-x86_64 -boot n I believe this would also mean setting-up network cards that can do PXE booting, like the e1000, which Qemu supports. If that is possible, then I would setup DHCP + tftp-hpa on another instance, wired on the same private network. Your thoughts? Is this possible? Cheers, Thomas Goirand (zigo) [1] https://salsa.debian.org/openstack-team/debian/openstack-cluster-installer From cboylan at sapwetik.org Wed Dec 1 23:51:43 2021 From: cboylan at sapwetik.org (Clark Boylan) Date: Wed, 01 Dec 2021 15:51:43 -0800 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: Message-ID: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> On Wed, Dec 1, 2021, at 2:33 PM, Thomas Goirand wrote: > Hi, > > In order to test OCI [1] within OpenStack itself, I would need my > instances to do PXE booting. This would be with something like this on > the qemu command line: > > qemu-system-x86_64 -boot n > > I believe this would also mean setting-up network cards that can do PXE > booting, like the e1000, which Qemu supports. > > If that is possible, then I would setup DHCP + tftp-hpa on another > instance, wired on the same private network. > > Your thoughts? Is this possible? I'm not super familiar with all the details but I believe ironic does something similar in their devstack plugin [2]. They do this so they can create fake baremetal instances that are actually VMs that ironic manages. But a similar setup will probably work for your use cases if PXE is what you are starting with. > > Cheers, > > Thomas Goirand (zigo) > > [1] > https://salsa.debian.org/openstack-team/debian/openstack-cluster-installer [2] https://opendev.org/openstack/ironic/src/branch/master/devstack/lib/ironic#L2109 From zigo at debian.org Thu Dec 2 00:04:03 2021 From: zigo at debian.org (Thomas Goirand) Date: Thu, 2 Dec 2021 01:04:03 +0100 Subject: [kolla] Plan to deprecate binary and unify on single distrubition In-Reply-To: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> References: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> Message-ID: On 11/18/21 1:33 PM, Micha? Nasiadka wrote: > Hello Koalas, > > On the PTG we have discussed two topics: > > 1) Deprecate and drop binary type of Kolla images > 2) Use a common base (single Linux distribution) for Kolla images > > This is a call for feedback - for people that have not been attending the PTG. > > What this essentially mean for consumers: > > 1) In Yoga cycle we will deprecate binary type of Kolla images, and in Z cycle those will be dropped. > 2) We are not going to support CentOS Stream 9 (cs9) as a base operating system, and the source type build will rely on CentOS Stream 8 in Z release. > 3) Beginning from A release Kolla will build only Debian source images - but Kolla-Ansible will still support deployment of those images on CentOS/Ubuntu/Debian Host operating systems (and Rocky Linux to be added in Yoga to that mix). > > Justification: > The Kolla project team is limited in numbers, therefore supporting current broad mix of operating systems (especially with CentOS Stream 9 ,,on the way??) is a significant maintenance burden. > By dropping binary type of images - users would be running more tested images (since Kolla/Kolla-Ansible CI runs source images jobs as voting). > In Xena we?ve already changed the default image type Kolla-Ansible uses to source. > We also feel that using a unified base OS for Kolla container images is a way to remove some of the maintenance burden (including CI cycles and > > Request for feedback: > If any of those changes is a no go from your perspective - we?d like to hear your opinions. > > Best regards, > Michal Nasiadka Hi, While I fully support moving to a Debian only world (anyone who knows me isn't surprised to read this...), I really wonder if moving to a "source only" world is the right step. If you don't have enough people, why don't you just move to a Debian package only type of install, rather than re-implementing all by yourself? The very point of packages is to abstract for you many things, like for example, the fact that Nova has a reserved UID/GID of 64060 in Debian, and will always use that one, and automatically set that up for you, plus put the Nova user in the right group... Or if you need another example, making sure that swift-proxy starts with the right "route = .* addheader:Date: ${httptime[]}" uwsgi parameter... etc. Now, if you have a limited amount of contributors, all these tiny tweaks that the package maintainers are doing for you, you'll have to re-implement them yourself. Worse: it's going to be even more complicated if you put stuff in containers. Using source only images, you also don't get all the testing that's done by distro on funny arch which you probably didn't think of. Yes, the openvswitch package in Debian is tested and working on the 10 official Debian arch. The lz4tools contains a 32 bits patch. Or is it that you don't care about anything else than x86_64? So why don't you just take the other direction, and drop the source images? I'm sure this will save you a lot of work. Plus myself and Michal will make sure everything works, and is available in time for every release, like we always did (did you notice Debian is often (if not always) the first distro to have available packages for each release?). Cheers, Thomas Goirand (zigo) P.S: I'm using puppet-openstack myself, so I wont ever contribute to Kolla and/or OSA, but I still would like to support you... From zigo at debian.org Thu Dec 2 00:09:58 2021 From: zigo at debian.org (Thomas Goirand) Date: Thu, 2 Dec 2021 01:09:58 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: Message-ID: <70a61878-a798-f833-48c8-f177ce4fa38b@debian.org> Hi Tony, Thanks for your reply. On 12/2/21 12:47 AM, Tony Liu wrote: > Here is what I would do, not validated. > * Create a network with DHCP disabled. > * Launch a VM to support PXE boot. > * Launch a VM on volume based on any image for PXE boot. > * The default virtio nic works fine with PXE boot. > * "virsh edit " to enable PXE boot on the compute node. This isn't an option. To do "virsh edit", you need to have root access to a compute node, and that's not what I would like to do: I'd like my CI to run with a normal OpenStack account only, to make it easy for people to reuse. > * "soft reboot" VM should bring VM to PXE boot. > > I prefer to do this with virsh/libvirt directly, which is much easier. > > Tony Cheers, Thomas Goirand (zigo) From openinfradn at gmail.com Thu Dec 2 03:28:15 2021 From: openinfradn at gmail.com (open infra) Date: Thu, 2 Dec 2021 08:58:15 +0530 Subject: cpu pinning Message-ID: Hi, I have created a flavor with following properties and created an instance. Instance failed with the error "No valid host was found. There are not enough hosts available." When I set the cpu policy as 'shared' I can create the instance. The host machine has two numa nodes and a total of 128 vcpu. I can not figure out what's missing here. controller-1:~$ openstack flavor show dn.large -c properties +------------+--------------------------------------------------------------------------------------------------------+ | Field | Value | +------------+--------------------------------------------------------------------------------------------------------+ | properties | hw:cpu_cores='2', hw:cpu_policy='dedicated', hw:cpu_sockets='1', hw:cpu_threads='2', hw:numa_nodes='1' | +------------+--------------------------------------------------------------------------------------------------------+ controller-1:~$ openstack hypervisor stats show +----------------------+--------+ | Field | Value | +----------------------+--------+ | count | 1 | | current_workload | 0 | | disk_available_least | 187 | | free_disk_gb | 199 | | free_ram_mb | 308787 | | local_gb | 219 | | local_gb_used | 20 | | memory_mb | 515443 | | memory_mb_used | 206656 | | running_vms | 7 | | vcpus | 126 | | vcpus_used | 49 | +----------------------+--------+ Regards, Danishka -------------- next part -------------- An HTML attachment was scrubbed... URL: From mnasiadka at gmail.com Thu Dec 2 07:37:34 2021 From: mnasiadka at gmail.com (=?utf-8?Q?Micha=C5=82_Nasiadka?=) Date: Thu, 2 Dec 2021 08:37:34 +0100 Subject: [kolla] Plan to deprecate binary and unify on single distrubition In-Reply-To: References: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> Message-ID: Hi Thomas, Thanks for your opinion - but we can?t really wait until each Debian OpenStack release to test all the features (and what will be broken) - we already trail long time after the official OpenStack release due to other issues. That would significantly lengthen the delay and the process involved. Kolla has been doing voting CI on source images for a long time - and we don?t want to change it. It?s also a direction to make the external dependencies smaller - not bigger - and I see the custom things added from packager perspective as a dependency. Best regards, Michal > On 2 Dec 2021, at 01:04, Thomas Goirand wrote: > > On 11/18/21 1:33 PM, Micha? Nasiadka wrote: >> Hello Koalas, >> >> On the PTG we have discussed two topics: >> >> 1) Deprecate and drop binary type of Kolla images >> 2) Use a common base (single Linux distribution) for Kolla images >> >> This is a call for feedback - for people that have not been attending the PTG. >> >> What this essentially mean for consumers: >> >> 1) In Yoga cycle we will deprecate binary type of Kolla images, and in Z cycle those will be dropped. >> 2) We are not going to support CentOS Stream 9 (cs9) as a base operating system, and the source type build will rely on CentOS Stream 8 in Z release. >> 3) Beginning from A release Kolla will build only Debian source images - but Kolla-Ansible will still support deployment of those images on CentOS/Ubuntu/Debian Host operating systems (and Rocky Linux to be added in Yoga to that mix). >> >> Justification: >> The Kolla project team is limited in numbers, therefore supporting current broad mix of operating systems (especially with CentOS Stream 9 ,,on the way??) is a significant maintenance burden. >> By dropping binary type of images - users would be running more tested images (since Kolla/Kolla-Ansible CI runs source images jobs as voting). >> In Xena we?ve already changed the default image type Kolla-Ansible uses to source. >> We also feel that using a unified base OS for Kolla container images is a way to remove some of the maintenance burden (including CI cycles and >> >> Request for feedback: >> If any of those changes is a no go from your perspective - we?d like to hear your opinions. >> >> Best regards, >> Michal Nasiadka > > Hi, > > While I fully support moving to a Debian only world (anyone who knows me > isn't surprised to read this...), I really wonder if moving to a "source > only" world is the right step. > > If you don't have enough people, why don't you just move to a Debian > package only type of install, rather than re-implementing all by > yourself? The very point of packages is to abstract for you many things, > like for example, the fact that Nova has a reserved UID/GID of 64060 in > Debian, and will always use that one, and automatically set that up for > you, plus put the Nova user in the right group... Or if you need another > example, making sure that swift-proxy starts with the right "route = .* > addheader:Date: ${httptime[]}" uwsgi parameter... etc. > > Now, if you have a limited amount of contributors, all these tiny tweaks > that the package maintainers are doing for you, you'll have to > re-implement them yourself. Worse: it's going to be even more > complicated if you put stuff in containers. > > Using source only images, you also don't get all the testing that's done > by distro on funny arch which you probably didn't think of. Yes, the > openvswitch package in Debian is tested and working on the 10 official > Debian arch. The lz4tools contains a 32 bits patch. Or is it that you > don't care about anything else than x86_64? > > So why don't you just take the other direction, and drop the source > images? I'm sure this will save you a lot of work. Plus myself and > Michal will make sure everything works, and is available in time for > every release, like we always did (did you notice Debian is often (if > not always) the first distro to have available packages for each release?). > > Cheers, > > Thomas Goirand (zigo) > > P.S: I'm using puppet-openstack myself, so I wont ever contribute to > Kolla and/or OSA, but I still would like to support you... > From zigo at debian.org Thu Dec 2 08:07:23 2021 From: zigo at debian.org (Thomas Goirand) Date: Thu, 2 Dec 2021 09:07:23 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> References: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> Message-ID: <5af59ddc-669f-09cf-233e-5f85f0e89999@debian.org> On 12/2/21 12:51 AM, Clark Boylan wrote: > I'm not super familiar with all the details but I believe ironic does something similar in their devstack plugin [2]. They do this so they can create fake baremetal instances that are actually VMs that ironic manages. But a similar setup will probably work for your use cases if PXE is what you are starting with. > > [2] https://opendev.org/openstack/ironic/src/branch/master/devstack/lib/ironic#L2109 Hi, Thanks Clark for your answer. I've read the script a little bit, and I have to admit I do not understand what it does... :( If some Ironic falks can explain, that'd be great! Cheers, Thomas Goirand (zigo) From zigo at debian.org Thu Dec 2 08:15:55 2021 From: zigo at debian.org (Thomas Goirand) Date: Thu, 2 Dec 2021 09:15:55 +0100 Subject: [kolla] Plan to deprecate binary and unify on single distrubition In-Reply-To: References: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> Message-ID: <0cb15dca-99d0-9358-233d-4e85b5d1104d@debian.org> Hi Michal, Thanks for your reply. I'm sorry to say that your reply shows a poor understanding of what distributions and packaging are in general, and what Debian does for OpenStack in particular. But I'm not surprised, this isn't the first time I see this. It also happened when I was employed by Mirantis, and with cloudwatt. People usually just dismiss packaging because they see it as a black-box they cannot act on, because they just happen to not do packaging themselves. On 12/2/21 8:37 AM, Micha? Nasiadka wrote: > Hi Thomas, > > Thanks for your opinion - but we can?t really wait until each Debian OpenStack release to test all the features (and what will be broken) - we already trail long time after the official OpenStack release due to other issues. > That would significantly lengthen the delay and the process involved. This isn't fair, and IMO isn't truth. Usually, Debian packages are ready *before* the final release, often just a week after the first RCs are out. The only reason why there's always one or 2 glitches, is because I couldn't install the new release until everything is finished (and from my side, including updating my puppet scripts for the new release of puppet-openstack). So, we're talking about just a few days delay after the release. This normally gives you enough time to fix things, and have Kolla ready the day of the release. I used to package all the beta releases, twice per OpenStack release. Though nobody ever consumed them. And these days, not all projects are release beta releases, so it is kind of useless. > It?s also a direction to make the external dependencies smaller - not bigger - and I see the custom things added from packager perspective as a dependency. IMO, that's a very wrong perspective. Package maintainers (that's how we're called, not "packager" please...) are adding automations and system integrations. If you refuse them, this means you'll have to implement them yourself. Remember that I use puppet myself, so I very well know what I'm talking about, when discussing config management and package integration. When I can make a choice of where to do the automation, system dependency, etc., then it goes to the packaging: this is the normal and natural place to do it. A config management system should not be the one taking care of any of the things that packages normally take care of (like dependencies). Handling them in your config management system is not only an heresy, it's also using the wrong tool for it, and achieving a poorer result. Also, handing dependency management using something like pip is a non-sense, because it only knows about Python dependencies (not anything else), and it's been demonstrated so many times how bad pip is as a dependency solver: for example, it broke the CI so many times, showing its defects. Another thing you're seeing wrong, is about smaller dependencies. First, having one more Python package will not hurt, and if it is there, it is for a reason. Second, packages are designed to have the smaller amount of dependency possible. Always. As a rule. If there's something wrong there, it's easy to fix (and you could contribute to it...). And we also make sure that everything works together, adding patches when needed (how are you going to manage patches?). I'm not hoping to convince you, I feel it's already too late. Though I can't help myself to let you know I believe you're doing the wrong technical choice. Cheers, Thomas Goirand (zigo) From mdemaced at redhat.com Thu Dec 2 11:10:09 2021 From: mdemaced at redhat.com (Maysa De Macedo Souza) Date: Thu, 2 Dec 2021 12:10:09 +0100 Subject: [kuryr] Propose to EOL stable branches (Stein, Rocky and Queens) for all the kuryr scope In-Reply-To: <3c2cbd0af92ff7dffb3fae4ce6c30015abd94b6b.camel@redhat.com> References: <3c2cbd0af92ff7dffb3fae4ce6c30015abd94b6b.camel@redhat.com> Message-ID: Hello, Here is the patch[1] proposing the eol for pike, queen, rocky and stein branches. If anyone wants to keep those branches, let us know. [1] https://review.opendev.org/c/openstack/releases/+/820144/ Cheers, Maysa. On Wed, Dec 1, 2021 at 9:59 AM Micha? Dulko wrote: > Yes, let's remove them. It has little sense to keep these versions and > we're actually actively working to keep the compatibility with older > OpenStack versions anyway. > > On Fri, 2021-11-26 at 23:50 +0100, Maysa De Macedo Souza wrote: > > Hello again, > > > > We also have Pike as a stable branch[0], so that is the older one, > > which would also be EOL. > > > > [0] > > > https://review.opendev.org/q/(project:%255Eopenstack/kuryr.*)AND((branch:stable/queens)OR(branch:stable/rocky)OR(branch:stable/stein)OR(branch:stable/pike)) > > > > Cheers, > > Maysa. > > > > On Fri, Nov 26, 2021 at 11:42 PM Maysa De Macedo Souza > > wrote: > > > Hello, > > > > > > We have plenty of stable branches in Kuryr repositories still open > > > with the oldest being stable/queens. > > > We haven't seen many backports[0] proposed to Stein, Rocky and > > > Queens, so I would like to propose retiring these branches. > > > > > > Let me know if anybody is interested in keeping any of those > > > branches. > > > > > > [0] > > > > https://review.opendev.org/q/(project:%255Eopenstack/kuryr.*)AND((branch:stable/queens)OR(branch:stable/rocky)OR(branch:stable/stein)) > > > > > > Thank you, > > > Maysa Macedo. > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From gael.therond at bitswalk.com Thu Dec 2 11:57:35 2021 From: gael.therond at bitswalk.com (=?UTF-8?Q?Ga=C3=ABl_THEROND?=) Date: Thu, 2 Dec 2021 11:57:35 +0000 Subject: Windows imaging process In-Reply-To: References: Message-ID: Hi Patryk, Thanks a lot for the link and sorry for the late answer! As stated in my previous email, this is not going to work on our side of things as our security process prohibits us from modifying the Original ISO. Not that I CAN'T do it as I obviously have the hand on the whole image creation, just that if I'm not respecting the workflow those images won't pass the qualification gate. What really astonish me is that KVM is CDRom IDE HW Bus aware and able to mount multiple of them for a same instance, Nova itself knows about those specific HW bus too as you can actually specify it on the image metadata, what's really missing in my process for now is the ability to let nova spawn a VM with both ISO and VIRTIO drivers as CDRom devices. I'm really wondering why we can't specify any additional HW resources to nova at spawn time (from Horizon and CLI) except using native nova client such as instructed in here: https://access.redhat.com/solutions/1225473 or did I miss something? Thanks a lot for your answer! Le mar. 30 nov. 2021 ? 00:13, Patryk Jakuszew a ?crit : > On Tue, 30 Nov 2021 at 01:11, Ga?l THEROND > wrote: > > > > Starting from an unaltered Microsoft originated ISO image is a mandatory > requirement for this project (Because of security constraints that I can?t > have any impact on). > > > > Does that mean you cannot alter the original image? There are methods > for inserting the virtio drivers into the installation media, but that > will require generating a new installer ISO out of the modified files. > > This article seems to be describing the procedure correctly: > > https://portal.nutanix.com/page/documents/kbs/details?targetId=kA00e000000bt28CAA > -------------- next part -------------- An HTML attachment was scrubbed... URL: From satish.txt at gmail.com Thu Dec 2 13:15:21 2021 From: satish.txt at gmail.com (Satish Patel) Date: Thu, 2 Dec 2021 08:15:21 -0500 Subject: cpu pinning In-Reply-To: References: Message-ID: Did you try hw:numa_nodes=2 ? Sent from my iPhone > On Dec 1, 2021, at 10:30 PM, open infra wrote: > > hw:numa_nodes='1' -------------- next part -------------- An HTML attachment was scrubbed... URL: From corey.bryant at canonical.com Thu Dec 2 13:25:16 2021 From: corey.bryant at canonical.com (Corey Bryant) Date: Thu, 2 Dec 2021 08:25:16 -0500 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <50abb4f5-af13-435c-87c9-1c14d74dcfb2@www.fastmail.com> Message-ID: On Wed, Dec 1, 2021 at 4:34 PM Clark Boylan wrote: > > > >> > >> If we want to be bold and try an experiment, we could also reduce > testing of the python versions to oldest and newest supported python > versions. For example python3.6 and python3.9 during Yoga. Don't worry > about python3.7 and python3.8 as it is likely they will continue to > function as long as 3.6 and 3.9 are happy. This would reduce the size of > the testing matrix. We can always reevaluate if problems for 3.7 and 3.8 > occur. > >> > > > > The issue here I think is trying to align on a default LTS distro > > python. CS8,18.04 is 3.6, Debian Buster is 3.7, 20.04 is 3.8 and CS9 > > will be 3.9. Personally I think 3.6/3.9 is sufficient because for the > > projects I contribute to this covers my cases. But this would exclude > > the 3.8 on 20.04. Would this be a problem for Ubuntu's UCA? > > Ubuntu 20.04 includes a python 3.9. I don't know how that would impact the > UCA or Ubuntu's packaging though. That said I have a strong suspicion that > if OpenStack runs on 3.6 and 3.9 that it should just work on 3.8 as well. > And as I mentioned if we prove that wrong through experience we can always > fix 3.8 and go back to explicitly testing it. I don't think this particular > change is super critical, but I know some people were concerned about the > number of pythons that would need testing. > > > Ubuntu will be supporting python3.8 (on 20.04) and python3.10 (on 22.04) for Yoga by default. Once python 3.10.1 is released, it will be backported to 20.04, so I was hoping to add non-voting py310 unit tests upstream. Tthoughts? We did that in the past, I forget what release. Python3.9 testing is probably good enough coverage for Python3.8. Generally testing with the min and max python versions seems sensible, although the gap feels like it's getting wider than it was in the past, likely due to python release cadence. Corey -------------- next part -------------- An HTML attachment was scrubbed... URL: From smooney at redhat.com Thu Dec 2 14:16:14 2021 From: smooney at redhat.com (Sean Mooney) Date: Thu, 02 Dec 2021 14:16:14 +0000 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <50abb4f5-af13-435c-87c9-1c14d74dcfb2@www.fastmail.com> Message-ID: <95d04b19475514d1132720ee1fc6064152b31710.camel@redhat.com> On Thu, 2021-12-02 at 08:25 -0500, Corey Bryant wrote: > On Wed, Dec 1, 2021 at 4:34 PM Clark Boylan wrote: > > > > > > > > > > > > > > > If we want to be bold and try an experiment, we could also reduce > > testing of the python versions to oldest and newest supported python > > versions. For example python3.6 and python3.9 during Yoga. Don't worry > > about python3.7 and python3.8 as it is likely they will continue to > > function as long as 3.6 and 3.9 are happy. This would reduce the size of > > the testing matrix. We can always reevaluate if problems for 3.7 and 3.8 > > occur. > > > > > > > > > > The issue here I think is trying to align on a default LTS distro > > > python. CS8,18.04 is 3.6, Debian Buster is 3.7, 20.04 is 3.8 and CS9 > > > will be 3.9. Personally I think 3.6/3.9 is sufficient because for the > > > projects I contribute to this covers my cases. But this would exclude > > > the 3.8 on 20.04. Would this be a problem for Ubuntu's UCA? > > > > Ubuntu 20.04 includes a python 3.9. I don't know how that would impact the > > UCA or Ubuntu's packaging though. That said I have a strong suspicion that > > if OpenStack runs on 3.6 and 3.9 that it should just work on 3.8 as well. > > And as I mentioned if we prove that wrong through experience we can always > > fix 3.8 and go back to explicitly testing it. I don't think this particular > > change is super critical, but I know some people were concerned about the > > number of pythons that would need testing. > > > > > > > Ubuntu will be supporting python3.8 (on 20.04) and python3.10 (on 22.04) > for Yoga by default. Once python 3.10.1 is released, it will be backported > to 20.04, so I was hoping to add non-voting py310 unit tests upstream. > Tthoughts? We did that in the past, I forget what release. i think that likely would be a good idea. i know eventlets are currently adding inital 3.10 support and have had to adapt to some changes sicne py3.6 by increasing the min version of some of there deps like dns python getting a head up early and ensuring we can get our upper-constraits in order i think woudl help make that smother. https://github.com/eventlet/eventlet/pull/715 > > Python3.9 testing is probably good enough coverage for Python3.8. Generally > testing with the min and max python versions seems sensible, although the > gap feels like it's getting wider than it was in the past, likely due to > python release cadence. ya it is wider then in the past for yoga we will effectlyve need to support 3.6, 3.7, 3.8, and 3.9 during development and then 3.10 once ubuntu 22.04 release. long term supporting 5 python version i dont think is sutainable so if we can reduce that to 3.8 3.9 and 3.10 for Z i think that is what we should aim for. we might event want to aim for 3.9 and 3.10 only if the distro support matrix makes sense. suse is perhaps the only distro that may not supprot 3.9+ ? and they nolonger ship an openstack distro so perhaps we coudl limit to 3.9+ in z > > Corey From gmann at ghanshyammann.com Thu Dec 2 15:01:52 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Thu, 02 Dec 2021 09:01:52 -0600 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <20211126143019.2o7rycs44ycnkgez@yuggoth.org> <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <50abb4f5-af13-435c-87c9-1c14d74dcfb2@www.fastmail.com> Message-ID: <17d7baa83ad.10b8913d2240727.1175999501627088346@ghanshyammann.com> ---- On Thu, 02 Dec 2021 07:25:16 -0600 Corey Bryant wrote ---- > > > On Wed, Dec 1, 2021 at 4:34 PM Clark Boylan wrote: > > >> > >> If we want to be bold and try an experiment, we could also reduce testing of the python versions to oldest and newest supported python versions. For example python3.6 and python3.9 during Yoga. Don't worry about python3.7 and python3.8 as it is likely they will continue to function as long as 3.6 and 3.9 are happy. This would reduce the size of the testing matrix. We can always reevaluate if problems for 3.7 and 3.8 occur. > >> > > > > The issue here I think is trying to align on a default LTS distro > > python. CS8,18.04 is 3.6, Debian Buster is 3.7, 20.04 is 3.8 and CS9 > > will be 3.9. Personally I think 3.6/3.9 is sufficient because for the > > projects I contribute to this covers my cases. But this would exclude > > the 3.8 on 20.04. Would this be a problem for Ubuntu's UCA? > > Ubuntu 20.04 includes a python 3.9. I don't know how that would impact the UCA or Ubuntu's packaging though. That said I have a strong suspicion that if OpenStack runs on 3.6 and 3.9 that it should just work on 3.8 as well. And as I mentioned if we prove that wrong through experience we can always fix 3.8 and go back to explicitly testing it. I don't think this particular change is super critical, but I know some people were concerned about the number of pythons that would need testing. > > > Ubuntu will be supporting python3.8 (on 20.04) and python3.10 (on 22.04) for Yoga by default. Once python 3.10.1 is released, it will be backported to 20.04, so I was hoping to add non-voting py310 unit tests upstream. Tthoughts? We did that in the past, I forget what release. Yeah, adding 3.10 as non-voting is a good idea like we did for py3.9. Anyways TC is going to discuss these in today's TC meeting, please join us there to conclude things. -gmann > Python3.9 testing is probably good enough coverage for Python3.8. Generally testing with the min and max python versions seems sensible, although the gap feels like it's getting wider than it was in the past, likely due to python release cadence. > > Corey From openinfradn at gmail.com Thu Dec 2 15:03:01 2021 From: openinfradn at gmail.com (open infra) Date: Thu, 2 Dec 2021 20:33:01 +0530 Subject: cpu pinning In-Reply-To: References: Message-ID: On Thu, Dec 2, 2021 at 6:45 PM Satish Patel wrote: > Did you try hw:numa_nodes=2 ? > Yes, I tried but I am getting the same error "No valid host was found. There are not enough hosts available." hw:cpu_cores='2', hw:cpu_policy='dedicated', hw:cpu_sockets='1', hw:cpu_threads='2', hw:numa_nodes='2' > > Sent from my iPhone > > On Dec 1, 2021, at 10:30 PM, open infra wrote: > > hw:numa_nodes='1' > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From openinfradn at gmail.com Thu Dec 2 15:06:47 2021 From: openinfradn at gmail.com (open infra) Date: Thu, 2 Dec 2021 20:36:47 +0530 Subject: cpu pinning In-Reply-To: References: Message-ID: I have only one worker and I hope it won't be an issue. On Thu, Dec 2, 2021 at 8:33 PM open infra wrote: > > > On Thu, Dec 2, 2021 at 6:45 PM Satish Patel wrote: > >> Did you try hw:numa_nodes=2 ? >> > > Yes, I tried but I am getting the same error "No valid host was found. > There are not enough hosts available." > > hw:cpu_cores='2', hw:cpu_policy='dedicated', hw:cpu_sockets='1', > hw:cpu_threads='2', hw:numa_nodes='2' > > > >> >> Sent from my iPhone >> >> On Dec 1, 2021, at 10:30 PM, open infra wrote: >> >> hw:numa_nodes='1' >> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: From smooney at redhat.com Thu Dec 2 15:46:18 2021 From: smooney at redhat.com (Sean Mooney) Date: Thu, 02 Dec 2021 15:46:18 +0000 Subject: cpu pinning In-Reply-To: References: Message-ID: On Thu, 2021-12-02 at 08:58 +0530, open infra wrote: > Hi, > > I have created a flavor with following properties and created an instance. > Instance failed with the error "No valid host was found. There are not > enough hosts available." > When I set the cpu policy as 'shared' I can create the instance. The host > machine has two numa nodes and a total of 128 vcpu. > I can not figure out what's missing here. i suspect the issue is not with the flavor but with yoru host configurtion. you likely need to defience cpu_dedicated_set and cpu_shared_set in the nova.conf we do not support mixing pinned and floating cpus on the same host unless you partion the cpu pool using cpu_dedicated_set and cpu_shared_set. as of train cpu_dedicated_set replaced vcpu_pin_set as the supported way to report the pool of cpus to be used for pinned vms to placment. if you do "openstack resource provider inventory show " it should detail the avaiabel pcpu and vcpu inventories. when you use hw:cpu_policy='dedicated' it will claim PCPUs not VCPUs in placment. That is likely the issue you are encountering. by default we have a fallback query to make this work while you are upgrading https://docs.openstack.org/nova/latest/configuration/config.html#workarounds.disable_fallback_pcpu_query which we should be disabling by default soon. but i suspect that is likely why you are getting the no valid host. to debug this properly you should enable debug logging on the nova schduler and then confirm if you got host back form placment and then if the numa toplogy filter is rejectign the host or not. without the schduler debug logs for the instance creation we cannt really help more then this since we do not have the info required. > > controller-1:~$ openstack flavor show dn.large -c properties > > +------------+--------------------------------------------------------------------------------------------------------+ > > > Field | Value > | > > +------------+--------------------------------------------------------------------------------------------------------+ > > > properties | hw:cpu_cores='2', hw:cpu_policy='dedicated', > hw:cpu_sockets='1', hw:cpu_threads='2', hw:numa_nodes='1' | > > +------------+--------------------------------------------------------------------------------------------------------+ > > controller-1:~$ openstack hypervisor stats show > > +----------------------+--------+ > > > Field | Value | > > +----------------------+--------+ > > > count | 1 | > > > current_workload | 0 | > > > disk_available_least | 187 | > > > free_disk_gb | 199 | > > > free_ram_mb | 308787 | > > > local_gb | 219 | > > > local_gb_used | 20 | > > > memory_mb | 515443 | > > > memory_mb_used | 206656 | > > > running_vms | 7 | > > > vcpus | 126 | > > > vcpus_used | 49 | > > +----------------------+--------+ > > > > Regards, > > Danishka From fungi at yuggoth.org Thu Dec 2 15:51:08 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Thu, 2 Dec 2021 15:51:08 +0000 Subject: python_requires >= 3.8 during Yoga In-Reply-To: References: <17d5cdde95a.b95752d01381326.4462262438602968824@ghanshyammann.com> <50abb4f5-af13-435c-87c9-1c14d74dcfb2@www.fastmail.com> Message-ID: <20211202155107.3rg2t5yxsni5btvs@yuggoth.org> On 2021-12-02 08:25:16 -0500 (-0500), Corey Bryant wrote: [...] > Python3.9 testing is probably good enough coverage for Python3.8. > Generally testing with the min and max python versions seems > sensible, although the gap feels like it's getting wider than it > was in the past, likely due to python release cadence. It seems like we're primarily talking about unit tests here. There will still be a fair amount of direct 3.8 coverage as DevStack/Tempest/Grenade jobs use the default python3 on Focal unless we tell them not to. Same goes for docs builds, release jobs, et cetera. -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From rosmaita.fossdev at gmail.com Thu Dec 2 16:07:04 2021 From: rosmaita.fossdev at gmail.com (Brian Rosmaita) Date: Thu, 2 Dec 2021 11:07:04 -0500 Subject: [cinder] R-17 midcycle followup meeting poll Message-ID: We had some very useful discussions at yesterday's midcycle but ran out of time before we could discuss Gorka's new quotas proposals. It turns out that due to 8 December being a public holiday in many locations, the next Cinder meeting won't work for a video followup, so we'll hold a special half-hour video meeting on this topic next week. The meeting will be 1400-1430 UTC on one of Tuesday, Thursday, or Friday. If you have a preference, please express it in this poll: https://doodle.com/poll/mnddwtp99uu6hsuq The poll closes at 1300 UTC on Monday 6 December, and I'll announce the selected time shortly thereafter. cheers, brian From cdilorenzo at gmail.com Thu Dec 2 17:00:41 2021 From: cdilorenzo at gmail.com (Chris DiLorenzo) Date: Thu, 2 Dec 2021 12:00:41 -0500 Subject: [kolla[ Issue running masakari-hostmonitor Message-ID: Hello, I am running Kolla-12.0.0 with Openstack Victoria. I have mostly successfully deployed masakari. The instance monitoring works, I can set up segments in Horizon and via CLI. But I can't get masakari-hostmonitor to start: 2021-12-02 16:50:44.999 7 INFO oslo.privsep.daemon [-] Running privsep helper: ['sudo', 'privsep-helper', '--config-file', '/etc/masakari-monitors/masakari -monitors.conf', '--privsep_context', 'masakarimonitors.privsep.monitors_priv', '--privsep_sock_path', '/tmp/tmpf_65hc7m/privsep.sock'] 2021-12-02 16:50:44.892 261 INFO oslo.privsep.daemon [-] privsep daemon starting 2021-12-02 16:50:44.898 261 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0 2021-12-02 16:50:44.902 261 ERROR oslo.privsep.daemon [-] [Errno 1] Operation not permitted Traceback (most recent call last): File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 593, in helper_main Daemon(channel, context).run() File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 403, in run self._drop_privs() File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 439, in _drop_privs capabilities.drop_all_caps_except(self.caps, self.caps, []) File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/capabilities.py", line 156, in drop_all_caps_except raise OSError(errno, os.strerror(errno)) PermissionError: [Errno 1] Operation not permitted 2021-12-02 16:50:45.001 7 DEBUG oslo_privsep.comm [-] EOF on privsep read channel _reader_main /var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep /comm.py:149 2021-12-02 16:50:45.475 7 WARNING oslo.privsep.daemon [-] privsep log: [Errno 1] Operation not permitted 2021-12-02 16:50:45.565 7 INFO oslo.privsep.daemon [-] Spawned new privsep daemon via rootwrap 2021-12-02 16:50:45.565 7 DEBUG oslo.privsep.daemon [-] Accepted privsep connection to /tmp/tmpf_65hc7m/privsep.sock __init__ /var/lib/kolla/venv/lib/pytho n3.8/site-packages/oslo_privsep/daemon.py:371 2021-12-02 16:50:45.467 267 INFO oslo.privsep.daemon [-] privsep daemon starting 2021-12-02 16:50:45.471 267 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0 2021-12-02 16:50:45.474 267 ERROR oslo.privsep.daemon [-] [Errno 1] Operation not permitted Traceback (most recent call last): File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 593, in helper_main Daemon(channel, context).run() File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 403, in run self._drop_privs() File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 439, in _drop_privs capabilities.drop_all_caps_except(self.caps, self.caps, []) File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/capabilities.py", line 156, in drop_all_caps_except raise OSError(errno, os.strerror(errno)) PermissionError: [Errno 1] Operation not permitted sudoers in the container appears to be set up correctly. Thanks Chris -------------- next part -------------- An HTML attachment was scrubbed... URL: From radoslaw.piliszek at gmail.com Thu Dec 2 19:28:08 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Thu, 2 Dec 2021 20:28:08 +0100 Subject: [kolla[ Issue running masakari-hostmonitor In-Reply-To: References: Message-ID: Hi Chris, as for the issue itself - check if the container is running privileged (it should). As for the other details - Kolla-Ansible 12.x is for Wallaby, not Victoria, that could be causing issues. 12.0.0 is not the latest release for Wallaby either (12.2.0 is for today). Fixing these might help with the other issue. -yoctozepto On Thu, 2 Dec 2021 at 18:08, Chris DiLorenzo wrote: > > Hello, > > I am running Kolla-12.0.0 with Openstack Victoria. I have mostly successfully deployed masakari. The instance monitoring works, I can set up segments in Horizon and via CLI. But I can't get masakari-hostmonitor to start: > > 2021-12-02 16:50:44.999 7 INFO oslo.privsep.daemon [-] Running privsep helper: ['sudo', 'privsep-helper', '--config-file', '/etc/masakari-monitors/masakari > -monitors.conf', '--privsep_context', 'masakarimonitors.privsep.monitors_priv', '--privsep_sock_path', '/tmp/tmpf_65hc7m/privsep.sock'] > 2021-12-02 16:50:44.892 261 INFO oslo.privsep.daemon [-] privsep daemon starting > 2021-12-02 16:50:44.898 261 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0 > 2021-12-02 16:50:44.902 261 ERROR oslo.privsep.daemon [-] [Errno 1] Operation not permitted > Traceback (most recent call last): > File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 593, in helper_main > Daemon(channel, context).run() > File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 403, in run > self._drop_privs() > File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 439, in _drop_privs > capabilities.drop_all_caps_except(self.caps, self.caps, []) > File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/capabilities.py", line 156, in drop_all_caps_except > raise OSError(errno, os.strerror(errno)) > PermissionError: [Errno 1] Operation not permitted > 2021-12-02 16:50:45.001 7 DEBUG oslo_privsep.comm [-] EOF on privsep read channel _reader_main /var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep > /comm.py:149 > 2021-12-02 16:50:45.475 7 WARNING oslo.privsep.daemon [-] privsep log: [Errno 1] Operation not permitted > 2021-12-02 16:50:45.565 7 INFO oslo.privsep.daemon [-] Spawned new privsep daemon via rootwrap > 2021-12-02 16:50:45.565 7 DEBUG oslo.privsep.daemon [-] Accepted privsep connection to /tmp/tmpf_65hc7m/privsep.sock __init__ /var/lib/kolla/venv/lib/pytho > n3.8/site-packages/oslo_privsep/daemon.py:371 > 2021-12-02 16:50:45.467 267 INFO oslo.privsep.daemon [-] privsep daemon starting > 2021-12-02 16:50:45.471 267 INFO oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0 > 2021-12-02 16:50:45.474 267 ERROR oslo.privsep.daemon [-] [Errno 1] Operation not permitted > Traceback (most recent call last): > File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 593, in helper_main > Daemon(channel, context).run() > File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 403, in run > self._drop_privs() > File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", line 439, in _drop_privs > capabilities.drop_all_caps_except(self.caps, self.caps, []) > File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/capabilities.py", line 156, in drop_all_caps_except > raise OSError(errno, os.strerror(errno)) > PermissionError: [Errno 1] Operation not permitted > > sudoers in the container appears to be set up correctly. > > Thanks > Chris From katonalala at gmail.com Thu Dec 2 19:40:22 2021 From: katonalala at gmail.com (Lajos Katona) Date: Thu, 2 Dec 2021 20:40:22 +0100 Subject: [neutron] Drivers meeting agenda - 03.12.2021 Message-ID: Hi Neutron Drivers, The agenda for tomorrow's drivers meeting is at [1]. I would like to discuss one bug which I feel is more an RFE: * https://bugs.launchpad.net/neutron/+bug/1952867 : [ml2][ovs] allow multiple physical networks map to one physical ovs bridge As physnet-bridge mapping is deep in Neutron I feel that a discussion is necessary to have a common understanding of the possible benefits and pitfalls of such shange. I wrote to Liu to ask him to join, but I'm not sure if he can participate. We have one more "On Demand" topic from Rodolfo: * https://review.opendev.org/c/openstack/python-openstackclient/+/819627/1: To deprecate "force" behaviour in OSC "quota" commands. This is currently the default behaviour in Neutron; what is proposed in this patch is to move Neutron to, by default, check the quota limits and use "--force" parameter to commit the quota limits without any check. [1] https://wiki.openstack.org/wiki/Meetings/NeutronDrivers#Agenda See you at the meeting tomorrow. Lajos Katona (lajoskatona) -------------- next part -------------- An HTML attachment was scrubbed... URL: From cdilorenzo at gmail.com Thu Dec 2 20:17:56 2021 From: cdilorenzo at gmail.com (Chris DiLorenzo) Date: Thu, 2 Dec 2021 15:17:56 -0500 Subject: [kolla[ Issue running masakari-hostmonitor In-Reply-To: References: Message-ID: It's not running it privileged mode: ubuntu at control001-poc:~$ docker inspect --format='{{.HostConfig.Privileged}}' masakari_hostmonitor false The config in ansible/roles/masakari/defaults/main.yml looks like this: masakari-hostmonitor: container_name: masakari_hostmonitor group: masakari-hostmonitor enabled: true ipc_mode: host image: "{{ masakari_monitors_image_full }}" volumes: "{{ masakari_hostmonitor_default_volumes + masakari_hostmonitor_extra_volumes }}" dimensions: "{{ masakari_hostmonitor_dimensions }}" The reason I am running 12.0.0 is that the hostmonitor code isn't in 11.x Thanks, Chris On Thu, Dec 2, 2021 at 2:28 PM Rados?aw Piliszek < radoslaw.piliszek at gmail.com> wrote: > Hi Chris, > > as for the issue itself - check if the container is running privileged > (it should). > > As for the other details - Kolla-Ansible 12.x is for Wallaby, not > Victoria, that could be causing issues. > 12.0.0 is not the latest release for Wallaby either (12.2.0 is for today). > Fixing these might help with the other issue. > > -yoctozepto > > On Thu, 2 Dec 2021 at 18:08, Chris DiLorenzo wrote: > > > > Hello, > > > > I am running Kolla-12.0.0 with Openstack Victoria. I have mostly > successfully deployed masakari. The instance monitoring works, I can set > up segments in Horizon and via CLI. But I can't get masakari-hostmonitor > to start: > > > > 2021-12-02 16:50:44.999 7 INFO oslo.privsep.daemon [-] Running privsep > helper: ['sudo', 'privsep-helper', '--config-file', > '/etc/masakari-monitors/masakari > > -monitors.conf', '--privsep_context', > 'masakarimonitors.privsep.monitors_priv', '--privsep_sock_path', > '/tmp/tmpf_65hc7m/privsep.sock'] > > 2021-12-02 16:50:44.892 261 INFO oslo.privsep.daemon [-] privsep daemon > starting > > 2021-12-02 16:50:44.898 261 INFO oslo.privsep.daemon [-] privsep process > running with uid/gid: 0/0 > > 2021-12-02 16:50:44.902 261 ERROR oslo.privsep.daemon [-] [Errno 1] > Operation not permitted > > Traceback (most recent call last): > > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", > line 593, in helper_main > > Daemon(channel, context).run() > > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", > line 403, in run > > self._drop_privs() > > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", > line 439, in _drop_privs > > capabilities.drop_all_caps_except(self.caps, self.caps, []) > > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/capabilities.py", > line 156, in drop_all_caps_except > > raise OSError(errno, os.strerror(errno)) > > PermissionError: [Errno 1] Operation not permitted > > 2021-12-02 16:50:45.001 7 DEBUG oslo_privsep.comm [-] EOF on privsep > read channel _reader_main > /var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep > > /comm.py:149 > > 2021-12-02 16:50:45.475 7 WARNING oslo.privsep.daemon [-] privsep log: > [Errno 1] Operation not permitted > > 2021-12-02 16:50:45.565 7 INFO oslo.privsep.daemon [-] Spawned new > privsep daemon via rootwrap > > 2021-12-02 16:50:45.565 7 DEBUG oslo.privsep.daemon [-] Accepted privsep > connection to /tmp/tmpf_65hc7m/privsep.sock __init__ > /var/lib/kolla/venv/lib/pytho > > n3.8/site-packages/oslo_privsep/daemon.py:371 > > 2021-12-02 16:50:45.467 267 INFO oslo.privsep.daemon [-] privsep daemon > starting > > 2021-12-02 16:50:45.471 267 INFO oslo.privsep.daemon [-] privsep process > running with uid/gid: 0/0 > > 2021-12-02 16:50:45.474 267 ERROR oslo.privsep.daemon [-] [Errno 1] > Operation not permitted > > Traceback (most recent call last): > > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", > line 593, in helper_main > > Daemon(channel, context).run() > > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", > line 403, in run > > self._drop_privs() > > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", > line 439, in _drop_privs > > capabilities.drop_all_caps_except(self.caps, self.caps, []) > > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/capabilities.py", > line 156, in drop_all_caps_except > > raise OSError(errno, os.strerror(errno)) > > PermissionError: [Errno 1] Operation not permitted > > > > sudoers in the container appears to be set up correctly. > > > > Thanks > > Chris > -------------- next part -------------- An HTML attachment was scrubbed... URL: From zhangbailin at inspur.com Fri Dec 3 01:45:10 2021 From: zhangbailin at inspur.com (=?gb2312?B?QnJpbiBaaGFuZyjVxbDZwdYp?=) Date: Fri, 3 Dec 2021 01:45:10 +0000 Subject: [cyborg] Cancel IRC meeting Message-ID: Hi all, Due to the conflict between the team members and the IRC meeting time, the IRC meeting of this week will be cancelled. If you have any questions, please feel free to ping us through ML or #openstack-cyborg channel. Thanks. brinzhang Inspur Electronic Information Industry Co.,Ltd. -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Fri Dec 3 02:19:47 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Thu, 02 Dec 2021 20:19:47 -0600 Subject: python_requires >= 3.8 during Yoga In-Reply-To: <17d7262e0ac.be45cbf0116583.6975429409607560951@ghanshyammann.com> References: <17d58f75012.106d120141328715.5251672452583667038@ghanshyammann.com> <20211126122942.deyi4tymy5xoodvz@lyarwood-laptop.usersys.redhat.com> <17d5ce532e1.eabdb7481381896.7024061515184177274@ghanshyammann.com> <20211126160515.4tvksnqcz3vaomf4@lyarwood-laptop.usersys.redhat.com> <17d5d0a4f3f.dc8bb48c1384810.8642013929754348916@ghanshyammann.com> <17d5d476490.e30e73db1388640.4406614125429012751@ghanshyammann.com> <20211129091718.fdgdwmowtp3jxf6u@lyarwood-laptop.usersys.redhat.com> <1319EEE3-AB75-495A-8EEF-E2985A4FBB10@binero.com> <17d7262e0ac.be45cbf0116583.6975429409607560951@ghanshyammann.com> Message-ID: <17d7e172c07.102013b53263302.7784256053937335636@ghanshyammann.com> ---- On Tue, 30 Nov 2021 13:47:02 -0600 Ghanshyam Mann wrote ---- > ---- On Mon, 29 Nov 2021 04:52:07 -0600 Tobias Urdin wrote ---- > > Hello,I agree with previous statement from Michal. > > > > The upgrade path in for example RDO has been very smooth previously with upgrading to new OpenStack releasethen switching out the distro version afterwards because they support both during a transition. > > Sure they will do that now as well, but if any project decides to break py36 there will be more work for the RDO teamand more arguments to not revert changes because py38 would be the ?real? supported runtime in OpenStack testing. > > The transition period is required to not break the upgrade path. > > Best regardsTobias > > On 29 Nov 2021, at 11:14, Micha? Nasiadka wrote: > > Hello, > > I?m strongly against dropping py36 support now, unless we?re going to find a solution that works on CentOS Stream 8.RHEL 9 is not out, and probably will not be in months - how do we expect users to use Yoga on production deployments (where they use CentOS Linux/equivalents today)? > > Dropping the runtime testing and supporting devstack - and negotiating on a per project basis to support py36 or not - is not a solution.Either Yoga supports py36 as a transition point/release to py38 - or not. > > In Kolla - we also did not anticipate (and don?t think it?s a good idea) to support CentOS Stream 9 in Yoga release.With the current decision - we are either forced with supporting CentOS Stream 9 (with no alternatives like Rocky Linux/Alma Linux in place - because RHEL 9 is not out) - or dropping CentOS support completely. > > If we pursue CS9 - we also need to support migration from CS8 to CS9 and that?s also a considerable amount of work - which is unplanned. > > Best regards,Michal > > I agree on all these points on keeping py36 for Yoga can be helpful for most people in migration to the new distro version. > I have added it to the TC meeting agenda which is scheduled for Dec 2nd and discuss more on this. Please feel free > to join TC meeting, details are below: > > https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Agenda_Suggestions The Technical Committee considered all the points made in this ML thread and discussed them in today's TC meeting[1]. Testing python 3.6 for one more cycle would not cost much. To have a smooth migration to new distro versions having different default python versions, we are ok to keep the python 3.6 testing for the Yoga cycle and keep CentOS Stream 8 and 9 both in runtime. This does not cost us much in terms of developer-power or infra resources. Below are the versions we will target for Yoga: Distro: * Ubuntu 20.04 * CentOS Stream 8 * CentOS Stream 9 * Debian 11 Python: (keeping only lower and highest versions in testing with the assumption that anything working in py3.6 and py3.9 will work in py3.7, py3.8 too) * Python 3.6 * Python 3.9 * We agree to add Python 3.10 as a non-voting job also. I have pushed the proposal in governance: - https://review.opendev.org/c/openstack/governance/+/820195 as well as in job template: - https://review.opendev.org/c/openstack/openstack-zuul-jobs/+/820286 Thanks to everyone for participating and providing more clarity/feedback from users/operator/distro-provide perspective. [1] - https://meetings.opendev.org/meetings/tc/2021/tc.2021-12-02-15.02.log.html#l-33 - https://www.youtube.com/watch?v=PwGYse8j4UY -gmann > > -gmann > > > > > On 29 Nov 2021, at 10:17, Lee Yarwood wrote: > > On 26-11-21 11:24:59, Ghanshyam Mann wrote: > > ---- On Fri, 26 Nov 2021 10:18:16 -0600 Ghanshyam Mann wrote ---- > > ---- On Fri, 26 Nov 2021 10:05:15 -0600 Lee Yarwood wrote ---- > > On 26-11-21 09:37:44, Ghanshyam Mann wrote: > > ---- On Fri, 26 Nov 2021 06:29:42 -0600 Lee Yarwood wrote ---- > > On 26-11-21 10:54:26, Alfredo Moralejo Alonso wrote: > > On Thu, Nov 25, 2021 at 10:23 PM Ghanshyam Mann > > wrote: > > > > ---- On Thu, 25 Nov 2021 13:58:28 -0600 Marcin Juszkiewicz < > > marcin.juszkiewicz at linaro.org> wrote ---- > > W dniu 25.11.2021 o 19:13, Stephen Finucane pisze: > > gmann has been helpfully proposing patches to change the > > versions of Python we're testing against in Yoga. I've > > suggested that we might want to bump 'python_requires' in > > 'setup.cfg' to indicate that we no longer support any version > > of Python before 3.8 > > > > CentOS Stream 8 has Python 3.6 by default and RDO team is doing > > CS8 -> CS9 migration during Yoga cycle. Can we postpone it to Z > > when there will be no distribution with Py 3.6 to care about? > > > > Stupid question that I should know the answer to but does RDO really > > support RPM based installations anymore? IOW couldn't we just workaround > > this by providing CS8 py38 based containers during the upgrade? > > > > As Marcin posted, the plan in RDO is to support both CentOS Stream 8 and > > CentOS Stream 9 in Yoga. This is how we have managed previous major CentOS > > version upgrades in the past providing support for both releases in an > > OpenStack version to ease the upgrade so I'd like to keep yoga working on > > py3.6 included in CS8 and CS9. > > > > If this was the plan why wasn't it made clear to the TC before they > > dropped CS8 from the Yoga runtimes? Would it even be possible for the TC > > to add CS8 and py36 back in to the Yoga runtimes? > > > > Postponing to Z, you mean dropping the py3.6 tests or bumping it in > > in 'setup.cfg' so that no one can install on py3.6 ? > > > > First one we already did and as per Yoga testing runtime we are > > targeting centos9-stream[1] in Yoga itself. > > > > For making 'python_requires' >=py3.8 in 'setup.cfg', I have no > > string opinion on this but I prefer to have flexible here that 'yes > > OpenStack is installable in py3.6 but we do not test it anymore from > > Yoga onwards so no guarantee'. Our testing runtime main goal is > > that we document the version we are testing *at least* which means > > it can work on lower or higher versions too but we just do not test > > them. > > > > > > May it be possible to keep py3.6 jobs to make sure patches are not > > introducing py3.8-only features that would break deployment in CS8? > > > > We should keep CS8 and py36 as supported runtimes if we are keeping the > > jobs, otherwise this just sets super confusing. > > > > Yeah, I think it create confusion as I can see in this ML thread so > > agree on keeping 'python_requires' also in sycn with what we test. > > > > Cool thanks! > > > > Now question on going back to centos stream 8 support in Yoga, is it > > not centos stream 9 is stable released or is it experimental only? If > > stable then we can keep the latest available version which can be > > centos stream 9. > > > > I honestly don't know and can't find any docs to point to. > > > > Our project interface testing doc clearly stats 'latest LTS' to > > consider for testing[1] whenever we are ready. I am not very strongly > > against of reverting back to centos stream 8 but we should not add two > > version of same distro in testing which can be a lot of we consider > > below three distro > > > > How do we expect operators to upgrade between Xena where CentOS 8 stream > > is a supported runtime and Yoga where CentOS 9 stream is currently the > > equivalent supported runtime without supporting both for a single > > release? > > > > This is really good question on upgrade testing we do at upstream and I remember > > it cameup and discussed a lot during py2.7 drop also that how we are testing the > > upgrade from py2.7 to py3. Can we do in grenade? But that we answered as we did > > not tested directly but stein and train tested both version so should not be any issue > > if you upgrade from there (one of FAQ in my blog[1]). > > > > But on distro upgrade testing, as you know we do not test those in upstream neither > > in grenade where upgrade are done on old node distro only not from old distro version to > > new distro version with new code. It is not like we do not want to test but if anyone > > from any distro would like to setup grenade for that and maintain then we are more happy. > > In summary, yes we cannot guarantee distro upgrade testing from OpenStack upstream testing > > due to resource bandwidth issue but we will welcome any help here. > > > > We discussed with amoralej about moving the testing runtime to CentOS > > stream 8 and py36 or not in TC IRC channel[1]. > > > > As we at upstream do not test distro two versions in same release, > > amoralej agreed to keep CentOS stream 9 if one to choose which is our > > current testing runtime is. So no change in the direction of current > > testing runtime and dropping the py3.6 but there is possibility of > > some trade off here. If any py3.6 breaking changes are happening then > > it is up to projects goodness, bandwidth, or flexibility about > > accepting the fix or not or even add a py36 unit test job. As our > > testing runtime is the minimum things to test and it does not put any > > max limit of testing, any project can extend their testing as per > > their bandwidth. > > > > In summary: > > > > (This is what we agreed today in TC channel but as most of the folks > > are on leave today, I will keep it open until next week so see if any > > objections from the community and will conclude it accordingly) > > > > * No change in Yoga testing runtime and we move to cs9 and drop py36. > > * We will not put hard stop on cs8 support and we can: > > ** Devstack keep supporting cs8 in Yoga > > ** It can be negotiated with project to add py36 job or fix if any > > py36 breaking changes are observed by RDO (or any distro interested in > > py36) but it depends on the project decision and bandwidth. > > > > As next, how we can improve the upgrade testing from distro versions > > is something we will explore next and see what all we can test to make > > upgrade easier. > > > > I'm against this, as I said in my setup.cfg >= py38 review for > > openstack/nova [1] we either list and support runtimes or don't. If RDO > > and others need CentOS 8 Stream support for a release then lets include > > it and py36 still for Yoga and make things explicit. > > > > As I've said elsewhere I think the TC really need to adjust their > > thinking on this topic and allow for one OpenStack release where both > > the old and new LTS distro release are supported. Ensuring we allow > > people to actually upgrade in place and later handle the distro upgrade > > itself. > > > > Cheers, > > > > Lee > > > > [1] https://review.opendev.org/c/openstack/nova/+/819415 > > > > -- > > Lee Yarwood A5D1 9385 88CB 7E5F BE64 6618 BCA6 6E33 F672 2D76 > > > > > > From lokendrarathour at gmail.com Fri Dec 3 05:37:07 2021 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Fri, 3 Dec 2021 11:07:07 +0530 Subject: [TripleO] IPV6 Support In-Reply-To: References: Message-ID: Hi Alex, we are trying on centos 7/8, and similar commands for centos 7/8 is not working. Can you please suggest related commands for the same? we have tried with yum/dnf in centos 7/8. Just to confirm again we are using the stein release on centos 7 where we want to use ipv6 or dual-stack if possible. Thanks once again for your help. -Lokendra On Thu, Nov 25, 2021 at 11:18 PM Alex Schultz wrote: > Yes IPV6 is supported, however the error you provided indicates problems > starting containers. Make sure you pin to container-tools:3.0 to ensure > you get the version we expect. > > dnf module -y disable container-tools:rhel8 ; > dnf module -y enable container-tools:3.0 > > On Thu, Nov 25, 2021 at 1:02 AM Lokendra Rathour < > lokendrarathour at gmail.com> wrote: > >> Hi, >> I am trying to install the undercloud using the ipv6 address on Centos 8 >> using Train release. >> It is seen that the deployment of undercloud is getting failed with error >> as mentioned below. Same deployment is working in the normal ipv4 case. >> So Questions: >> >> 1. is IPV6 supported in TripleO , if yes then please suggest how. >> >> >> Error: >> nied", >> " attempt(s): 2", >> "2021-11-25 07:25:31,535 WARNING: 65355 -- Retrying running >> container: zaqar", >> "2021-11-25 07:25:34,887 ERROR: 65355 -- ['/usr/bin/podman', 'run', >> '--user', '0', '--name', 'container-puppet-zaqar', '--env', >> 'PUPPET_TAGS=file,file_line,concat,augeas,cron,zaqar_config', '--env', >> 'NAME=zaqar', '--env', 'HOSTNAME=undercloud', '--env', 'NO_ARCHIVE=', >> '--env', 'STEP=6', '--env', 'NET_HOST=true', '--env', 'DEBUG=False', >> '--volume', '/etc/localtime:/etc/localtime:ro', '--volume', >> '/tmp/tmp06nvxzzz:/etc/config.pp:ro', '--volume', >> '/etc/puppet/:/tmp/puppet-etc/:ro', '--volume', >> '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '--volume', >> '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', >> '--volume', >> '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', >> '--volume', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '--volume', >> '/var/lib/config-data:/var/lib/config-data/:rw', '--volume', >> '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', >> '--volume', '/var/lib/container-puppet/puppetlabs/:/opt/puppetlabs/:ro', >> '--volume', '/dev/log:/dev/log:rw', '--rm', '--log-driver', 'k8s-file', >> '--log-opt', 'path=/var/log/containers/stdouts/container-puppet-zaqar.log', >> '--security-opt', 'label=disable', '--volume', >> '/usr/share/openstack-puppet/modules/:/usr/share/openstack-puppet/modules/:ro', >> '--entrypoint', '/var/lib/container-puppet/container-puppet.sh', '--net', >> 'host', '--volume', '/etc/hosts:/etc/hosts:ro', '--volume', >> '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', >> 'undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-zaqar-wsgi:current-tripleo'] >> run failed after Error: container_linux.go:370: starting container process >> caused: error adding seccomp filter rule for syscall bdflush: permission >> denied: OCI permission denied", >> " attempt(s): 3", >> "2021-11-25 07:25:34,888 WARNING: 65355 -- Retrying running >> container: zaqar", >> "2021-11-25 07:25:34,888 ERROR: 65355 -- Failed running container for >> zaqar", >> "2021-11-25 07:25:34,888 INFO: 65355 -- Finished processing puppet >> configs for zaqar", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring crond", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >> glance_api", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring heat_api", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring heat", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >> ironic_api", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring ironic", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >> ironic_inspector", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring neutron", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring iscsid", >> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring keystone", >> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring memcached", >> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring mistral", >> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring mysql", >> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring nova", >> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring rabbitmq", >> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring placement", >> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring swift", >> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >> swift_ringbuilder", >> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring zaqar" >> ], >> "stderr_lines": [], >> "_ansible_no_log": false, >> "attempts": 15 >> } >> ] >> ] >> Not cleaning working directory >> /home/stack/tripleo-heat-installer-templates >> Not cleaning ansible directory /home/stack/undercloud-ansible-mw4crw92 >> Install artifact is located at >> /home/stack/undercloud-install-20211125072536.tar.bzip2 >> >> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >> >> Deployment Failed! >> >> ERROR: Heat log files: /var/log/heat-launcher/undercloud_deploy-o_qf1b4w >> >> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >> >> Deployment failed. >> Traceback (most recent call last): >> File >> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >> 1345, in _standalone_deploy >> raise exceptions.DeploymentError('Deployment failed') >> tripleoclient.exceptions.DeploymentError: Deployment failed >> >> During handling of the above exception, another exception occurred: >> >> Traceback (most recent call last): >> File >> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >> 1363, in _standalone_deploy >> raise exceptions.DeploymentError(six.text_type(e)) >> tripleoclient.exceptions.DeploymentError: Deployment failed >> >> During handling of the above exception, another exception occurred: >> >> Traceback (most recent call last): >> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 401, in >> run_subcommand >> result = cmd.run(parsed_args) >> File "/usr/lib/python3.6/site-packages/cliff/command.py", line 185, in >> run >> return_code = self.take_action(parsed_args) or 0 >> File >> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >> 1451, in take_action >> if self._standalone_deploy(parsed_args) != 0: >> File >> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >> 1400, in _standalone_deploy >> raise exceptions.DeploymentError('Deployment failed.') >> tripleoclient.exceptions.DeploymentError: Deployment failed. >> clean_up Deploy: Deployment failed. >> Traceback (most recent call last): >> File >> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >> 1345, in _standalone_deploy >> raise exceptions.DeploymentError('Deployment failed') >> tripleoclient.exceptions.DeploymentError: Deployment failed >> >> During handling of the above exception, another exception occurred: >> >> Traceback (most recent call last): >> File >> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >> 1363, in _standalone_deploy >> raise exceptions.DeploymentError(six.text_type(e)) >> tripleoclient.exceptions.DeploymentError: Deployment failed >> >> During handling of the above exception, another exception occurred: >> >> Traceback (most recent call last): >> File "/usr/lib/python3.6/site-packages/osc_lib/shell.py", line 136, in >> run >> ret_val = super(OpenStackShell, self).run(argv) >> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 281, in run >> result = self.run_subcommand(remainder) >> File "/usr/lib/python3.6/site-packages/osc_lib/shell.py", line 176, in >> run_subcommand >> ret_value = super(OpenStackShell, self).run_subcommand(argv) >> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 401, in >> run_subcommand >> result = cmd.run(parsed_args) >> File "/usr/lib/python3.6/site-packages/cliff/command.py", line 185, in >> run >> return_code = self.take_action(parsed_args) or 0 >> File >> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >> 1451, in take_action >> if self._standalone_deploy(parsed_args) != 0: >> File >> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >> 1400, in _standalone_deploy >> raise exceptions.DeploymentError('Deployment failed.') >> tripleoclient.exceptions.DeploymentError: Deployment failed. >> >> END return value: 1 >> >> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >> >> An error has occured while deploying the Undercloud. >> >> See the previous output for details about what went wrong. >> >> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >> >> Command '['sudo', '--preserve-env', 'openstack', 'tripleo', 'deploy', >> '--standalone', '--standalone-role', 'Undercloud', '--stack', 'undercloud', >> '--local-domain=localdomain', '--local-ip=abce:abce:abce::1/64', >> '--templates=/usr/share/openstack-tripleo-heat-templates/', >> '--networks-file=network_data_undercloud.yaml', '--heat-native', '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/undercloud.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/use-dns-for-vips.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/podman.yaml', >> '-e', '/home/stack/containers-prepare-parameter.yaml', '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/mistral.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/zaqar-swift-backend.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/disable-telemetry.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/tempest.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/ssl/no-tls-endpoints-public-ip.yaml', >> '--deployment-user', 'stack', '--output-dir=/home/stack', '-e', >> '/home/stack/tripleo-config-generated-env-files/undercloud_parameters.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/tripleo-validations.yaml', >> '--debug', '--log-file=install-undercloud.log', '-e', >> '/usr/share/openstack-tripleo-heat-templates/undercloud-stack-vstate-dropin.yaml']' >> returned non-zero exit status 1. >> Command '['sudo', '--preserve-env', 'openstack', 'tripleo', 'deploy', >> '--standalone', '--standalone-role', 'Undercloud', '--stack', 'undercloud', >> '--local-domain=localdomain', '--local-ip=abce:abce:abce::1/64', >> '--templates=/usr/share/openstack-tripleo-heat-templates/', >> '--networks-file=network_data_undercloud.yaml', '--heat-native', '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/undercloud.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/use-dns-for-vips.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/podman.yaml', >> '-e', '/home/stack/containers-prepare-parameter.yaml', '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/mistral.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/zaqar-swift-backend.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/disable-telemetry.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/services/tempest.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/ssl/no-tls-endpoints-public-ip.yaml', >> '--deployment-user', 'stack', '--output-dir=/home/stack', '-e', >> '/home/stack/tripleo-config-generated-env-files/undercloud_parameters.yaml', >> '-e', >> '/usr/share/openstack-tripleo-heat-templates/environments/tripleo-validations.yaml', >> '--debug', '--log-file=install-undercloud.log', '-e', >> '/usr/share/openstack-tripleo-heat-templates/undercloud-stack-vstate-dropin.yaml']' >> returned non-zero exit status 1. >> END return value: 1 >> [stack at undercloud ~]$ >> >> >> -- -------------- next part -------------- An HTML attachment was scrubbed... URL: From aschultz at redhat.com Fri Dec 3 05:57:49 2021 From: aschultz at redhat.com (Alex Schultz) Date: Thu, 2 Dec 2021 22:57:49 -0700 Subject: [TripleO] IPV6 Support In-Reply-To: References: Message-ID: The commands I provided were for Train on CentOS 8. They were not related to the original topic of the email (ipv6) but rather the container startup issues. Stein is no longer supported and it would not be recommended to use it. On Thu, Dec 2, 2021 at 10:36 PM Lokendra Rathour wrote: > Hi Alex, > we are trying on centos 7/8, and similar commands for centos 7/8 is not > working. Can you please suggest related commands for the same? > we have tried with yum/dnf in centos 7/8. > > Just to confirm again we are using the stein release on centos 7 where we > want to use ipv6 or dual-stack if possible. > Thanks once again for your help. > > -Lokendra > > On Thu, Nov 25, 2021 at 11:18 PM Alex Schultz wrote: > >> Yes IPV6 is supported, however the error you provided indicates problems >> starting containers. Make sure you pin to container-tools:3.0 to ensure >> you get the version we expect. >> >> dnf module -y disable container-tools:rhel8 ; >> dnf module -y enable container-tools:3.0 >> >> On Thu, Nov 25, 2021 at 1:02 AM Lokendra Rathour < >> lokendrarathour at gmail.com> wrote: >> >>> Hi, >>> I am trying to install the undercloud using the ipv6 address on Centos 8 >>> using Train release. >>> It is seen that the deployment of undercloud is getting failed with >>> error as mentioned below. Same deployment is working in the normal ipv4 >>> case. >>> So Questions: >>> >>> 1. is IPV6 supported in TripleO , if yes then please suggest how. >>> >>> >>> Error: >>> nied", >>> " attempt(s): 2", >>> "2021-11-25 07:25:31,535 WARNING: 65355 -- Retrying running >>> container: zaqar", >>> "2021-11-25 07:25:34,887 ERROR: 65355 -- ['/usr/bin/podman', 'run', >>> '--user', '0', '--name', 'container-puppet-zaqar', '--env', >>> 'PUPPET_TAGS=file,file_line,concat,augeas,cron,zaqar_config', '--env', >>> 'NAME=zaqar', '--env', 'HOSTNAME=undercloud', '--env', 'NO_ARCHIVE=', >>> '--env', 'STEP=6', '--env', 'NET_HOST=true', '--env', 'DEBUG=False', >>> '--volume', '/etc/localtime:/etc/localtime:ro', '--volume', >>> '/tmp/tmp06nvxzzz:/etc/config.pp:ro', '--volume', >>> '/etc/puppet/:/tmp/puppet-etc/:ro', '--volume', >>> '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '--volume', >>> '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', >>> '--volume', >>> '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', >>> '--volume', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '--volume', >>> '/var/lib/config-data:/var/lib/config-data/:rw', '--volume', >>> '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', >>> '--volume', '/var/lib/container-puppet/puppetlabs/:/opt/puppetlabs/:ro', >>> '--volume', '/dev/log:/dev/log:rw', '--rm', '--log-driver', 'k8s-file', >>> '--log-opt', 'path=/var/log/containers/stdouts/container-puppet-zaqar.log', >>> '--security-opt', 'label=disable', '--volume', >>> '/usr/share/openstack-puppet/modules/:/usr/share/openstack-puppet/modules/:ro', >>> '--entrypoint', '/var/lib/container-puppet/container-puppet.sh', '--net', >>> 'host', '--volume', '/etc/hosts:/etc/hosts:ro', '--volume', >>> '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', >>> 'undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-zaqar-wsgi:current-tripleo'] >>> run failed after Error: container_linux.go:370: starting container process >>> caused: error adding seccomp filter rule for syscall bdflush: permission >>> denied: OCI permission denied", >>> " attempt(s): 3", >>> "2021-11-25 07:25:34,888 WARNING: 65355 -- Retrying running >>> container: zaqar", >>> "2021-11-25 07:25:34,888 ERROR: 65355 -- Failed running container >>> for zaqar", >>> "2021-11-25 07:25:34,888 INFO: 65355 -- Finished processing puppet >>> configs for zaqar", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring crond", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>> glance_api", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring heat_api", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring heat", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>> ironic_api", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring ironic", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>> ironic_inspector", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring neutron", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring iscsid", >>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring keystone", >>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>> memcached", >>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring mistral", >>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring mysql", >>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring nova", >>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring rabbitmq", >>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>> placement", >>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring swift", >>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>> swift_ringbuilder", >>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring zaqar" >>> ], >>> "stderr_lines": [], >>> "_ansible_no_log": false, >>> "attempts": 15 >>> } >>> ] >>> ] >>> Not cleaning working directory >>> /home/stack/tripleo-heat-installer-templates >>> Not cleaning ansible directory /home/stack/undercloud-ansible-mw4crw92 >>> Install artifact is located at >>> /home/stack/undercloud-install-20211125072536.tar.bzip2 >>> >>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>> >>> Deployment Failed! >>> >>> ERROR: Heat log files: /var/log/heat-launcher/undercloud_deploy-o_qf1b4w >>> >>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>> >>> Deployment failed. >>> Traceback (most recent call last): >>> File >>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>> 1345, in _standalone_deploy >>> raise exceptions.DeploymentError('Deployment failed') >>> tripleoclient.exceptions.DeploymentError: Deployment failed >>> >>> During handling of the above exception, another exception occurred: >>> >>> Traceback (most recent call last): >>> File >>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>> 1363, in _standalone_deploy >>> raise exceptions.DeploymentError(six.text_type(e)) >>> tripleoclient.exceptions.DeploymentError: Deployment failed >>> >>> During handling of the above exception, another exception occurred: >>> >>> Traceback (most recent call last): >>> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 401, in >>> run_subcommand >>> result = cmd.run(parsed_args) >>> File "/usr/lib/python3.6/site-packages/cliff/command.py", line 185, in >>> run >>> return_code = self.take_action(parsed_args) or 0 >>> File >>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>> 1451, in take_action >>> if self._standalone_deploy(parsed_args) != 0: >>> File >>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>> 1400, in _standalone_deploy >>> raise exceptions.DeploymentError('Deployment failed.') >>> tripleoclient.exceptions.DeploymentError: Deployment failed. >>> clean_up Deploy: Deployment failed. >>> Traceback (most recent call last): >>> File >>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>> 1345, in _standalone_deploy >>> raise exceptions.DeploymentError('Deployment failed') >>> tripleoclient.exceptions.DeploymentError: Deployment failed >>> >>> During handling of the above exception, another exception occurred: >>> >>> Traceback (most recent call last): >>> File >>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>> 1363, in _standalone_deploy >>> raise exceptions.DeploymentError(six.text_type(e)) >>> tripleoclient.exceptions.DeploymentError: Deployment failed >>> >>> During handling of the above exception, another exception occurred: >>> >>> Traceback (most recent call last): >>> File "/usr/lib/python3.6/site-packages/osc_lib/shell.py", line 136, in >>> run >>> ret_val = super(OpenStackShell, self).run(argv) >>> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 281, in run >>> result = self.run_subcommand(remainder) >>> File "/usr/lib/python3.6/site-packages/osc_lib/shell.py", line 176, in >>> run_subcommand >>> ret_value = super(OpenStackShell, self).run_subcommand(argv) >>> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 401, in >>> run_subcommand >>> result = cmd.run(parsed_args) >>> File "/usr/lib/python3.6/site-packages/cliff/command.py", line 185, in >>> run >>> return_code = self.take_action(parsed_args) or 0 >>> File >>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>> 1451, in take_action >>> if self._standalone_deploy(parsed_args) != 0: >>> File >>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>> 1400, in _standalone_deploy >>> raise exceptions.DeploymentError('Deployment failed.') >>> tripleoclient.exceptions.DeploymentError: Deployment failed. >>> >>> END return value: 1 >>> >>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>> >>> An error has occured while deploying the Undercloud. >>> >>> See the previous output for details about what went wrong. >>> >>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>> >>> Command '['sudo', '--preserve-env', 'openstack', 'tripleo', 'deploy', >>> '--standalone', '--standalone-role', 'Undercloud', '--stack', 'undercloud', >>> '--local-domain=localdomain', '--local-ip=abce:abce:abce::1/64', >>> '--templates=/usr/share/openstack-tripleo-heat-templates/', >>> '--networks-file=network_data_undercloud.yaml', '--heat-native', '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/undercloud.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/use-dns-for-vips.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/podman.yaml', >>> '-e', '/home/stack/containers-prepare-parameter.yaml', '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/mistral.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/zaqar-swift-backend.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/disable-telemetry.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/tempest.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/ssl/no-tls-endpoints-public-ip.yaml', >>> '--deployment-user', 'stack', '--output-dir=/home/stack', '-e', >>> '/home/stack/tripleo-config-generated-env-files/undercloud_parameters.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/tripleo-validations.yaml', >>> '--debug', '--log-file=install-undercloud.log', '-e', >>> '/usr/share/openstack-tripleo-heat-templates/undercloud-stack-vstate-dropin.yaml']' >>> returned non-zero exit status 1. >>> Command '['sudo', '--preserve-env', 'openstack', 'tripleo', 'deploy', >>> '--standalone', '--standalone-role', 'Undercloud', '--stack', 'undercloud', >>> '--local-domain=localdomain', '--local-ip=abce:abce:abce::1/64', >>> '--templates=/usr/share/openstack-tripleo-heat-templates/', >>> '--networks-file=network_data_undercloud.yaml', '--heat-native', '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/undercloud.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/use-dns-for-vips.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/podman.yaml', >>> '-e', '/home/stack/containers-prepare-parameter.yaml', '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/mistral.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/zaqar-swift-backend.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/disable-telemetry.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/services/tempest.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/ssl/no-tls-endpoints-public-ip.yaml', >>> '--deployment-user', 'stack', '--output-dir=/home/stack', '-e', >>> '/home/stack/tripleo-config-generated-env-files/undercloud_parameters.yaml', >>> '-e', >>> '/usr/share/openstack-tripleo-heat-templates/environments/tripleo-validations.yaml', >>> '--debug', '--log-file=install-undercloud.log', '-e', >>> '/usr/share/openstack-tripleo-heat-templates/undercloud-stack-vstate-dropin.yaml']' >>> returned non-zero exit status 1. >>> END return value: 1 >>> [stack at undercloud ~]$ >>> >>> >>> > > -- > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From lokendrarathour at gmail.com Fri Dec 3 06:59:09 2021 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Fri, 3 Dec 2021 12:29:09 +0530 Subject: [TripleO] IPV6 Support In-Reply-To: References: Message-ID: Hi Alex, >From your message I could understand that Stein is no longer supported, but does it support ipv6, please advise. We are also using stein to test the Upgrade feature of Triple0. We can start with Train as well and install and upgrade to Ussurie. But ipv6 support is something that we find a miss in the undercloud.conf or somewhere we can specify ipv6 details. Request you to please share any document/link which helps us try train on centos 8 with ipv6. For now, I am referring to this document. Undercloud Installation ? TripleO 3.0.0 documentation (openstack.org) Thanks again for your quick response. Your response will help me decide things at the initial point only. -Lokendra On Fri, Dec 3, 2021 at 11:28 AM Alex Schultz wrote: > The commands I provided were for Train on CentOS 8. They were not related > to the original topic of the email (ipv6) but rather the container startup > issues. Stein is no longer supported and it would not be recommended to > use it. > > On Thu, Dec 2, 2021 at 10:36 PM Lokendra Rathour < > lokendrarathour at gmail.com> wrote: > >> Hi Alex, >> we are trying on centos 7/8, and similar commands for centos 7/8 is not >> working. Can you please suggest related commands for the same? >> we have tried with yum/dnf in centos 7/8. >> >> Just to confirm again we are using the stein release on centos 7 where we >> want to use ipv6 or dual-stack if possible. >> Thanks once again for your help. >> >> -Lokendra >> >> On Thu, Nov 25, 2021 at 11:18 PM Alex Schultz >> wrote: >> >>> Yes IPV6 is supported, however the error you provided indicates problems >>> starting containers. Make sure you pin to container-tools:3.0 to ensure >>> you get the version we expect. >>> >>> dnf module -y disable container-tools:rhel8 ; >>> dnf module -y enable container-tools:3.0 >>> >>> On Thu, Nov 25, 2021 at 1:02 AM Lokendra Rathour < >>> lokendrarathour at gmail.com> wrote: >>> >>>> Hi, >>>> I am trying to install the undercloud using the ipv6 address on Centos >>>> 8 using Train release. >>>> It is seen that the deployment of undercloud is getting failed with >>>> error as mentioned below. Same deployment is working in the normal ipv4 >>>> case. >>>> So Questions: >>>> >>>> 1. is IPV6 supported in TripleO , if yes then please suggest how. >>>> >>>> >>>> Error: >>>> nied", >>>> " attempt(s): 2", >>>> "2021-11-25 07:25:31,535 WARNING: 65355 -- Retrying running >>>> container: zaqar", >>>> "2021-11-25 07:25:34,887 ERROR: 65355 -- ['/usr/bin/podman', 'run', >>>> '--user', '0', '--name', 'container-puppet-zaqar', '--env', >>>> 'PUPPET_TAGS=file,file_line,concat,augeas,cron,zaqar_config', '--env', >>>> 'NAME=zaqar', '--env', 'HOSTNAME=undercloud', '--env', 'NO_ARCHIVE=', >>>> '--env', 'STEP=6', '--env', 'NET_HOST=true', '--env', 'DEBUG=False', >>>> '--volume', '/etc/localtime:/etc/localtime:ro', '--volume', >>>> '/tmp/tmp06nvxzzz:/etc/config.pp:ro', '--volume', >>>> '/etc/puppet/:/tmp/puppet-etc/:ro', '--volume', >>>> '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '--volume', >>>> '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', >>>> '--volume', >>>> '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', >>>> '--volume', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '--volume', >>>> '/var/lib/config-data:/var/lib/config-data/:rw', '--volume', >>>> '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', >>>> '--volume', '/var/lib/container-puppet/puppetlabs/:/opt/puppetlabs/:ro', >>>> '--volume', '/dev/log:/dev/log:rw', '--rm', '--log-driver', 'k8s-file', >>>> '--log-opt', 'path=/var/log/containers/stdouts/container-puppet-zaqar.log', >>>> '--security-opt', 'label=disable', '--volume', >>>> '/usr/share/openstack-puppet/modules/:/usr/share/openstack-puppet/modules/:ro', >>>> '--entrypoint', '/var/lib/container-puppet/container-puppet.sh', '--net', >>>> 'host', '--volume', '/etc/hosts:/etc/hosts:ro', '--volume', >>>> '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', >>>> 'undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-zaqar-wsgi:current-tripleo'] >>>> run failed after Error: container_linux.go:370: starting container process >>>> caused: error adding seccomp filter rule for syscall bdflush: permission >>>> denied: OCI permission denied", >>>> " attempt(s): 3", >>>> "2021-11-25 07:25:34,888 WARNING: 65355 -- Retrying running >>>> container: zaqar", >>>> "2021-11-25 07:25:34,888 ERROR: 65355 -- Failed running container >>>> for zaqar", >>>> "2021-11-25 07:25:34,888 INFO: 65355 -- Finished processing puppet >>>> configs for zaqar", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring crond", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>> glance_api", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>> heat_api", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring heat", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>> ironic_api", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring ironic", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>> ironic_inspector", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring neutron", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring iscsid", >>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>> keystone", >>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>>> memcached", >>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring mistral", >>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring mysql", >>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring nova", >>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>>> rabbitmq", >>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>>> placement", >>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring swift", >>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>>> swift_ringbuilder", >>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring zaqar" >>>> ], >>>> "stderr_lines": [], >>>> "_ansible_no_log": false, >>>> "attempts": 15 >>>> } >>>> ] >>>> ] >>>> Not cleaning working directory >>>> /home/stack/tripleo-heat-installer-templates >>>> Not cleaning ansible directory /home/stack/undercloud-ansible-mw4crw92 >>>> Install artifact is located at >>>> /home/stack/undercloud-install-20211125072536.tar.bzip2 >>>> >>>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>>> >>>> Deployment Failed! >>>> >>>> ERROR: Heat log files: /var/log/heat-launcher/undercloud_deploy-o_qf1b4w >>>> >>>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>>> >>>> Deployment failed. >>>> Traceback (most recent call last): >>>> File >>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>> 1345, in _standalone_deploy >>>> raise exceptions.DeploymentError('Deployment failed') >>>> tripleoclient.exceptions.DeploymentError: Deployment failed >>>> >>>> During handling of the above exception, another exception occurred: >>>> >>>> Traceback (most recent call last): >>>> File >>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>> 1363, in _standalone_deploy >>>> raise exceptions.DeploymentError(six.text_type(e)) >>>> tripleoclient.exceptions.DeploymentError: Deployment failed >>>> >>>> During handling of the above exception, another exception occurred: >>>> >>>> Traceback (most recent call last): >>>> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 401, in >>>> run_subcommand >>>> result = cmd.run(parsed_args) >>>> File "/usr/lib/python3.6/site-packages/cliff/command.py", line 185, >>>> in run >>>> return_code = self.take_action(parsed_args) or 0 >>>> File >>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>> 1451, in take_action >>>> if self._standalone_deploy(parsed_args) != 0: >>>> File >>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>> 1400, in _standalone_deploy >>>> raise exceptions.DeploymentError('Deployment failed.') >>>> tripleoclient.exceptions.DeploymentError: Deployment failed. >>>> clean_up Deploy: Deployment failed. >>>> Traceback (most recent call last): >>>> File >>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>> 1345, in _standalone_deploy >>>> raise exceptions.DeploymentError('Deployment failed') >>>> tripleoclient.exceptions.DeploymentError: Deployment failed >>>> >>>> During handling of the above exception, another exception occurred: >>>> >>>> Traceback (most recent call last): >>>> File >>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>> 1363, in _standalone_deploy >>>> raise exceptions.DeploymentError(six.text_type(e)) >>>> tripleoclient.exceptions.DeploymentError: Deployment failed >>>> >>>> During handling of the above exception, another exception occurred: >>>> >>>> Traceback (most recent call last): >>>> File "/usr/lib/python3.6/site-packages/osc_lib/shell.py", line 136, >>>> in run >>>> ret_val = super(OpenStackShell, self).run(argv) >>>> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 281, in run >>>> result = self.run_subcommand(remainder) >>>> File "/usr/lib/python3.6/site-packages/osc_lib/shell.py", line 176, >>>> in run_subcommand >>>> ret_value = super(OpenStackShell, self).run_subcommand(argv) >>>> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 401, in >>>> run_subcommand >>>> result = cmd.run(parsed_args) >>>> File "/usr/lib/python3.6/site-packages/cliff/command.py", line 185, >>>> in run >>>> return_code = self.take_action(parsed_args) or 0 >>>> File >>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>> 1451, in take_action >>>> if self._standalone_deploy(parsed_args) != 0: >>>> File >>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>> 1400, in _standalone_deploy >>>> raise exceptions.DeploymentError('Deployment failed.') >>>> tripleoclient.exceptions.DeploymentError: Deployment failed. >>>> >>>> END return value: 1 >>>> >>>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>>> >>>> An error has occured while deploying the Undercloud. >>>> >>>> See the previous output for details about what went wrong. >>>> >>>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>>> >>>> Command '['sudo', '--preserve-env', 'openstack', 'tripleo', 'deploy', >>>> '--standalone', '--standalone-role', 'Undercloud', '--stack', 'undercloud', >>>> '--local-domain=localdomain', '--local-ip=abce:abce:abce::1/64', >>>> '--templates=/usr/share/openstack-tripleo-heat-templates/', >>>> '--networks-file=network_data_undercloud.yaml', '--heat-native', '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/undercloud.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/use-dns-for-vips.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/podman.yaml', >>>> '-e', '/home/stack/containers-prepare-parameter.yaml', '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/mistral.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/zaqar-swift-backend.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/disable-telemetry.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/tempest.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/ssl/no-tls-endpoints-public-ip.yaml', >>>> '--deployment-user', 'stack', '--output-dir=/home/stack', '-e', >>>> '/home/stack/tripleo-config-generated-env-files/undercloud_parameters.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/tripleo-validations.yaml', >>>> '--debug', '--log-file=install-undercloud.log', '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/undercloud-stack-vstate-dropin.yaml']' >>>> returned non-zero exit status 1. >>>> Command '['sudo', '--preserve-env', 'openstack', 'tripleo', 'deploy', >>>> '--standalone', '--standalone-role', 'Undercloud', '--stack', 'undercloud', >>>> '--local-domain=localdomain', '--local-ip=abce:abce:abce::1/64', >>>> '--templates=/usr/share/openstack-tripleo-heat-templates/', >>>> '--networks-file=network_data_undercloud.yaml', '--heat-native', '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/undercloud.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/use-dns-for-vips.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/podman.yaml', >>>> '-e', '/home/stack/containers-prepare-parameter.yaml', '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/mistral.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/zaqar-swift-backend.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/disable-telemetry.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/tempest.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/ssl/no-tls-endpoints-public-ip.yaml', >>>> '--deployment-user', 'stack', '--output-dir=/home/stack', '-e', >>>> '/home/stack/tripleo-config-generated-env-files/undercloud_parameters.yaml', >>>> '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/environments/tripleo-validations.yaml', >>>> '--debug', '--log-file=install-undercloud.log', '-e', >>>> '/usr/share/openstack-tripleo-heat-templates/undercloud-stack-vstate-dropin.yaml']' >>>> returned non-zero exit status 1. >>>> END return value: 1 >>>> [stack at undercloud ~]$ >>>> >>>> >>>> >> >> -- >> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: From radoslaw.piliszek at gmail.com Fri Dec 3 07:55:25 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Fri, 3 Dec 2021 08:55:25 +0100 Subject: [kolla[ Issue running masakari-hostmonitor In-Reply-To: References: Message-ID: Ah, sorry for the confusion. It's instancemonitor that should run privileged, not hostmonitor. That's right, hostmonitor is really supported only since Wallaby because that's when it was patched to run in containerised deployments. I suggest you simply upgrade your stack to Wallaby and enjoy it being supported. -yoctozepto On Thu, 2 Dec 2021 at 21:18, Chris DiLorenzo wrote: > It's not running it privileged mode: > > ubuntu at control001-poc:~$ docker inspect > --format='{{.HostConfig.Privileged}}' masakari_hostmonitor > false > > The config in ansible/roles/masakari/defaults/main.yml looks like this: > > masakari-hostmonitor: > container_name: masakari_hostmonitor > group: masakari-hostmonitor > enabled: true > ipc_mode: host > image: "{{ masakari_monitors_image_full }}" > volumes: "{{ masakari_hostmonitor_default_volumes + > masakari_hostmonitor_extra_volumes }}" > dimensions: "{{ masakari_hostmonitor_dimensions }}" > > The reason I am running 12.0.0 is that the hostmonitor code isn't in 11.x > > Thanks, > Chris > > On Thu, Dec 2, 2021 at 2:28 PM Rados?aw Piliszek < > radoslaw.piliszek at gmail.com> wrote: > >> Hi Chris, >> >> as for the issue itself - check if the container is running privileged >> (it should). >> >> As for the other details - Kolla-Ansible 12.x is for Wallaby, not >> Victoria, that could be causing issues. >> 12.0.0 is not the latest release for Wallaby either (12.2.0 is for today). >> Fixing these might help with the other issue. >> >> -yoctozepto >> >> On Thu, 2 Dec 2021 at 18:08, Chris DiLorenzo >> wrote: >> > >> > Hello, >> > >> > I am running Kolla-12.0.0 with Openstack Victoria. I have mostly >> successfully deployed masakari. The instance monitoring works, I can set >> up segments in Horizon and via CLI. But I can't get masakari-hostmonitor >> to start: >> > >> > 2021-12-02 16:50:44.999 7 INFO oslo.privsep.daemon [-] Running privsep >> helper: ['sudo', 'privsep-helper', '--config-file', >> '/etc/masakari-monitors/masakari >> > -monitors.conf', '--privsep_context', >> 'masakarimonitors.privsep.monitors_priv', '--privsep_sock_path', >> '/tmp/tmpf_65hc7m/privsep.sock'] >> > 2021-12-02 16:50:44.892 261 INFO oslo.privsep.daemon [-] privsep daemon >> starting >> > 2021-12-02 16:50:44.898 261 INFO oslo.privsep.daemon [-] privsep >> process running with uid/gid: 0/0 >> > 2021-12-02 16:50:44.902 261 ERROR oslo.privsep.daemon [-] [Errno 1] >> Operation not permitted >> > Traceback (most recent call last): >> > File >> "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", >> line 593, in helper_main >> > Daemon(channel, context).run() >> > File >> "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", >> line 403, in run >> > self._drop_privs() >> > File >> "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", >> line 439, in _drop_privs >> > capabilities.drop_all_caps_except(self.caps, self.caps, []) >> > File >> "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/capabilities.py", >> line 156, in drop_all_caps_except >> > raise OSError(errno, os.strerror(errno)) >> > PermissionError: [Errno 1] Operation not permitted >> > 2021-12-02 16:50:45.001 7 DEBUG oslo_privsep.comm [-] EOF on privsep >> read channel _reader_main >> /var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep >> > /comm.py:149 >> > 2021-12-02 16:50:45.475 7 WARNING oslo.privsep.daemon [-] privsep log: >> [Errno 1] Operation not permitted >> > 2021-12-02 16:50:45.565 7 INFO oslo.privsep.daemon [-] Spawned new >> privsep daemon via rootwrap >> > 2021-12-02 16:50:45.565 7 DEBUG oslo.privsep.daemon [-] Accepted >> privsep connection to /tmp/tmpf_65hc7m/privsep.sock __init__ >> /var/lib/kolla/venv/lib/pytho >> > n3.8/site-packages/oslo_privsep/daemon.py:371 >> > 2021-12-02 16:50:45.467 267 INFO oslo.privsep.daemon [-] privsep daemon >> starting >> > 2021-12-02 16:50:45.471 267 INFO oslo.privsep.daemon [-] privsep >> process running with uid/gid: 0/0 >> > 2021-12-02 16:50:45.474 267 ERROR oslo.privsep.daemon [-] [Errno 1] >> Operation not permitted >> > Traceback (most recent call last): >> > File >> "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", >> line 593, in helper_main >> > Daemon(channel, context).run() >> > File >> "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", >> line 403, in run >> > self._drop_privs() >> > File >> "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/daemon.py", >> line 439, in _drop_privs >> > capabilities.drop_all_caps_except(self.caps, self.caps, []) >> > File >> "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_privsep/capabilities.py", >> line 156, in drop_all_caps_except >> > raise OSError(errno, os.strerror(errno)) >> > PermissionError: [Errno 1] Operation not permitted >> > >> > sudoers in the container appears to be set up correctly. >> > >> > Thanks >> > Chris >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From ricolin at ricolky.com Fri Dec 3 09:37:19 2021 From: ricolin at ricolky.com (Rico Lin) Date: Fri, 3 Dec 2021 17:37:19 +0800 Subject: [tc][all] Follow up pain point discussion Message-ID: Dear all The pain point collection etherpad [1] seems to get good attention from teams. We plan to host a one-hour video meeting to keep the discussion going. Please fill in the doodle to select your preferred time if you can join us: https://doodle.com/poll/kur2mfywceerqcvh?utm_source=poll&utm_medium=link The discussion will be hosted in video format. Here are some items (but not limited to) I expect we will discuss and hopefully put some actions/efforts on. - General Pain point: RabbitMQ - General Pain point: OpenStackClient support - General project pain point discussion: How to encourage teams to work/reply on pain points. Meanwhile, we encourage all to check the etherpad and provide your feedback. Also if teams can help to answer questions/concerns on etherpad (like Nova team does) will be greatly appreciated. [1] https://etherpad.opendev.org/p/pain-point-elimination *Rico Lin* -------------- next part -------------- An HTML attachment was scrubbed... URL: From katonalala at gmail.com Fri Dec 3 12:27:22 2021 From: katonalala at gmail.com (Lajos Katona) Date: Fri, 3 Dec 2021 13:27:22 +0100 Subject: [neutron] Drivers meeting agenda - 03.12.2021 In-Reply-To: References: Message-ID: Hi, I found the logs from the meeting when we discussed the quota limit_check from ralonsoh: https://meetings.opendev.org/meetings/neutron_drivers/2021/neutron_drivers.2021-08-27-14.00.log.html#l-61 Regards Lajos Katona (lajoskatona) Lajos Katona ezt ?rta (id?pont: 2021. dec. 2., Cs, 20:40): > Hi Neutron Drivers, > > The agenda for tomorrow's drivers meeting is at [1]. > > I would like to discuss one bug which I feel is more an RFE: > * https://bugs.launchpad.net/neutron/+bug/1952867 : [ml2][ovs] allow > multiple physical networks map to one physical ovs bridge > As physnet-bridge mapping is deep in Neutron I feel that a discussion is > necessary to have a common understanding of the possible benefits and > pitfalls of such shange. > I wrote to Liu to ask him to join, but I'm not sure if he can participate. > > We have one more "On Demand" topic from Rodolfo: > * https://review.opendev.org/c/openstack/python-openstackclient/+/819627/1: > To deprecate "force" behaviour in OSC "quota" commands. This is currently > the default behaviour in Neutron; what is proposed in this patch is to move > Neutron to, by default, check the quota limits and use "--force" parameter > to commit the quota limits without any check. > > [1] https://wiki.openstack.org/wiki/Meetings/NeutronDrivers#Agenda > > See you at the meeting tomorrow. > Lajos Katona (lajoskatona) > -------------- next part -------------- An HTML attachment was scrubbed... URL: From elfosardo at gmail.com Fri Dec 3 13:27:28 2021 From: elfosardo at gmail.com (Riccardo Pittau) Date: Fri, 3 Dec 2021 14:27:28 +0100 Subject: [ironic][meeting] Postpone meeting time until end of March 2022 In-Reply-To: <1955640444.77903.1638308022081@cernmail.cern.ch> References: <1955640444.77903.1638308022081@cernmail.cern.ch> Message-ID: Hello everyone! Since there were no objections we're going to move the meeting time 1 hour later New ironic meeting time is 16:00-17:00 UTC starting on monday December 6th Thank you! Ciao, Riccardo On Tue, Nov 30, 2021 at 10:33 PM wrote: > Sounds good to me as well! > > On 11/30/2021 6:22 PM Dmitry Tantsur wrote: > > > No objections from me either. > > On Tue, Nov 30, 2021 at 3:07 PM Riccardo Pittau > wrote: > > Hello ironicers! > > Due to some overlapping time with other meetings that cause me to attend > to 2 (or sometimes 3.....) meetings at the same time, I'd like to propose > to move the time for the weekly ironic upstream meeting 1 hour later, so > from the current 15:00-16:00 UTC to 16:00-17:00 UTC, until the next > Daylight saving time change that should happen at the end of March 2022. > > Thank you! > > Ciao, > Riccardo > > > > > -- > Red Hat GmbH, https://de.redhat.com/ , Registered seat: Grasbrunn, > Commercial register: Amtsgericht Muenchen, HRB 153243, > Managing Directors: Charles Cachera, Brian Klemm, Laurie Krebs, Michael > O'Neill > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From dtantsur at redhat.com Fri Dec 3 14:17:28 2021 From: dtantsur at redhat.com (Dmitry Tantsur) Date: Fri, 3 Dec 2021 15:17:28 +0100 Subject: [ironic][meeting] Postpone meeting time until end of March 2022 In-Reply-To: References: <1955640444.77903.1638308022081@cernmail.cern.ch> Message-ID: We need to update https://opendev.org/opendev/irc-meetings/src/branch/master/meetings/ironic-bare-metal-team-meeting.yaml to put it in effect. On Fri, Dec 3, 2021 at 2:27 PM Riccardo Pittau wrote: > Hello everyone! > Since there were no objections we're going to move the meeting time 1 hour > later > New ironic meeting time is 16:00-17:00 UTC starting on monday December 6th > Thank you! > > Ciao, > Riccardo > > On Tue, Nov 30, 2021 at 10:33 PM wrote: > >> Sounds good to me as well! >> >> On 11/30/2021 6:22 PM Dmitry Tantsur wrote: >> >> >> No objections from me either. >> >> On Tue, Nov 30, 2021 at 3:07 PM Riccardo Pittau >> wrote: >> >> Hello ironicers! >> >> Due to some overlapping time with other meetings that cause me to attend >> to 2 (or sometimes 3.....) meetings at the same time, I'd like to propose >> to move the time for the weekly ironic upstream meeting 1 hour later, so >> from the current 15:00-16:00 UTC to 16:00-17:00 UTC, until the next >> Daylight saving time change that should happen at the end of March 2022. >> >> Thank you! >> >> Ciao, >> Riccardo >> >> >> >> >> -- >> Red Hat GmbH, https://de.redhat.com/ , Registered seat: Grasbrunn, >> Commercial register: Amtsgericht Muenchen, HRB 153243, >> Managing Directors: Charles Cachera, Brian Klemm, Laurie Krebs, Michael >> O'Neill >> >> -- Red Hat GmbH, https://de.redhat.com/ , Registered seat: Grasbrunn, Commercial register: Amtsgericht Muenchen, HRB 153243, Managing Directors: Charles Cachera, Brian Klemm, Laurie Krebs, Michael O'Neill -------------- next part -------------- An HTML attachment was scrubbed... URL: From elod.illes at est.tech Fri Dec 3 14:29:24 2021 From: elod.illes at est.tech (=?UTF-8?B?RWzFkWQgSWxsw6lz?=) Date: Fri, 3 Dec 2021 15:29:24 +0100 Subject: [release] Release countdown for week R-16, Dec 06-10 Message-ID: <1fee9c14-d1dc-1ac2-7dc4-6324a204c6c9@est.tech> Development Focus ---------------------------- The Yoga-2 milestone will happen in next month, on 06 January, 2022. Yoga-related specs should now be finalized so that teams can move to implementation ASAP. Some teams observe specific deadlines on the second milestone (mostly spec freezes): please refer to https://releases.openstack.org /yoga /schedule.html for details. General Information ---------------------------- Please remember that libraries need to be released at least once per milestone period. At milestone 2, the release team will propose releases for any library that has not been otherwise released since milestone 1. Other non-library deliverables that follow the cycle-with-intermediary release model should have an intermediary release before milestone-2. Those who haven't will be proposed to switch to the cycle-with-rc model, which is more suited to deliverables that are released only once per cycle. At milestone-2 we also freeze the contents of the final release. If you have a new deliverable that should be included in the final release, you should make sure it has a deliverable file in: https://opendev.org/openstack/releases/src/branch/master/deliverables/ yoga You should request a beta release (or intermediary release) for those new deliverables by milestone-2. We understand some may not be quite ready for a full release yet, but if you have something minimally viable to get released it would be good to do a 0.x release to exercise the release tooling for your deliverables. See the MembershipFreeze description for more details: https://releases.openstack.org/ yoga /schedule.html# y -mf Finally, now may be a good time for teams to check on any stable releases that need to be done for your deliverables. If you have bugfixes that have been backported, but no stable release getting those. If you are unsure what is out there committed but not released, in the openstack/releases repo, running the command "tools/list_stable_unreleased_changes.sh " gives a nice report. Upcoming Deadlines & Dates ----------------------------------------- Yoga-2 Milestone:06 January, 2022 El?d Ill?s irc: elodilles -------------- next part -------------- An HTML attachment was scrubbed... URL: From rodrigo.lima at o2sistemas.com Fri Dec 3 18:56:02 2021 From: rodrigo.lima at o2sistemas.com (Rodrigo Lima) Date: Fri, 3 Dec 2021 15:56:02 -0300 Subject: [kolla-ansible] Default masakari configuration should evacuate instances across nodes in wallaby? Message-ID: I'm working on upgrading an openstack farm from victoria to wallaby. After successful upgrade, I would like to enable hacluster and masakari to test HA between failing compute nodes. Everything seems to be running (pacemaker with the remote nodes OK, corosync without errors, masakari-monitors detects the 2 compute nodes online), but... when I simulate the failure of a node with shutdown, the failure and notification appears in the hostmonitor log, but the instances that were on the failed node don't evacuate, and I couldn't find documentation that explains how to do this specific configuration, if necessary. Does anyone have any ideas? -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Fri Dec 3 19:27:18 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Fri, 03 Dec 2021 13:27:18 -0600 Subject: [all][tc] What's happening in Technical Committee: summary 3rd Dec, 21: Reading: 10 min Message-ID: <17d81c3e3c2.1104058bc321556.6008677995285032244@ghanshyammann.com> Hello Everyone, Here is this week's summary of the Technical Committee activities. 1. TC Meetings: ============ * Most of the meeting discussions are summarized below (Completed or in-progress activities section). Meeting full logs are available @ - Recording: https://www.youtube.com/watch?v=PwGYse8j4UY - Topic Info logs: https://meetings.opendev.org/meetings/tc/2021/tc.2021-12-02-15.02.log.html * Next week's meeting is on IRC on 9th Dec, Thursday 15:00 UTC, feel free the topic on the agenda[1] by 8th Dec. 2. What we completed this week: ========================= * Goal proposal for RBAC is merged and the next step is to select it[2] * Add ansible-collection-kolla repo to Kolla project[3] 3. Activities In progress: ================== TC Tracker for Yoga cycle ------------------------------ * This etherpad includes the Yoga cycle targets/working items for TC[4]. Open Reviews ----------------- * 9 open reviews for ongoing activities[5]. TC position on release cadence ------------------------------------- You might have seen the ML thread on the release cadence[6]. We discussed it in this week's TC meeting so that we can provide the TC stand on this. There are mixed amounts of opinions on the different proposals like doing 1-year release, no coordinated release, more releases, mixed LTS release and keeping it same which is 6-month release etc. One of the key things we discussed is how to improve the upgrade process which is one of the key points to revise the current release model. Dan made a good proposal on grenade testing the 'n-1' -> 'n' (current) and 'n-2' -> 'n' (new) so that we make sure 'n' release is upgrade-able from 'n-1' as well as 'n-2' release. For that, we need to think about the deprecation phase to be n+2 release then n+1. Along with that, we can add/improve our upgrade documentation. These are a few ideas we discussed but due to time constraints, we have not concluded yet. We will set up a separate call to continue the discussion sometime in Jan start(after the Christmas holidays) Yoga testing runtime ------------------------- We have an agreement now on Yoga testing runtime and keeping python3.6 support in Yoga, you can see the details in this ML[7] Fixing Zuul config error -------------------------- If your project is listed in zuul config error, please start planning to fix those[8] RBAC discussion: continuing from PTG ---------------------------------------------- The goal proposal is not merged and the next step is to select this goal which is in progress[8] with a good amount of positive votes and no objection yet. Community-wide goal updates ------------------------------------ * Proposed goals: ** RBAC goal is pretty much in good shape now, feel free to review it[9] * Under review: ** FIPS compatibility and compliance [10]. Adjutant need maintainers and PTLs ------------------------------------------- We are waiting to hear from Braden, Albert on permission to work on this project[11]. New project 'Skyline' proposal ------------------------------------ * TC is still waiting for skyline team to respond to open queries[12]. TC tags analysis ------------------- * TC agreed to remove the framework and it is communicated in ML[13]. Project updates ------------------- * Add NVidia vGPU plugin charm to OpenStack charms[14] * Retire js-openstack-lib [15] 4. How to contact the TC: ==================== If you would like to discuss or give feedback to TC, you can reach out to us in multiple ways: 1. Email: you can send the email with tag [tc] on openstack-discuss ML[16]. 2. Weekly meeting: The Technical Committee conduct a weekly meeting every Thursday 15 UTC [17] 3. Ping us using 'tc-members' nickname on #openstack-tc IRC channel. [1] https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting [2] https://review.opendev.org/c/openstack/governance/+/815158 [3] https://review.opendev.org/c/openstack/governance/+/819331 [4] https://etherpad.opendev.org/p/tc-yoga-tracker [5] https://review.opendev.org/q/projects:openstack/governance+status:open [6] http://lists.openstack.org/pipermail/openstack-discuss/2021-November/025684.html [7] http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026164.html [8] https://etherpad.opendev.org/p/zuul-config-error-openstack [9] https://review.opendev.org/c/openstack/governance/+/818817 [10] https://review.opendev.org/c/openstack/governance/+/816587 [11] http://lists.openstack.org/pipermail/openstack-discuss/2021-November/025786.html [12] https://review.opendev.org/c/openstack/governance/+/814037 [13] http://lists.openstack.org/pipermail/openstack-discuss/2021-October/025571.html [14] https://review.opendev.org/c/openstack/governance/+/819856 [15] https://review.opendev.org/c/openstack/governance/+/807163 [16] http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-discuss [17] http://eavesdrop.openstack.org/#Technical_Committee_Meeting -gmann From yipikai7 at gmail.com Fri Dec 3 20:01:05 2021 From: yipikai7 at gmail.com (Cedric Lemarchand) Date: Fri, 3 Dec 2021 21:01:05 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: Message-ID: We faced the same use case, and we end up doing something like this: - disable security on the network provider (or on the instance port where pxe boot is needed) - deploy or rescue the intance with an ipxe image Cheers On Wed, Dec 1, 2021, 23:36 Thomas Goirand wrote: > Hi, > > In order to test OCI [1] within OpenStack itself, I would need my > instances to do PXE booting. This would be with something like this on > the qemu command line: > > qemu-system-x86_64 -boot n > > I believe this would also mean setting-up network cards that can do PXE > booting, like the e1000, which Qemu supports. > > If that is possible, then I would setup DHCP + tftp-hpa on another > instance, wired on the same private network. > > Your thoughts? Is this possible? > > Cheers, > > Thomas Goirand (zigo) > > [1] > https://salsa.debian.org/openstack-team/debian/openstack-cluster-installer > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From aschultz at redhat.com Fri Dec 3 21:07:43 2021 From: aschultz at redhat.com (Alex Schultz) Date: Fri, 3 Dec 2021 14:07:43 -0700 Subject: [TripleO] IPV6 Support In-Reply-To: References: Message-ID: On Thu, Dec 2, 2021 at 11:58 PM Lokendra Rathour wrote: > Hi Alex, > From your message I could understand that Stein is no longer supported, > but does it support ipv6, please advise. > We are also using stein to test the Upgrade feature of Triple0. We can > start with Train as well and install and upgrade to Ussurie. But ipv6 > support is something that we find a miss in the undercloud.conf or > somewhere we can specify ipv6 details. > > Request you to please share any document/link which helps us try train on > centos 8 with ipv6. For now, I am referring to this document. > Undercloud Installation ? TripleO 3.0.0 documentation (openstack.org) > > > Unfortunately our documentation is usually for newer versions and train documentation may not be there for ipv6 as we've reworked the network templates in recent releases. At this point if you're starting from scratch using Wallaby is a better idea. If you are looking for some example templates, we actually have a job in CI that tests this and would have example configurations and commands. This job only uses ipv6 for the overcloud networks and not for the provisioning network. I don't think ipv6 provisioning is fully supported in Train. It might work but it's something you'd need to do some advanced configuration in order to get the interfaces correctly configured. For our CI job, we have example templates in tripleo-heat-templates. These could be used as a starting point. https://opendev.org/openstack/tripleo-heat-templates/src/branch/stable/train/ci/environments/network/multiple-nics-ipv6 Our ipv6 job is: https://review.rdoproject.org/zuul/builds?job_name=periodic-tripleo-ci-centos-8-ovb-3ctlr_1comp-featureset035-train If you click on the last successful build and click "View log", it'll take you to our CI logs. Example: https://review.rdoproject.org/zuul/build/3190c53f516543edb1844f28004d5c33 View Log takes you to: https://logserver.rdoproject.org/openstack-periodic-integration-stable4/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-ci-centos-8-ovb-3ctlr_1comp-featureset035-train/3190c53/ All the deployment templates and configurations are located under the logs/undercloud/home/zuul path. https://logserver.rdoproject.org/openstack-periodic-integration-stable4/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-ci-centos-8-ovb-3ctlr_1comp-featureset035-train/3190c53/logs/undercloud/home/zuul/ For ipv6 the network configuration, the important flags/options to `openstack overcloud deploy` would be: -n network_data.yaml This by default uses network_data.yaml from tripleo-heat-templates. This has the ip information for the various networks. You can take the default and copy/update it to fit your needs. This file is used to generate some of the network configurations during the deployment. -e /usr/share/openstack-tripleo-heat-templates/environments/network-isolation-v6.yaml This environment file enables ipv6 for ports and configures all the networks for use. This also has examples on overriding default ips. -e /usr/share/openstack-tripleo-heat-templates/ci/environments/network/multiple-nics-ipv6/network-environment.yaml -e /home/zuul/network-environment.yaml These configuration files contain the references for the network interface configs as well as setting the ranges necessary for the pools. Good luck. > Thanks again for your quick response. Your response will help me decide > things at the initial point only. > -Lokendra > > > On Fri, Dec 3, 2021 at 11:28 AM Alex Schultz wrote: > >> The commands I provided were for Train on CentOS 8. They were not >> related to the original topic of the email (ipv6) but rather the container >> startup issues. Stein is no longer supported and it would not be >> recommended to use it. >> >> On Thu, Dec 2, 2021 at 10:36 PM Lokendra Rathour < >> lokendrarathour at gmail.com> wrote: >> >>> Hi Alex, >>> we are trying on centos 7/8, and similar commands for centos 7/8 is not >>> working. Can you please suggest related commands for the same? >>> we have tried with yum/dnf in centos 7/8. >>> >>> Just to confirm again we are using the stein release on centos 7 where >>> we want to use ipv6 or dual-stack if possible. >>> Thanks once again for your help. >>> >>> -Lokendra >>> >>> On Thu, Nov 25, 2021 at 11:18 PM Alex Schultz >>> wrote: >>> >>>> Yes IPV6 is supported, however the error you provided indicates >>>> problems starting containers. Make sure you pin to container-tools:3.0 to >>>> ensure you get the version we expect. >>>> >>>> dnf module -y disable container-tools:rhel8 ; >>>> dnf module -y enable container-tools:3.0 >>>> >>>> On Thu, Nov 25, 2021 at 1:02 AM Lokendra Rathour < >>>> lokendrarathour at gmail.com> wrote: >>>> >>>>> Hi, >>>>> I am trying to install the undercloud using the ipv6 address on Centos >>>>> 8 using Train release. >>>>> It is seen that the deployment of undercloud is getting failed with >>>>> error as mentioned below. Same deployment is working in the normal ipv4 >>>>> case. >>>>> So Questions: >>>>> >>>>> 1. is IPV6 supported in TripleO , if yes then please suggest how. >>>>> >>>>> >>>>> Error: >>>>> nied", >>>>> " attempt(s): 2", >>>>> "2021-11-25 07:25:31,535 WARNING: 65355 -- Retrying running >>>>> container: zaqar", >>>>> "2021-11-25 07:25:34,887 ERROR: 65355 -- ['/usr/bin/podman', >>>>> 'run', '--user', '0', '--name', 'container-puppet-zaqar', '--env', >>>>> 'PUPPET_TAGS=file,file_line,concat,augeas,cron,zaqar_config', '--env', >>>>> 'NAME=zaqar', '--env', 'HOSTNAME=undercloud', '--env', 'NO_ARCHIVE=', >>>>> '--env', 'STEP=6', '--env', 'NET_HOST=true', '--env', 'DEBUG=False', >>>>> '--volume', '/etc/localtime:/etc/localtime:ro', '--volume', >>>>> '/tmp/tmp06nvxzzz:/etc/config.pp:ro', '--volume', >>>>> '/etc/puppet/:/tmp/puppet-etc/:ro', '--volume', >>>>> '/etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:ro', '--volume', >>>>> '/etc/pki/tls/certs/ca-bundle.crt:/etc/pki/tls/certs/ca-bundle.crt:ro', >>>>> '--volume', >>>>> '/etc/pki/tls/certs/ca-bundle.trust.crt:/etc/pki/tls/certs/ca-bundle.trust.crt:ro', >>>>> '--volume', '/etc/pki/tls/cert.pem:/etc/pki/tls/cert.pem:ro', '--volume', >>>>> '/var/lib/config-data:/var/lib/config-data/:rw', '--volume', >>>>> '/var/lib/container-puppet/puppetlabs/facter.conf:/etc/puppetlabs/facter/facter.conf:ro', >>>>> '--volume', '/var/lib/container-puppet/puppetlabs/:/opt/puppetlabs/:ro', >>>>> '--volume', '/dev/log:/dev/log:rw', '--rm', '--log-driver', 'k8s-file', >>>>> '--log-opt', 'path=/var/log/containers/stdouts/container-puppet-zaqar.log', >>>>> '--security-opt', 'label=disable', '--volume', >>>>> '/usr/share/openstack-puppet/modules/:/usr/share/openstack-puppet/modules/:ro', >>>>> '--entrypoint', '/var/lib/container-puppet/container-puppet.sh', '--net', >>>>> 'host', '--volume', '/etc/hosts:/etc/hosts:ro', '--volume', >>>>> '/var/lib/container-puppet/container-puppet.sh:/var/lib/container-puppet/container-puppet.sh:ro', >>>>> 'undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-zaqar-wsgi:current-tripleo'] >>>>> run failed after Error: container_linux.go:370: starting container process >>>>> caused: error adding seccomp filter rule for syscall bdflush: permission >>>>> denied: OCI permission denied", >>>>> " attempt(s): 3", >>>>> "2021-11-25 07:25:34,888 WARNING: 65355 -- Retrying running >>>>> container: zaqar", >>>>> "2021-11-25 07:25:34,888 ERROR: 65355 -- Failed running container >>>>> for zaqar", >>>>> "2021-11-25 07:25:34,888 INFO: 65355 -- Finished processing puppet >>>>> configs for zaqar", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring crond", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>>> glance_api", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>>> heat_api", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring heat", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>>> ironic_api", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring ironic", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>>> ironic_inspector", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>>> neutron", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring iscsid", >>>>> "2021-11-25 07:25:34,889 ERROR: 65345 -- ERROR configuring >>>>> keystone", >>>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>>>> memcached", >>>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>>>> mistral", >>>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring mysql", >>>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring nova", >>>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>>>> rabbitmq", >>>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>>>> placement", >>>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring swift", >>>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring >>>>> swift_ringbuilder", >>>>> "2021-11-25 07:25:34,890 ERROR: 65345 -- ERROR configuring zaqar" >>>>> ], >>>>> "stderr_lines": [], >>>>> "_ansible_no_log": false, >>>>> "attempts": 15 >>>>> } >>>>> ] >>>>> ] >>>>> Not cleaning working directory >>>>> /home/stack/tripleo-heat-installer-templates >>>>> Not cleaning ansible directory /home/stack/undercloud-ansible-mw4crw92 >>>>> Install artifact is located at >>>>> /home/stack/undercloud-install-20211125072536.tar.bzip2 >>>>> >>>>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>>>> >>>>> Deployment Failed! >>>>> >>>>> ERROR: Heat log files: >>>>> /var/log/heat-launcher/undercloud_deploy-o_qf1b4w >>>>> >>>>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>>>> >>>>> Deployment failed. >>>>> Traceback (most recent call last): >>>>> File >>>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>>> 1345, in _standalone_deploy >>>>> raise exceptions.DeploymentError('Deployment failed') >>>>> tripleoclient.exceptions.DeploymentError: Deployment failed >>>>> >>>>> During handling of the above exception, another exception occurred: >>>>> >>>>> Traceback (most recent call last): >>>>> File >>>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>>> 1363, in _standalone_deploy >>>>> raise exceptions.DeploymentError(six.text_type(e)) >>>>> tripleoclient.exceptions.DeploymentError: Deployment failed >>>>> >>>>> During handling of the above exception, another exception occurred: >>>>> >>>>> Traceback (most recent call last): >>>>> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 401, in >>>>> run_subcommand >>>>> result = cmd.run(parsed_args) >>>>> File "/usr/lib/python3.6/site-packages/cliff/command.py", line 185, >>>>> in run >>>>> return_code = self.take_action(parsed_args) or 0 >>>>> File >>>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>>> 1451, in take_action >>>>> if self._standalone_deploy(parsed_args) != 0: >>>>> File >>>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>>> 1400, in _standalone_deploy >>>>> raise exceptions.DeploymentError('Deployment failed.') >>>>> tripleoclient.exceptions.DeploymentError: Deployment failed. >>>>> clean_up Deploy: Deployment failed. >>>>> Traceback (most recent call last): >>>>> File >>>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>>> 1345, in _standalone_deploy >>>>> raise exceptions.DeploymentError('Deployment failed') >>>>> tripleoclient.exceptions.DeploymentError: Deployment failed >>>>> >>>>> During handling of the above exception, another exception occurred: >>>>> >>>>> Traceback (most recent call last): >>>>> File >>>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>>> 1363, in _standalone_deploy >>>>> raise exceptions.DeploymentError(six.text_type(e)) >>>>> tripleoclient.exceptions.DeploymentError: Deployment failed >>>>> >>>>> During handling of the above exception, another exception occurred: >>>>> >>>>> Traceback (most recent call last): >>>>> File "/usr/lib/python3.6/site-packages/osc_lib/shell.py", line 136, >>>>> in run >>>>> ret_val = super(OpenStackShell, self).run(argv) >>>>> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 281, in >>>>> run >>>>> result = self.run_subcommand(remainder) >>>>> File "/usr/lib/python3.6/site-packages/osc_lib/shell.py", line 176, >>>>> in run_subcommand >>>>> ret_value = super(OpenStackShell, self).run_subcommand(argv) >>>>> File "/usr/lib/python3.6/site-packages/cliff/app.py", line 401, in >>>>> run_subcommand >>>>> result = cmd.run(parsed_args) >>>>> File "/usr/lib/python3.6/site-packages/cliff/command.py", line 185, >>>>> in run >>>>> return_code = self.take_action(parsed_args) or 0 >>>>> File >>>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>>> 1451, in take_action >>>>> if self._standalone_deploy(parsed_args) != 0: >>>>> File >>>>> "/usr/lib/python3.6/site-packages/tripleoclient/v1/tripleo_deploy.py", line >>>>> 1400, in _standalone_deploy >>>>> raise exceptions.DeploymentError('Deployment failed.') >>>>> tripleoclient.exceptions.DeploymentError: Deployment failed. >>>>> >>>>> END return value: 1 >>>>> >>>>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>>>> >>>>> An error has occured while deploying the Undercloud. >>>>> >>>>> See the previous output for details about what went wrong. >>>>> >>>>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! >>>>> >>>>> Command '['sudo', '--preserve-env', 'openstack', 'tripleo', 'deploy', >>>>> '--standalone', '--standalone-role', 'Undercloud', '--stack', 'undercloud', >>>>> '--local-domain=localdomain', '--local-ip=abce:abce:abce::1/64', >>>>> '--templates=/usr/share/openstack-tripleo-heat-templates/', >>>>> '--networks-file=network_data_undercloud.yaml', '--heat-native', '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/undercloud.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/use-dns-for-vips.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/podman.yaml', >>>>> '-e', '/home/stack/containers-prepare-parameter.yaml', '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/mistral.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/zaqar-swift-backend.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/disable-telemetry.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/tempest.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/ssl/no-tls-endpoints-public-ip.yaml', >>>>> '--deployment-user', 'stack', '--output-dir=/home/stack', '-e', >>>>> '/home/stack/tripleo-config-generated-env-files/undercloud_parameters.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/tripleo-validations.yaml', >>>>> '--debug', '--log-file=install-undercloud.log', '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/undercloud-stack-vstate-dropin.yaml']' >>>>> returned non-zero exit status 1. >>>>> Command '['sudo', '--preserve-env', 'openstack', 'tripleo', 'deploy', >>>>> '--standalone', '--standalone-role', 'Undercloud', '--stack', 'undercloud', >>>>> '--local-domain=localdomain', '--local-ip=abce:abce:abce::1/64', >>>>> '--templates=/usr/share/openstack-tripleo-heat-templates/', >>>>> '--networks-file=network_data_undercloud.yaml', '--heat-native', '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/undercloud.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/use-dns-for-vips.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/podman.yaml', >>>>> '-e', '/home/stack/containers-prepare-parameter.yaml', '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/mistral.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/zaqar-swift-backend.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/disable-telemetry.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/services/tempest.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/ssl/no-tls-endpoints-public-ip.yaml', >>>>> '--deployment-user', 'stack', '--output-dir=/home/stack', '-e', >>>>> '/home/stack/tripleo-config-generated-env-files/undercloud_parameters.yaml', >>>>> '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/environments/tripleo-validations.yaml', >>>>> '--debug', '--log-file=install-undercloud.log', '-e', >>>>> '/usr/share/openstack-tripleo-heat-templates/undercloud-stack-vstate-dropin.yaml']' >>>>> returned non-zero exit status 1. >>>>> END return value: 1 >>>>> [stack at undercloud ~]$ >>>>> >>>>> >>>>> >>> >>> -- >>> >>> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From juliaashleykreger at gmail.com Fri Dec 3 21:41:24 2021 From: juliaashleykreger at gmail.com (Julia Kreger) Date: Fri, 3 Dec 2021 13:41:24 -0800 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: <5af59ddc-669f-09cf-233e-5f85f0e89999@debian.org> References: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> <5af59ddc-669f-09cf-233e-5f85f0e89999@debian.org> Message-ID: On Thu, Dec 2, 2021 at 12:10 AM Thomas Goirand wrote: > > On 12/2/21 12:51 AM, Clark Boylan wrote: > > I'm not super familiar with all the details but I believe ironic does something similar in their devstack plugin [2]. They do this so they can create fake baremetal instances that are actually VMs that ironic manages. But a similar setup will probably work for your use cases if PXE is what you are starting with. > > > > [2] https://opendev.org/openstack/ironic/src/branch/master/devstack/lib/ironic#L2109 > > Hi, > > Thanks Clark for your answer. > > I've read the script a little bit, and I have to admit I do not > understand what it does... :( > > If some Ironic falks can explain, that'd be great! > > Cheers, > > Thomas Goirand (zigo) > Extreme magic - OpenStack Edition. At least that feels like the answer. In essence, we create VMs. We attach them to the network, then configure virtualbmc or sushy-tools to play the role of a BMC and one of the things that library understand is the giant switch of "boot to network". From there we ipxe boot. The VMs play the role of baremetal machines through the emulation of the BMC and they are just hooked up to the network. This does require the job to use root privileges to setup networking/vms. And the virtualbmc or sushy-tools service needs enough rights to toggle the VM's config. Would be happy to try and discuss specifics of the devstack plugin, but I'm unsure if that would really help. Regardless, you know where to find us. :) As an aside, when I first read the post I was thinking of booting from ISO images. I've known people to embed ipxe into disk images, and you can embed static config/instructions for booting into an ipxe binary. It does require custom binary builds, but it might make what you're trying to achieve easier. From zigo at debian.org Fri Dec 3 22:20:41 2021 From: zigo at debian.org (Thomas Goirand) Date: Fri, 3 Dec 2021 23:20:41 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> <5af59ddc-669f-09cf-233e-5f85f0e89999@debian.org> Message-ID: On 12/3/21 10:41 PM, Julia Kreger wrote: > Extreme magic - OpenStack Edition. At least that feels like the answer. > > In essence, we create VMs. We attach them to the network, then > configure virtualbmc or sushy-tools to play the role of a BMC and one > of the things that library understand is the giant switch of "boot to > network". From there we ipxe boot. The VMs play the role of baremetal > machines through the emulation of the BMC and they are just hooked up > to the network. > > This does require the job to use root privileges to setup > networking/vms. And the virtualbmc or sushy-tools service needs enough > rights to toggle the VM's config. Would be happy to try and discuss > specifics of the devstack plugin, but I'm unsure if that would really > help. Regardless, you know where to find us. :) > > As an aside, when I first read the post I was thinking of booting from > ISO images. I've known people to embed ipxe into disk images, and you > can embed static config/instructions for booting into an ipxe binary. > It does require custom binary builds, but it might make what you're > trying to achieve easier. Hi Julia! Thanks for your answer. Well, I've read multiple times your reply, but I have to admit I still don't get all the details. For example, how would you "embed ipxe into disk images" for example? Should I for example run the iPXE binary from something like syslinux (or grub?)? But isn't iPXE usually running *after* the BIOS of a PC gets an IP from DHCP? I've just read a post where someone boots the iPXE binary from Grub, so that part shouldn't be hard: format a disk, make a ext4 partition, put the iPXE binary there, and tell grub to execute it. Is that all?!? Cheers, Thomas Goirand (zigo) P.S: I do have some knowledge on how to prepare disk images. After all, I've maintained the Debian OpenStack images for 3 Debian releases. So I probably just need a few pointers... From aschultz at redhat.com Fri Dec 3 22:34:45 2021 From: aschultz at redhat.com (Alex Schultz) Date: Fri, 3 Dec 2021 15:34:45 -0700 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> <5af59ddc-669f-09cf-233e-5f85f0e89999@debian.org> Message-ID: On Fri, Dec 3, 2021 at 3:22 PM Thomas Goirand wrote: > > On 12/3/21 10:41 PM, Julia Kreger wrote: > > Extreme magic - OpenStack Edition. At least that feels like the answer. > > > > In essence, we create VMs. We attach them to the network, then > > configure virtualbmc or sushy-tools to play the role of a BMC and one > > of the things that library understand is the giant switch of "boot to > > network". From there we ipxe boot. The VMs play the role of baremetal > > machines through the emulation of the BMC and they are just hooked up > > to the network. > > > > This does require the job to use root privileges to setup > > networking/vms. And the virtualbmc or sushy-tools service needs enough > > rights to toggle the VM's config. Would be happy to try and discuss > > specifics of the devstack plugin, but I'm unsure if that would really > > help. Regardless, you know where to find us. :) > > > > As an aside, when I first read the post I was thinking of booting from > > ISO images. I've known people to embed ipxe into disk images, and you > > can embed static config/instructions for booting into an ipxe binary. > > It does require custom binary builds, but it might make what you're > > trying to achieve easier. > > Hi Julia! > > Thanks for your answer. > > Well, I've read multiple times your reply, but I have to admit I still > don't get all the details. For example, how would you "embed ipxe into > disk images" for example? Should I for example run the iPXE binary from > something like syslinux (or grub?)? But isn't iPXE usually running > *after* the BIOS of a PC gets an IP from DHCP? > I'm sure there are other examples around but the one we used to (maybe still do) use for ovb can be built via: https://opendev.org/openstack/openstack-virtual-baremetal/src/branch/master/ipxe You basically build an image that just attempts ipxe on repeat. $ git clone --recurse-submodules --remote-submodules https://opendev.org/openstack/openstack-virtual-baremetal $ cd openstack-virtual-baremetal/ipxe $ make ??? $ openstack image create --file ipxe-boot.img ipxe-boot > I've just read a post where someone boots the iPXE binary from Grub, so > that part shouldn't be hard: format a disk, make a ext4 partition, put > the iPXE binary there, and tell grub to execute it. Is that all?!? > > Cheers, > > Thomas Goirand (zigo) > > P.S: I do have some knowledge on how to prepare disk images. After all, > I've maintained the Debian OpenStack images for 3 Debian releases. So I > probably just need a few pointers... > From zigo at debian.org Fri Dec 3 22:43:33 2021 From: zigo at debian.org (Thomas Goirand) Date: Fri, 3 Dec 2021 23:43:33 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> <5af59ddc-669f-09cf-233e-5f85f0e89999@debian.org> Message-ID: <865ebbe2-1615-08f3-da9e-8f95a8552754@debian.org> On 12/3/21 11:34 PM, Alex Schultz wrote: > On Fri, Dec 3, 2021 at 3:22 PM Thomas Goirand wrote: >> >> On 12/3/21 10:41 PM, Julia Kreger wrote: >>> Extreme magic - OpenStack Edition. At least that feels like the answer. >>> >>> In essence, we create VMs. We attach them to the network, then >>> configure virtualbmc or sushy-tools to play the role of a BMC and one >>> of the things that library understand is the giant switch of "boot to >>> network". From there we ipxe boot. The VMs play the role of baremetal >>> machines through the emulation of the BMC and they are just hooked up >>> to the network. >>> >>> This does require the job to use root privileges to setup >>> networking/vms. And the virtualbmc or sushy-tools service needs enough >>> rights to toggle the VM's config. Would be happy to try and discuss >>> specifics of the devstack plugin, but I'm unsure if that would really >>> help. Regardless, you know where to find us. :) >>> >>> As an aside, when I first read the post I was thinking of booting from >>> ISO images. I've known people to embed ipxe into disk images, and you >>> can embed static config/instructions for booting into an ipxe binary. >>> It does require custom binary builds, but it might make what you're >>> trying to achieve easier. >> >> Hi Julia! >> >> Thanks for your answer. >> >> Well, I've read multiple times your reply, but I have to admit I still >> don't get all the details. For example, how would you "embed ipxe into >> disk images" for example? Should I for example run the iPXE binary from >> something like syslinux (or grub?)? But isn't iPXE usually running >> *after* the BIOS of a PC gets an IP from DHCP? >> > > I'm sure there are other examples around but the one we used to (maybe > still do) use for ovb can be built via: > > https://opendev.org/openstack/openstack-virtual-baremetal/src/branch/master/ipxe > > You basically build an image that just attempts ipxe on repeat. Hi Alex! In fact, I just found out my solution: http://boot.ipxe.org/ipxe.iso Just download this, then: openstack image create \ --disk-format iso \ --container-format bare \ --file ipxe.iso ipxe.iso Et voila! The VM boots from PXE (I checked on the VNC console, and it also gets DHCP). I didn't go further, but it should be easy to setup the rest of the components (ie: a private network without DHCP, and a VM acting as DHCPd + tftp-hpa). I'll let you guys know how much progress I'm making with all of this (probably I'll start working on this next week...). Cheers, Thomas Goirand (zigo) From aschultz at redhat.com Fri Dec 3 22:54:10 2021 From: aschultz at redhat.com (Alex Schultz) Date: Fri, 3 Dec 2021 15:54:10 -0700 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: <865ebbe2-1615-08f3-da9e-8f95a8552754@debian.org> References: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> <5af59ddc-669f-09cf-233e-5f85f0e89999@debian.org> <865ebbe2-1615-08f3-da9e-8f95a8552754@debian.org> Message-ID: On Fri, Dec 3, 2021 at 3:45 PM Thomas Goirand wrote: > > On 12/3/21 11:34 PM, Alex Schultz wrote: > > On Fri, Dec 3, 2021 at 3:22 PM Thomas Goirand wrote: > >> > >> On 12/3/21 10:41 PM, Julia Kreger wrote: > >>> Extreme magic - OpenStack Edition. At least that feels like the answer. > >>> > >>> In essence, we create VMs. We attach them to the network, then > >>> configure virtualbmc or sushy-tools to play the role of a BMC and one > >>> of the things that library understand is the giant switch of "boot to > >>> network". From there we ipxe boot. The VMs play the role of baremetal > >>> machines through the emulation of the BMC and they are just hooked up > >>> to the network. > >>> > >>> This does require the job to use root privileges to setup > >>> networking/vms. And the virtualbmc or sushy-tools service needs enough > >>> rights to toggle the VM's config. Would be happy to try and discuss > >>> specifics of the devstack plugin, but I'm unsure if that would really > >>> help. Regardless, you know where to find us. :) > >>> > >>> As an aside, when I first read the post I was thinking of booting from > >>> ISO images. I've known people to embed ipxe into disk images, and you > >>> can embed static config/instructions for booting into an ipxe binary. > >>> It does require custom binary builds, but it might make what you're > >>> trying to achieve easier. > >> > >> Hi Julia! > >> > >> Thanks for your answer. > >> > >> Well, I've read multiple times your reply, but I have to admit I still > >> don't get all the details. For example, how would you "embed ipxe into > >> disk images" for example? Should I for example run the iPXE binary from > >> something like syslinux (or grub?)? But isn't iPXE usually running > >> *after* the BIOS of a PC gets an IP from DHCP? > >> > > > > I'm sure there are other examples around but the one we used to (maybe > > still do) use for ovb can be built via: > > > > https://opendev.org/openstack/openstack-virtual-baremetal/src/branch/master/ipxe > > > > You basically build an image that just attempts ipxe on repeat. > > Hi Alex! > > In fact, I just found out my solution: http://boot.ipxe.org/ipxe.iso > Yea I think that's what gets built with make and then it qemu-img converts it for you. > Just download this, then: > > openstack image create \ > --disk-format iso \ > --container-format bare \ > --file ipxe.iso ipxe.iso > > Et voila! The VM boots from PXE (I checked on the VNC console, and it > also gets DHCP). I didn't go further, but it should be easy to setup the > rest of the components (ie: a private network without DHCP, and a VM > acting as DHCPd + tftp-hpa). > Yea as previously mentioned you may just need to disable port security and it should be fine. This is basically what we (tripleo) do in our 3rd party testing where the dhcpd is from the undercloud. I've also used the same image locally in libvirt + sushy-emulator for testing as well. > I'll let you guys know how much progress I'm making with all of this > (probably I'll start working on this next week...). > > Cheers, > > Thomas Goirand (zigo) > From mnaser at vexxhost.com Sat Dec 4 10:48:53 2021 From: mnaser at vexxhost.com (Mohammed Naser) Date: Sat, 4 Dec 2021 14:48:53 +0400 Subject: [magnum] dropping mesos code Message-ID: Hi there, Does it make sense for us to drop the Mesos driver code from the code base? I don't know of anyone that actually uses it and I believe it's likely not been tested for years. Mohammed -- Mohammed Naser VEXXHOST, Inc. From radoslaw.piliszek at gmail.com Sat Dec 4 11:48:12 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Sat, 4 Dec 2021 12:48:12 +0100 Subject: [kolla-ansible] Default masakari configuration should evacuate instances across nodes in wallaby? In-Reply-To: References: Message-ID: What do the Masakari API and Masakari Engine logs show? -yoctozepto On Fri, 3 Dec 2021 at 19:57, Rodrigo Lima wrote: > I'm working on upgrading an openstack farm from victoria to wallaby. After > successful upgrade, I would like to enable hacluster and masakari to test > HA between failing compute nodes. Everything seems to be running (pacemaker > with the remote nodes OK, corosync without errors, masakari-monitors > detects the 2 compute nodes online), but... when I simulate the failure of > a node with shutdown, the failure and notification appears in the > hostmonitor log, but the instances that were on the failed node don't > evacuate, and I couldn't find documentation that explains how to do this > specific configuration, if necessary. Does anyone have any ideas? > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From thierry at openstack.org Sat Dec 4 14:56:24 2021 From: thierry at openstack.org (Thierry Carrez) Date: Sat, 4 Dec 2021 15:56:24 +0100 Subject: [all][tc] Relmgt team position on release cadence In-Reply-To: <6536dfae-cb75-4263-856d-f4dbdfa2db01@www.fastmail.com> References: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> <8301ef05-4011-43e9-beec-66f6aeeb00f9@www.fastmail.com> <20211129130957.3poysdeqsufkp3pg@yuggoth.org> <17d72ae5ee2.b80bfd38119150.5945079300853303270@ghanshyammann.com> <20211130220913.ydk5zhyvkdl7g7zc@yuggoth.org> <6536dfae-cb75-4263-856d-f4dbdfa2db01@www.fastmail.com> Message-ID: Jean-Philippe Evrard wrote: > [...] > Here is what I read: > 1) Many want more releases, not less. I haven't seen a complaint about tagging more releases. > 2) More than one person is proposing to abandon the integrated release, and nobody has complained about it. Here are a few drawbacks of abandoning the integrated release (which is really a "synchronized release"): - You would no longer have a baseline of components that are heavily tested together which we do community QA on and (collectively) clearly sign off on, so downstream integrators are on their own and have much more work to do - You would no longer have clearly-comparable points between distributions. Everyone knows what "running Ussuri" means, which facilitates communication and bug report handling in the community. - You would no longer have clear community support commitments. We currently maintain and fix bug reports from people running vanilla "Ussuri"... but do we want to care about every combination of components under the sun? (maybe we do already) - You would no longer have "OpenStack" released, so you miss the regular marketing opportunity to remind the rest of the world that it still exists. The OpenStack brand fades, and it gets more complicated to get development resources to work on it. Without the synchronized release, OpenStack essentially becomes a rolling distribution of cloud components on which we make very limited guarantees. I guess it is suitable to build maintained distributions on, but it really is no longer directly usable beyond development. Is that what we want? > 3) Many people seem eager to carry "stable branches" for "critical patches", but no new definition of such criticality was done. Note that a common stable branch cut is *exactly* the same thing as a synchronized release... So I see 2 and 3 as being opposed views. -- Thierry Carrez (ttx) From suzhengwei at inspur.com Sun Dec 5 08:04:51 2021 From: suzhengwei at inspur.com (=?utf-8?B?U2FtIFN1ICjoi4/mraPkvJ8p?=) Date: Sun, 5 Dec 2021 08:04:51 +0000 Subject: =?utf-8?B?562U5aSNOiBbbGlzdHMub3BlbnN0YWNrLm9yZ+S7o+WPkV1ba29sbGEtYW5z?= =?utf-8?B?aWJsZV0gRGVmYXVsdCBtYXNha2FyaSBjb25maWd1cmF0aW9uIHNob3VsZCBl?= =?utf-8?Q?vacuate_instances_across_nodes_in_wallaby=3F?= In-Reply-To: References: <2e9bcb5dc06f50c55e5b65c775fd4d01@sslemail.net> Message-ID: <651d5a37cf8d40269cfdb56d21b1dd51@inspur.com> Hi, Rodrigo Lima? Have you found the failure notification in the masakari dashboard or DB( Table notifications)? If so, the failure notification would trigger the recovery workflow, and you need to check the masakari-api and masakari-engine log to find the detail. If not, it could be a deployment problem, and you need to check the connectivity between masakari-api and masakari-hostmonitor. ???: Rodrigo Lima [mailto:rodrigo.lima at o2sistemas.com] ????: 2021?12?4? 2:56 ???: openstack-discuss at lists.openstack.org ??: [lists.openstack.org??][kolla-ansible] Default masakari configuration should evacuate instances across nodes in wallaby? I'm working on upgrading an openstack farm from victoria to wallaby. After successful upgrade, I would like to enable hacluster and masakari to test HA between failing compute nodes. Everything seems to be running (pacemaker with the remote nodes OK, corosync without errors, masakari-monitors detects the 2 compute nodes online), but... when I simulate the failure of a node with shutdown, the failure and notification appears in the hostmonitor log, but the instances that were on the failed node don't evacuate, and I couldn't find documentation that explains how to do this specific configuration, if necessary. Does anyone have any ideas? -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: ~WRD000.jpg Type: image/jpeg Size: 823 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3606 bytes Desc: not available URL: From suzhengwei at inspur.com Sun Dec 5 08:18:02 2021 From: suzhengwei at inspur.com (=?utf-8?B?U2FtIFN1ICjoi4/mraPkvJ8p?=) Date: Sun, 5 Dec 2021 08:18:02 +0000 Subject: =?utf-8?B?562U5aSNOiBbbGlzdHMub3BlbnN0YWNrLm9yZ+S7o+WPkV1ba29sbGEtYW5z?= =?utf-8?B?aWJsZV0gRGVmYXVsdCBtYXNha2FyaSBjb25maWd1cmF0aW9uIHNob3VsZCBl?= =?utf-8?Q?vacuate_instances_across_nodes_in_wallaby=3F?= References: <2e9bcb5dc06f50c55e5b65c775fd4d01@sslemail.net> Message-ID: <664003e28619450aa0c7529a673b0503@inspur.com> Hi, Rodrigo Lima? Have you create the failover segment and add the hosts into the segment? What is the ?enabled?flag of the segment? ???: Sam Su (???) ????: 2021?12?5? 16:05 ???: 'rodrigo.lima at o2sistemas.com' ??: 'openstack-discuss at lists.openstack.org' ??: ??: [lists.openstack.org??][kolla-ansible] Default masakari configuration should evacuate instances across nodes in wallaby? Hi, Rodrigo Lima? Have you found the failure notification in the masakari dashboard or DB( Table notifications)? If so, the failure notification would trigger the recovery workflow, and you need to check the masakari-api and masakari-engine log to find the detail. If not, it could be a deployment problem, and you need to check the connectivity between masakari-api and masakari-hostmonitor. ???: Rodrigo Lima [mailto:rodrigo.lima at o2sistemas.com] ????: 2021?12?4? 2:56 ???: openstack-discuss at lists.openstack.org ??: [lists.openstack.org??][kolla-ansible] Default masakari configuration should evacuate instances across nodes in wallaby? I'm working on upgrading an openstack farm from victoria to wallaby. After successful upgrade, I would like to enable hacluster and masakari to test HA between failing compute nodes. Everything seems to be running (pacemaker with the remote nodes OK, corosync without errors, masakari-monitors detects the 2 compute nodes online), but... when I simulate the failure of a node with shutdown, the failure and notification appears in the hostmonitor log, but the instances that were on the failed node don't evacuate, and I couldn't find documentation that explains how to do this specific configuration, if necessary. Does anyone have any ideas? -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: image001.jpg Type: image/jpeg Size: 823 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3606 bytes Desc: not available URL: From smooney at redhat.com Mon Dec 6 08:15:29 2021 From: smooney at redhat.com (Sean Mooney) Date: Mon, 06 Dec 2021 08:15:29 +0000 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> <5af59ddc-669f-09cf-233e-5f85f0e89999@debian.org> <865ebbe2-1615-08f3-da9e-8f95a8552754@debian.org> Message-ID: <3aea3ceacbe81b1ff05c88130a9fd26b677aa623.camel@redhat.com> On Fri, 2021-12-03 at 15:54 -0700, Alex Schultz wrote: > On Fri, Dec 3, 2021 at 3:45 PM Thomas Goirand wrote: > > > > On 12/3/21 11:34 PM, Alex Schultz wrote: > > > On Fri, Dec 3, 2021 at 3:22 PM Thomas Goirand wrote: > > > > > > > > On 12/3/21 10:41 PM, Julia Kreger wrote: > > > > > Extreme magic - OpenStack Edition. At least that feels like the answer. > > > > > > > > > > In essence, we create VMs. We attach them to the network, then > > > > > configure virtualbmc or sushy-tools to play the role of a BMC and one > > > > > of the things that library understand is the giant switch of "boot to > > > > > network". From there we ipxe boot. The VMs play the role of baremetal > > > > > machines through the emulation of the BMC and they are just hooked up > > > > > to the network. > > > > > > > > > > This does require the job to use root privileges to setup > > > > > networking/vms. And the virtualbmc or sushy-tools service needs enough > > > > > rights to toggle the VM's config. Would be happy to try and discuss > > > > > specifics of the devstack plugin, but I'm unsure if that would really > > > > > help. Regardless, you know where to find us. :) > > > > > > > > > > As an aside, when I first read the post I was thinking of booting from > > > > > ISO images. I've known people to embed ipxe into disk images, and you > > > > > can embed static config/instructions for booting into an ipxe binary. > > > > > It does require custom binary builds, but it might make what you're > > > > > trying to achieve easier. > > > > > > > > Hi Julia! > > > > > > > > Thanks for your answer. > > > > > > > > Well, I've read multiple times your reply, but I have to admit I still > > > > don't get all the details. For example, how would you "embed ipxe into > > > > disk images" for example? Should I for example run the iPXE binary from > > > > something like syslinux (or grub?)? But isn't iPXE usually running > > > > *after* the BIOS of a PC gets an IP from DHCP? > > > > > > > > > > I'm sure there are other examples around but the one we used to (maybe > > > still do) use for ovb can be built via: > > > > > > https://opendev.org/openstack/openstack-virtual-baremetal/src/branch/master/ipxe > > > > > > You basically build an image that just attempts ipxe on repeat. > > > > Hi Alex! > > > > In fact, I just found out my solution: http://boot.ipxe.org/ipxe.iso > > > > Yea I think that's what gets built with make and then it qemu-img > converts it for you. > > > Just download this, then: > > > > openstack image create \ > > --disk-format iso \ > > --container-format bare \ > > --file ipxe.iso ipxe.iso > > > > Et voila! The VM boots from PXE (I checked on the VNC console, and it > > also gets DHCP). I didn't go further, but it should be easy to setup the > > rest of the components (ie: a private network without DHCP, and a VM > > acting as DHCPd + tftp-hpa). > > > > Yea as previously mentioned you may just need to disable port security > and it should be fine. This is basically what we (tripleo) do in our > 3rd party testing where the dhcpd is from the undercloud. I've also > used the same image locally in libvirt + sushy-emulator for testing as > well. disabling port security might help but in general pxe booting nova instance is not supported. as in this is intentally not a supproted usecase currently, not that it coudl not be enable but we dont intend it to work outside of a ironic context.?disbaling port security really shoudl only be required if you want to run your own dhcp server, if you willing ot manually pxe boot interactivly you have ohter options. we also dont expclitly block pxe booting if the qemu seabios/uefi firmware support it so you can enable the boot menu vai hw:boot_menu=True extra spec or hw_boot_menu=True image property https://github.com/openstack/glance/blob/master/etc/metadefs/compute-libvirt.json#L24-L29 and select pxe boot if the firmware on the host loaded by qemu support it, but nova makes no guarentee that it will work with the libvirt driver. ironic is sort of a special case with regards to allowing/using pxe booting when its operating as a nova virt driver. in ironics case it more that nova does not care how the ironic servcei interally boots the instance and ironci just happens to use ipmi and pxe/ipxe as one of the options. you do not nesssiasarly need to use an ipxe boot image to pxeboot either as the seabios image, depening on your host useually has pxe supprot built in. your milage will vary however as this is outside the bounds of what is intended to work so if you encounterd issues adressing those would be new feature blueprint not bugs so they would not be backportable. if your intent is to just use this for testing then following in the path of tripleo and ironic by using virtualbmc to provide ipmi access to the vm might be ok but that assuems you are deploying an openstack cloud which you can then ssh into the hosts to install an run the virtual bmc. if you really want to do this in an unaltered cloud we would nee to configre the vm to boot form the network https://libvirt.org/formatdomain.html#bios-bootloader https://libvirt.org/formatdomain.html#specifying-boot-order im not sure what the best way to extend nova to support that would be, likely a neutron extention to mark ports a bootable which nova would read and then use to update the xml. host config is not really an option as it would be problematic for move operations, but flavor extra specs or image properites would be viable alternitive to the neutron approch. although neutorn ports would give a better ux. if you dont make progress with other approchs then you should consider fileing a spec for network boot support in nova/neutron. > > > I'll let you guys know how much progress I'm making with all of this > > (probably I'll start working on this next week...). > > > > Cheers, > > > > Thomas Goirand (zigo) > > > > From mark at stackhpc.com Mon Dec 6 08:40:33 2021 From: mark at stackhpc.com (Mark Goddard) Date: Mon, 6 Dec 2021 08:40:33 +0000 Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: References: <7d613f7f8f71497db28b5e27ddcb8224@verisign.com> Message-ID: On Wed, 1 Dec 2021 at 21:27, Pierre Riteau wrote: > > Hello Albert, > > The amqp_durable_queues configuration option comes from > oslo.messaging, which is a library used by many OpenStack projects. > You can set this option in the [oslo_messaging_rabbit] section of each > of these OpenStack projects (check their configuration reference for > more details). > > As for how to configure it with Kolla Ansible: you can either set it > directly for each service (e.g. in /etc/kolla/config/nova.conf to > configure it for Nova) or for all projects at once using > /etc/kolla/config/global.conf. Projects that don't support this option > should just ignore it. > > Read this section of the documentation for more details: > https://docs.openstack.org/kolla-ansible/latest/admin/advanced-configuration.html#openstack-service-configuration-in-kolla > > Best wishes, > Pierre Riteau Should we (Kolla) consider adding more of these to our defaults? Or documenting them in our RabbitMQ guide? Mark > > On Wed, 1 Dec 2021 at 19:20, Braden, Albert wrote: > > > > I read this with great interest because we are seeing this issue. Questions: > > > > 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? > > 2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? > > 3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? > > > > Does anyone have a sample set of RMQ config files that they can share? > > > > It looks like my Outlook has ruined the link; reposting: > > [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > > > -----Original Message----- > > From: Arnaud Morin > > Sent: Monday, November 29, 2021 2:04 PM > > To: Bogdan Dobrelya > > Cc: DHilsbos at performair.com; openstack-discuss at lists.openstack.org > > Subject: [EXTERNAL] Re: [ops]RabbitMQ High Availability > > > > Caution: This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe. > > > > Hi, > > > > After a talk on this ml (which starts at [1]), we endup building a > > documentation with Large Scale group. > > The doc is accessible at [2]. > > > > Hope this will help. > > > > [1] http://secure-web.cisco.com/1gFccuTyEVGnFd9aBOZ-RTPG0hbVIPGAbuLBNnoXP4onSZGFG1umIn0EtkEpBJWko4mi6yOUZ8Vsm-5sDGmIVl8rC2sOHv3Z2I1s9lFIkVFyn16CXJgcJbQQ7SBU8wEz5I_TysLtIY6YrmiC3PkKdG4oVCZk6n_KqYPYjmYUmDn9BD6JcXKUbFujVfugbjewZDY4HDCBnTe43tPSqkIZRVarApPiwsFtHu5PQ5riSoSgTpupqZHZdPnnGz7sbVGzx/http%3A%2F%2Flists.openstack.org%2Fpipermail%2Fopenstack-discuss%2F2020-August%2F016362.html > > [2] https://secure-web.cisco.com/1OtQ3pcnPPBNwevAFxS8yOS2xFlkHo0tY4SmkFE-wpAU_YPYS-BxRX5omcjCPZ3cMOxefnaO0vc3qlVm_SvI3DpkhejUkQUrrRbBJ72ki_ly13bYzC_QKd0-VERmSnlx8SFUB_DWewMYIZ7JfaURBYN9QvJgwD0b0aG-hYgvxcN1ZCt7qHTDqneGTtpx-5gRUMvld2dFz5uXsPj7QzohumP5bAoTblw7xLJy3zXhlfvrg6aHhQIR4xw9_y8E5Lt7d/https%3A%2F%2Fwiki.openstack.org%2Fwiki%2FLarge_Scale_Configuration_Rabbit > > > > On 24.11.21 - 11:31, Bogdan Dobrelya wrote: > > > On 11/24/21 12:34 AM, DHilsbos at performair.com wrote: > > > > All; > > > > > > > > In the time I've been part of this mailing list, the subject of RabbitMQ high availability has come up several times, and each time specific recommendations for both Rabbit and Open Stack are provided. I remember it being an A or B kind of recommendation (i.e. configure Rabbit like A1, and Open Stack like A2, OR configure Rabbit like B1, and Open Stack like B2). > > > > > > There is no special recommendations for rabbitmq setup for openstack, > > > but probably a few, like instead of putting it behind a haproxy, or the > > > like, list the rabbit cluster nodes in the oslo messaging config > > > settings directly. Also, it seems that durable queues makes a very > > > little sense for highly ephemeral RPC calls, just by design. I would > > > also add that the raft quorum queues feature of rabbitmq >=3.18 does > > > neither fit well into the oslo messaging design for RPC calls. > > > > > > A discussable and highly opinionated thing is also configuring > > > ha/mirror queue policy params for queues used for RPC calls vs > > > broad-casted notifications. > > > > > > And my biased personal humble recommendation is: use the upstream OCF RA > > > [0][1], if configuring rabbitmq cluster by pacemaker. > > > > > > [0] https://secure-web.cisco.com/1N1wD9gW7NZho0LdTVNuiU2ZIB7NW-eJMfDgVzBH3D3E6URzGYPKa-uhcLHxy3tRvRXopjnLAd2CECD1urJyRpg8NBSxTOEUSPxOlS0cQyULtSQuDbVWr-W7Bl3ZRcdWPrF9EuX_b40IM7zTjqS40gImsEouTqtD1vlCuEoaFgpptDEuMuaNTqBJ0IAtiZHuWiW6E7ufTtgxmVbkGLjXCZw5ZNhibbu-kGVyA-7MQsxQ-RBgSq5peTcLBR2Vx-f9k/https%3A%2F%2Fwww.rabbitmq.com%2Fpacemaker.html%23auto-pacemaker > > > > > > [1] > > > https://secure-web.cisco.com/1iDK1NnL9JTkQqkpBda06xTQNrWY2W0pVOTDwUoadfQbSXn5r0g_GH8PB8wZC5-JmHW2-m1YWoj1Z86jFcmWT0m9W9Sax5fJE5G7MbvQN2JM0EbAVHJDCmiBkMZlrSLoTgmh30RGhvmF9ww7jAjVnas3_AYFmwc65P-YtpdcswFC8rYcg5HlE2d979gf2OQUeftP3lfClkVou7hnELIFanDq07MfOJc2exHIfBo2ZQyUXRqXWUqnTsj7df-jCySkz/https%3A%2F%2Fgithub.com%2FClusterLabs%2Fresource-agents%2Fblob%2Fmaster%2Fheartbeat%2Frabbitmq-server-ha > > > > > > > > > > > Unfortunately, I can't find the previous threads on this topic. > > > > > > > > Does anyone have this information, that they would care to share with me? > > > > > > > > Thank you, > > > > > > > > Dominic L. Hilsbos, MBA > > > > Vice President - Information Technology > > > > Perform Air International Inc. > > > > DHilsbos at PerformAir.com > > > > www.PerformAir.com > > > > > > > > > > > > > > > > > > > > > -- > > > Best regards, > > > Bogdan Dobrelya, > > > Irc #bogdando > > > > > > > > > > > > > From mark at stackhpc.com Mon Dec 6 08:53:45 2021 From: mark at stackhpc.com (Mark Goddard) Date: Mon, 6 Dec 2021 08:53:45 +0000 Subject: [kolla] Plan to deprecate binary and unify on single distrubition In-Reply-To: <0cb15dca-99d0-9358-233d-4e85b5d1104d@debian.org> References: <7B0C2125-975E-437C-B332-9E88F725D248@gmail.com> <0cb15dca-99d0-9358-233d-4e85b5d1104d@debian.org> Message-ID: On Thu, 2 Dec 2021 at 08:19, Thomas Goirand wrote: > > Hi Michal, > > Thanks for your reply. > > I'm sorry to say that your reply shows a poor understanding of what > distributions and packaging are in general, and what Debian does for > OpenStack in particular. But I'm not surprised, this isn't the first > time I see this. It also happened when I was employed by Mirantis, and > with cloudwatt. People usually just dismiss packaging because they see > it as a black-box they cannot act on, because they just happen to not do > packaging themselves. > > On 12/2/21 8:37 AM, Micha? Nasiadka wrote: > > Hi Thomas, > > > > Thanks for your opinion - but we can?t really wait until each Debian OpenStack release to test all the features (and what will be broken) - we already trail long time after the official OpenStack release due to other issues. > > That would significantly lengthen the delay and the process involved. > > This isn't fair, and IMO isn't truth. > > Usually, Debian packages are ready *before* the final release, often > just a week after the first RCs are out. The only reason why there's > always one or 2 glitches, is because I couldn't install the new release > until everything is finished (and from my side, including updating my > puppet scripts for the new release of puppet-openstack). > > So, we're talking about just a few days delay after the release. This > normally gives you enough time to fix things, and have Kolla ready the > day of the release. > > I used to package all the beta releases, twice per OpenStack release. > Though nobody ever consumed them. And these days, not all projects are > release beta releases, so it is kind of useless. Currently, we provide 'binary' images for CentOS, Ubuntu and Debian. CentOS provides delorean builds from master, so we can test those on R-26. Ubuntu UCA appears soon after. Only for Debian do we need to wait until ~release. It's not a problem, since we're dropping binary images, I'm just providing some context. > > > It?s also a direction to make the external dependencies smaller - not bigger - and I see the custom things added from packager perspective as a dependency. > > IMO, that's a very wrong perspective. Package maintainers (that's how > we're called, not "packager" please...) are adding automations and > system integrations. If you refuse them, this means you'll have to > implement them yourself. Remember that I use puppet myself, so I very > well know what I'm talking about, when discussing config management and > package integration. When I can make a choice of where to do the > automation, system dependency, etc., then it goes to the packaging: this > is the normal and natural place to do it. > > A config management system should not be the one taking care of any of > the things that packages normally take care of (like dependencies). > Handling them in your config management system is not only an heresy, > it's also using the wrong tool for it, and achieving a poorer result. > > Also, handing dependency management using something like pip is a > non-sense, because it only knows about Python dependencies (not anything > else), and it's been demonstrated so many times how bad pip is as a > dependency solver: for example, it broke the CI so many times, showing > its defects. > > Another thing you're seeing wrong, is about smaller dependencies. First, > having one more Python package will not hurt, and if it is there, it is > for a reason. Second, packages are designed to have the smaller amount > of dependency possible. Always. As a rule. If there's something wrong > there, it's easy to fix (and you could contribute to it...). > > And we also make sure that everything works together, adding patches > when needed (how are you going to manage patches?). > > I'm not hoping to convince you, I feel it's already too late. Though I > can't help myself to let you know I believe you're doing the wrong > technical choice. This is all true - we do incur some additional work to build OpenStack from source. OTOH, this has been done and has been stable for many years. A major benefit we get from source images is being able to point at a local source repo when building images. Thankfully we need to do this less and less frequently as time goes on, but it is still sometimes necessary. > > Cheers, > > Thomas Goirand (zigo) > From zigo at debian.org Mon Dec 6 10:28:00 2021 From: zigo at debian.org (Thomas Goirand) Date: Mon, 6 Dec 2021 11:28:00 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: <3aea3ceacbe81b1ff05c88130a9fd26b677aa623.camel@redhat.com> References: <620fe1cf-cd03-4d76-8614-aeeb83cf103d@www.fastmail.com> <5af59ddc-669f-09cf-233e-5f85f0e89999@debian.org> <865ebbe2-1615-08f3-da9e-8f95a8552754@debian.org> <3aea3ceacbe81b1ff05c88130a9fd26b677aa623.camel@redhat.com> Message-ID: <47b722d3-312a-7941-6d8a-6babd3c36d2f@debian.org> Hi Sean, Thanks a lot for all of your valuable info. On 12/6/21 9:15 AM, Sean Mooney wrote: > disbaling port security really shoudl only be required if you want to run your own dhcp server This really is what I would like to do indeed. > you do not nesssiasarly need to use an ipxe boot image to pxeboot either as the seabios image, depening on your host useually has pxe supprot built in. Right, but by default, OpenStack will not do "-boot n" to make the BIOS do PXE. So booting the ipxe.iso image is what I found as the most easy way. Is there another way? > your milage will vary however as this is outside the bounds of what is intended to work so if you encounterd issues adressing those would be new feature blueprint > not bugs so they would not be backportable. Well, my only concern is if I must disable port security (because this needs admin credentials), though it should be possible to workaround this using "opensack port set --allowed-address" as a non-admin. > if your intent is to just use this for testing then following in the path of tripleo and ironic by using virtualbmc to provide ipmi access to the vm might be ok but that assuems you are deploying an > openstack cloud which you can then ssh into the hosts to install an run the virtual bmc. Instead of the virtual BMC which I didn't like much, I currently use ipmi_sim from openipmi. This feels the right choice. Maybe supporting it would be a nice addition to OpenStack. > if you really want to do this in an unaltered cloud we would nee to configre the vm to boot form the network > https://libvirt.org/formatdomain.html#bios-bootloader > https://libvirt.org/formatdomain.html#specifying-boot-order > > im not sure what the best way to extend nova to support that would be, likely a neutron extention to mark ports a bootable which nova would read and then use to update the xml. > host config is not really an option as it would be problematic for move operations, but flavor extra specs or image properites would be viable alternitive to the neutron approch. > although neutorn ports would give a better ux. > if you dont make progress with other approchs then you should consider fileing a spec for network boot support in nova/neutron. Ok, thanks for your advice. I'll try first with the ipxe.iso image, as this looks like the most easy path for the moment, and I'll see if later on I need more. Cheers, Thomas Goirand (zigo) From stephenfin at redhat.com Mon Dec 6 10:57:06 2021 From: stephenfin at redhat.com (Stephen Finucane) Date: Mon, 06 Dec 2021 10:57:06 +0000 Subject: [tc][all] Follow up pain point discussion In-Reply-To: References: Message-ID: On Fri, 2021-12-03 at 17:37 +0800, Rico Lin wrote: > Dear all > > The pain point collection etherpad [1] seems to get good attention from teams. > We plan to host a one-hour video meeting to keep the discussion going. > Please fill in the doodle to?select your preferred?time if you can join us: > https://doodle.com/poll/kur2mfywceerqcvh?utm_source=poll&utm_medium=link > The discussion will be hosted?in video?format. > > Here are some items (but not limited to) I expect we will discuss and > hopefully put some actions/efforts on. > * General Pain point: RabbitMQ > * General?Pain point:?OpenStackClient support I will try to join for this discussion, but one point on this for anyone else attending: please take the time before this discussion to try out a recent version of OSC. At the moment, our focus is on switching the underlying library used in OSC compute-related commands from novaclient to SDK and we plan to do the same for cinder next. We're doing this because afaict, there aren't any significant CLI gaps left for most of the core services. My understanding of the feature parity situation is as follows: * neutronclient???full feature parity since...a long time ago - Kudos, neutron team, for your long-term investment here * novaclient?? full feature parity ?since 5.5.0 - We finally introduced a proper '--block-device' option for 'server create' in 5.5.0, which closed our remaining RFE. There have been bugfixes since but it's feature complete, if not bug free. * cinderclient?? (effectively) full feature parity - The only remaining gap that I'm aware of is for configurable filters and volume service clusters. There are patches open for these?here?and?here?respectively that we simply need to merge to complete this, but it's consider low-ish priority * glanceclient?? not sure, as I've yet to investigate this. I suspect there are a few gaps here. * keystoneclient?? not sure, as I've yet to investigate this. I suspect we're very close if not there already though. So in summary, I don't think the OSC problem that has existed historically is still a real thing. If there are gaps that I'm not aware of, please be sure to communicate them here or on the Etherpad so we can track them and encourage people (future students, other contributors) to close them. Stephen > * General project pain point discussion: How to encourage teams to work/reply > on pain points. > Meanwhile, we encourage all to check the etherpad and provide your feedback. > Also if teams can help to answer questions/concerns on etherpad (like Nova > team does) will be greatly appreciated. > > [1] https://etherpad.opendev.org/p/pain-point-elimination > Rico Lin -------------- next part -------------- An HTML attachment was scrubbed... URL: From katonalala at gmail.com Mon Dec 6 12:19:04 2021 From: katonalala at gmail.com (Lajos Katona) Date: Mon, 6 Dec 2021 13:19:04 +0100 Subject: [neutron] Bug deputy report for week of November 29 Message-ID: Hi Neutron Team I was bug deputy in neutron last week. Summary ========== 3 bugs need more attention, one is related to db upgrade (upgrade from Rocky to Stein), one bug is an issue between Nova and Neutron and related to network-vif-plug event, and one seems to be related to IPv6 Metadata. If you have time please check them and help solve them and check the reviews I linked. Needs attention ================ * "neutron-db-manage upgrade" is interrupted with error about neutron.portuplinkstatuspropagation ( https://bugs.launchpad.net/neutron/+bug/1952675 ) * VirtualInterfaceCreateException due to "Timeout waiting for [('network-vif-plugged'..." (https://bugs.launchpad.net/neutron/+bug/1944779 ) * DHCP agent fails to fully configure DHCP namespaces because of duplicate address detected (https://bugs.launchpad.net/neutron/+bug/1953165 ) ** Seems to be an issue with IPv6 metadata. Bugs with assignees =================== * Neurton linuxbridge agent does not create VXLAN interface ( https://bugs.launchpad.net/neutron/+bug/1952611 ) ** Fix proposed by mnaser to oslo.privsep: https://review.opendev.org/c/openstack/oslo.privsep/+/819996 * Segment updates may cause unnecessary overload ( https://bugs.launchpad.net/neutron/+bug/1952730 ) ** fix from rubasov: https://review.opendev.org/c/openstack/neutron/+/819777 * Cannot setup IPv6 overlay network in xena ( https://bugs.launchpad.net/neutron/+bug/1952897 ) ** https://review.opendev.org/c/openstack/neutron/+/820376 * Allow modification of max retries in OVSNeutronAgent ( https://bugs.launchpad.net/neutron/+bug/1952898 ) ** https://review.opendev.org/c/openstack/neutron/+/819816 * Gratuitous ARPs are not sent during master transition ( https://bugs.launchpad.net/neutron/+bug/1952907 ): ** Possible fix from Liu: https://review.opendev.org/c/openstack/neutron/+/712474 * On IPv6 overlay networks linuxbridge vxlans are created always on loopback device (https://bugs.launchpad.net/neutron/+bug/1953139 ): ** https://review.opendev.org/c/openstack/neutron/+/820484 RFEs ====== * [RFE] Unify quota engine API ( https://bugs.launchpad.net/neutron/+bug/1953170 ) ** We discussed the need for this rfe on last Friday's drivers meeting: https://meetings.opendev.org/meetings/neutron_drivers/2021/neutron_drivers.2021-12-03-14.01.log.html#l-74 Regards Lajos Katona (lajoskatona) -------------- next part -------------- An HTML attachment was scrubbed... URL: From elod.illes at est.tech Mon Dec 6 13:51:52 2021 From: elod.illes at est.tech (=?UTF-8?B?RWzFkWQgSWxsw6lz?=) Date: Mon, 6 Dec 2021 14:51:52 +0100 Subject: [cinder][OpenstackAnsible] Xena Cycle-Trailing Release Deadline Message-ID: <4321ec6c-0acf-ae0c-19d2-6fb0291f233e@est.tech> Hello teams with trailing projects, Next week is the Xena cycle-trailing release deadline [1], and all projects following the cycle-trailing release model must release their Xena deliverables by 16 December, 2021. The following trailing projects haven't been released yet for Xena: Cinder team's deliverables: - cinderlib OSA team's deliverables: - openstack-ansible-roles - openstack-ansible Patches were generated 2 weeks ago [2], feel free to update those if they are not up to date anymore. This is just a friendly reminder to allow you to release these projects in time. Do not hesitate to ping us if you have any questions or concerns. [1] https://releases.openstack.org/yoga/schedule.html#y-cycle-trail [2] https://review.opendev.org/q/topic:trailing-xena El?d Ill?s irc: elodilles -------------- next part -------------- An HTML attachment was scrubbed... URL: From bdobreli at redhat.com Mon Dec 6 14:18:47 2021 From: bdobreli at redhat.com (Bogdan Dobrelya) Date: Mon, 6 Dec 2021 15:18:47 +0100 Subject: [ops] [kolla] RabbitMQ High Availability Message-ID: > I read this with great interest because we are seeing this issue. Questions: > > 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? > 2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? > 3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? Note that even having rabbit HA policies adjusted like that and its HA replication factor [0] decreased (e.g. to a 2), there still might be high churn caused by a large enough number of replicated durable RPC topic queues. And that might cripple the cloud down with the incurred I/O overhead because a durable queue requires all messages in it to be persisted to a disk (for all the messaging cluster replicas) before they are ack'ed by the broker. Given that said, Oslo messaging would likely require a more granular control for topic exchanges and the durable queues flag - to tell it to declare as durable only the most critical paths of a service. A single config setting and a single control exchange per a service might be not enough. There are also race conditions with durable queues enabled, like [1]. A solution could be where each service declare its own dedicated control exchange with its own configuration. Finally, openstack components should add perhaps a *.next CI job to test it with durable queues, like [2] [0] https://www.rabbitmq.com/ha.html#replication-factor [1] https://zuul.opendev.org/t/openstack/build/aa514dd788f34cc1be3800e6d7dba0e8/log/controller/logs/screen-n-cpu.txt [2] https://review.opendev.org/c/openstack/nova/+/820523 > > Does anyone have a sample set of RMQ config files that they can share? > > It looks like my Outlook has ruined the link; reposting: > [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit -- Best regards, Bogdan Dobrelya, Irc #bogdando From rosmaita.fossdev at gmail.com Mon Dec 6 14:20:57 2021 From: rosmaita.fossdev at gmail.com (Brian Rosmaita) Date: Mon, 6 Dec 2021 09:20:57 -0500 Subject: [cinder] midcycle R-17 followup meeting thursday Message-ID: <6d97efb4-e451-2577-4e19-65a2c25bd2f1@gmail.com> The poll results are in and our midcycle followup meeting will be held: DATE: Thursday, 9 December 2021 TIME: 1400-1430 UTC LOCATION: https://bluejeans.com/3228528973 The meeting will be recorded. The midcycle etherpad is here: https://etherpad.opendev.org/p/cinder-yoga-midcycles (Followup info starts at line 34, unless some joker has added a bunch of stuff at the top of the etherpad.) The primary topic is Gorka's proposal to revamp the quota system. There are links to the proposed spec and some POC patches on the etherpad. It's worth taking a look in advance so that we can have a short and productive meeting! cheers, brian From hanguangyu at uniontech.com Mon Dec 6 11:49:37 2021 From: hanguangyu at uniontech.com (=?utf-8?B?6Z+p5YWJ5a6H?=) Date: Mon, 6 Dec 2021 19:49:37 +0800 Subject: [neutron] Instances can't get IP from the DHCP server in OpenStack Message-ID: Hi, I'm trying to create instance in OpenStack Victoria. But, I am facing a issues about dhcp of neutron. I have OpenStack Victoria running in one baremetl server with uniontech os(a downstream of centos 8). I have a Flat network created in the range of 10.12.21.190-10.12.21.195. I selected to have a DHCP. The instance ran, and neutron-dhcp-agent.service had allocated a IP to it.  # openstack server list +--------------------------------------+-------+--------+-----------------------+----------------+--------+ | ID | Name | Status | Networks | Image | Flavor | +--------------------------------------+-------+--------+-----------------------+----------------+--------+ | 4ccae37e-fbfe-4acb-a109-e1bc9175c2e0 | inst1 | ACTIVE | provider=10.12.21.192 | centos8-server | h2 | +--------------------------------------+-------+--------+-----------------------+----------------+--------+ But the instance can't get a response from DHCP.  No any error in log. And if I setup the ip manually in the instance I can get access to the gateway, dhcp port and external network. I can catpure the DHCP request in any network device in the server, include dhcp port(10.12.21.190). But no response.  Dnsmasq process was running and had  # openstack port list +--------------------------------------+------+-------------------+-----------------------------------------------------------------------------+--------+ | ID | Name | MAC Address | Fixed IP Addresses | Status | +--------------------------------------+------+-------------------+-----------------------------------------------------------------------------+--------+ | 5ab32ca8-2d1d-47cf-9a62-7c01d21abaf0 | | fa:16:3e:1d:48:45 | ip_address='10.12.21.192', subnet_id='3daf5a55-e76a-4093-8533-78d9464b1beb' | ACTIVE | | 63fb2306-7759-4933-8d93-590e3a56f315 | | fa:16:3e:c5:59:9a | ip_address='10.12.21.190', subnet_id='3daf5a55-e76a-4093-8533-78d9464b1beb' | ACTIVE | +--------------------------------------+------+-------------------+-----------------------------------------------------------------------------+--------+ # ip netns qdhcp-ec2c4d9d-888b-4312-b0af-ab2127b76e0e (id: 0) # sudo ip netns exec qdhcp-ec2c4d9d-888b-4312-b0af-ab2127b76e0e sudo tcpdump -n -S -i tap63fb2306-77|grep DHCP ## notes: fa:16:3e:1d:48:45 is the mac of nic of inst1 ? dropped privs to tcpdump tcpdump: verbose output suppressed, use -v or -vv for full protocol decode listening on tap63fb2306-77, link-type EN10MB (Ethernet), capture size 262144 bytes 16:47:18.194028 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, Request from fa:16:3e:1d:48:45, length 276 16:47:21.146736 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, Request from fa:16:3e:1d:48:45, length 276 16:47:25.226387 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, Request from fa:16:3e:1d:48:45, length 276 more detail informations: # ps -ef|grep dnsmasq dnsmasq 7441 1 0 12?03 ? 00:00:00 dnsmasq --no-hosts --no-resolv --pid-file=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/pid --dhcp-hostsfile=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/host --addn-hosts=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/addn_hosts --dhcp-optsfile=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/opts --dhcp-leasefile=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/leases --dhcp-match=set:ipxe,175 --dhcp-userclass=set:ipxe6,iPXE --local-service --bind-dynamic --dhcp-range=set:subnet-3daf5a55-e76a-4093-8533-78d9464b1beb,10.12.21.0,static,255.255.255.0,86400s --dhcp-option-force=option:mtu,1500 --dhcp-lease-max=256 --conf-file= --domain=openstacklocal dnsmasq 317868 1 0 14:33 ? 00:00:00 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/libexec/libvirt_leaseshelper root 317869 317868 0 14:33 ? 00:00:00 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/libexec/libvirt_leaseshelper root 340315 317310 0 18:43 pts/2 00:00:00 grep --color=auto dnsmasq [root at a16 ~]# cat /var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/host fa:16:3e:c5:59:9a,host-10-12-21-190.openstacklocal,10.12.21.190 fa:16:3e:1d:48:45,host-10-12-21-192.openstacklocal,10.12.21.192 # ovs-vsctl show adbd4b0c-cb78-4b8f-b8a6-197c5948312c Manager "ptcp:6640:127.0.0.1" is_connected: true Bridge br0 Controller "tcp:127.0.0.1:6633" is_connected: true fail_mode: secure datapath_type: system Port enp5s0 Interface enp5s0 Port phy-br0 Interface phy-br0 type: patch options: {peer=int-br0} Port br0 Interface br0 type: internal Bridge br-tun Controller "tcp:127.0.0.1:6633" is_connected: true fail_mode: secure datapath_type: system Port br-tun Interface br-tun type: internal Port patch-int Interface patch-int type: patch options: {peer=patch-tun} Bridge br-int Controller "tcp:127.0.0.1:6633" is_connected: true fail_mode: secure datapath_type: system Port int-br0 Interface int-br0 type: patch options: {peer=phy-br0} Port br-int Interface br-int type: internal Port tap63fb2306-77 tag: 1 Interface tap63fb2306-77 type: internal Port patch-tun Interface patch-tun type: patch options: {peer=patch-int} Port qvo5ab32ca8-2d tag: 1 Interface qvo5ab32ca8-2d ovs_version: "2.13.0". # ip a 1: lo: -------------- next part -------------- A non-text attachment was scrubbed... Name: 3008DC07 at 74C0114D.D1F8AD61 Type: application/octet-stream Size: 31847 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: 3307DD07 at EEDEB41E.D1F8AD61 Type: application/octet-stream Size: 8758 bytes Desc: not available URL: From hanguangyu at uniontech.com Mon Dec 6 12:49:11 2021 From: hanguangyu at uniontech.com (=?utf-8?B?6Z+p5YWJ5a6H?=) Date: Mon, 6 Dec 2021 20:49:11 +0800 Subject: [neutron] Instances can't get DHCP reply from neutron In-Reply-To: References: Message-ID: Hi, I'm trying to create instance in OpenStack Victoria. But, I am facing an issues about dhcp of neutron.I have OpenStack Victoria running in one baremetl server with uniontech os (a downstream of centos 8). I have a Flat network created in the range of 10.12.21.190- 10.12.21.195. I selected to have a DHCP. The instance ran, and neutron-dhcp-agent.service had allocated an IP to it. # openstack server list +--------------------------------------+-------+--------+-----------------------+----------------+--------+ | ID | Name | Status | Networks | Image | Flavor | +--------------------------------------+-------+--------+-----------------------+----------------+--------+ | 4ccae37e-fbfe-4acb-a109-e1bc9175c2e0 | inst1 | ACTIVE | provider=10.12.21.192 | centos8-server | h2 | +--------------------------------------+-------+--------+-----------------------+----------------+--------+ But the instance can't get a response from DHCP. No any error in log. And if I setup the ip manually in the instance I can get access to the gateway, dhcp port and external network. I can catpure the DHCP request in any network device in the server, include dhcp port. But no response. Dnsmasq process was running and had allocated an IP for instance. For details information: https://paste.opendev.org/show/811477/ I would very very appreciate any kind of guidance or help. Thanks, Han Guangyu From christian.rohmann at inovex.de Mon Dec 6 17:05:46 2021 From: christian.rohmann at inovex.de (Christian Rohmann) Date: Mon, 6 Dec 2021 18:05:46 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: Message-ID: Hey all, On 01/12/2021 23:33, Thomas Goirand wrote: > In order to test OCI [1] within OpenStack itself, I would need my > instances to do PXE booting. This would be with something like this on > the qemu command line: > > qemu-system-x86_64 -boot n > > I believe this would also mean setting-up network cards that can do PXE > booting, like the e1000, which Qemu supports. > > If that is possible, then I would setup DHCP + tftp-hpa on another > instance, wired on the same private network. On 03/12/2021 21:01, Cedric Lemarchand wrote: > We faced the same use case, and we end up doing something like this: > > - disable security on the network provider (or on the instance port > where pxe boot is needed) > - deploy or rescue the intance with an ipxe image I ran into this exact issue in the past, searching for possibilities to boot Nova instances using PXE. Certainly the actual boot server should not be provided by Nova in any way - it's more about getting a VM to network boot. Actually the only change required would be to enable netboot via the boot order: ? * https://libvirt.org/formatdomain.html#specifying-boot-order There also have been blueprints or implementation attempts for nova before: ?* https://blueprints.launchpad.net/nova/+spec/boot-order-for-instance ?* https://review.opendev.org/c/openstack/nova-specs/+/133254/ Maybe if someone from the Nova devs would give an indication if supporting network boot would? be worth discussing? The most basic use-case certainly are staging environments for physical (server) systems that do network boot. But I believe there likely are more ideas others could contribute. Christian From ashlee at openstack.org Mon Dec 6 18:36:14 2021 From: ashlee at openstack.org (Ashlee Ferguson) Date: Mon, 6 Dec 2021 12:36:14 -0600 Subject: OpenInfra Live - December 9, 2021 at 9am CT Message-ID: <81A13845-D57D-4232-97C4-4B6402D47F14@openstack.org> Hi everyone, This week?s OpenInfra Live episode is brought to you by the OpenStack community. This new episode of the ?Large Scale OpenStack? show will discuss operators' tricks and tools. In a live and direct discussion, our guests will reveal tricks and homegrown tools that they use in the trenches to keep their OpenStack clusters ticking like clockwork. Episode: Large Scale OpenStack: Operators? tricks and tools Date and time: December 9, 2021 at 9am CT (1500 UTC) You can watch us live on: YouTube: https://www.youtube.com/watch?v=F_9KKAQE4fc LinkedIn: https://www.linkedin.com/feed/update/urn:li:ugcPost:6873686260419055616/ Facebook: https://www.facebook.com/104139126308032/posts/4596356560419577/ WeChat: recording will be posted on OpenStack WeChat after the live stream Speakers: Shatadru Bandyopadhyay (Workday) Thierry Carrez (OpenInfra Foundation) Thomas Goirand (Infomaniak) Axel Jacquet (Infomaniak) Gene Kuo (LINE) Belmiro Moreira (CERN) Arnaud Morin (OVHCloud) Adrien Pensart (OVHCloud) Have an idea for a future episode? Share it now at ideas.openinfra.live. Thanks! Ashlee Ashlee Ferguson Carlson Community & Events Open Infrastructure Foundation -------------- next part -------------- An HTML attachment was scrubbed... URL: From Ricardo.SoaresSarto at windriver.com Mon Dec 6 18:57:57 2021 From: Ricardo.SoaresSarto at windriver.com (Soares Sarto, Ricardo) Date: Mon, 6 Dec 2021 18:57:57 +0000 Subject: [dev][panko] deprecation doubts Message-ID: Hi Everyone Now that Panko has been deprecated, where are the events stored? What is the evolution strategy for OpenStack Telemetry? What is a current list of data generated by core openstack services (nova, neutron, keystone, glance, cinder, heat, etc.) to ceilometer ? (e.g. this link is from ocata ? where is the current one ? https://docs.openstack.org/ocata/admin-guide/telemetry-data-collection.html ) Thank you! -------------- next part -------------- An HTML attachment was scrubbed... URL: From rodrigo.lima at o2sistemas.com Mon Dec 6 20:11:22 2021 From: rodrigo.lima at o2sistemas.com (Rodrigo Lima) Date: Mon, 6 Dec 2021 17:11:22 -0300 Subject: [kolla-ansible] Default masakari configuration should evacuate instances across nodes in wallaby? In-Reply-To: References: Message-ID: Hi! It?s a lot of information. From engine: 2021-12-06 20:02:33.502 7 INFO masakari.engine.manager [req-034e0a9e-1b34-4db3-a2e0-6d61a369ba6c 6c128366b66346d38fb5493adf0cf666 e39a7ea7d17046b5b97b2253bf8195bc - - -] Processing notification 2468563d-70c5-41ef-ad04-a51b4ad3dd4d of type: COMPUTE_HOST 2021-12-06 20:02:34.052 7 INFO masakari.compute.nova [req-882d4151-0549-4cb2-acbb-90509073179c nova - - - -] Disable nova-compute on ctl02-hml.amt.net.br 2021-12-06 20:02:34.113 7 INFO masakari.engine.drivers.taskflow.host_failure [req-882d4151-0549-4cb2-acbb-90509073179c nova - - - -] Sleeping 180 sec before starting recovery thread until nova recognizes the node down. 2021-12-06 20:05:34.130 7 INFO masakari.compute.nova [req-c570bdef-0786-47bd-94fc-e6c1396eabb1 nova - - - -] Fetch Server list on ctl02-hml.amt.net.br 2021-12-06 20:05:35.166 7 INFO masakari.compute.nova [req-adec6e54-15ff-4a28-98c3-01070ba2cf87 nova - - - -] Call get server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:35.949 7 INFO masakari.compute.nova [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Call get server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:35.955 7 INFO masakari.compute.nova [req-e548525d-74f7-42c4-8045-2f63d645f76f nova - - - -] Call get server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:36.739 7 INFO masakari.compute.nova [req-9cea2ff2-0d32-42a1-8605-b889d95cadc0 nova - - - -] Call lock server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:36.747 7 INFO masakari.compute.nova [req-bc3bdea3-e923-4c08-bf84-bea711465dce nova - - - -] Call get server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:37.391 7 INFO masakari.compute.nova [req-9cea2ff2-0d32-42a1-8605-b889d95cadc0 nova - - - -] Call evacuate command for instance 618e44e8-248f-4f50-a760-581972352af8 on host None 2021-12-06 20:05:37.478 7 INFO masakari.compute.nova [req-84d38844-ae87-4767-b822-0099bc40aa16 nova - - - -] Call lock server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:38.059 7 INFO masakari.compute.nova [req-d57bd685-2e42-48e0-bf7d-9978ac516451 nova - - - -] Call get server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:38.102 7 INFO masakari.compute.nova [req-84d38844-ae87-4767-b822-0099bc40aa16 nova - - - -] Call evacuate command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 on host None 2021-12-06 20:05:38.734 7 INFO masakari.compute.nova [req-25b4bb9c-2e05-409a-bae8-fd7c8348ba5b nova - - - -] Call get server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:39.062 7 INFO masakari.compute.nova [req-3d69363b-aa6c-4752-be52-285bc36f25c2 nova - - - -] Call get server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:39.732 7 INFO masakari.compute.nova [req-425752db-0599-42a8-bf2a-0229148410cc nova - - - -] Call get server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall [req-3d69363b-aa6c-4752-be52-285bc36f25c2 nova - - - -] Fixed interval looping call 'masakari.engine.drivers.taskflow.host_failure.EvacuateInstancesTask._evacuate_and_confirm.._wait_for_evacuation_confirmation' failed: masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall Traceback (most recent call last): 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_service/loopingcall.py", line 150, in _run_loop 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall result = func(*self.args, **self.kw) 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall File "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", line 207, in _wait_for_evacuation_confirmation 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall raise exception.InstanceEvacuateFailed( 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall 2021-12-06 20:05:39.790 7 WARNING masakari.engine.drivers.taskflow.host_failure [req-d537ed5e-f5c8-4ba5-b559-0156883ca02b nova - - - -] Failed to evacuate instance 618e44e8-248f-4f50-a760-581972352af8: masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:39.793 7 INFO masakari.compute.nova [req-c02a47f3-fc23-4d5b-9167-cf21bf8923db nova - - - -] Call unlock server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall [req-425752db-0599-42a8-bf2a-0229148410cc nova - - - -] Fixed interval looping call 'masakari.engine.drivers.taskflow.host_failure.EvacuateInstancesTask._evacuate_and_confirm.._wait_for_evacuation_confirmation' failed: masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall Traceback (most recent call last): 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_service/loopingcall.py", line 150, in _run_loop 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall result = func(*self.args, **self.kw) 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall File "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", line 207, in _wait_for_evacuation_confirmation 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall raise exception.InstanceEvacuateFailed( 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall 2021-12-06 20:05:40.423 7 WARNING masakari.engine.drivers.taskflow.host_failure [req-f8061ba7-353b-487c-8813-30f85216335f nova - - - -] Failed to evacuate instance 746178b2-14ce-4ce2-83f6-2bf9d613a887: masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:40.425 7 INFO masakari.compute.nova [req-7f0c16fe-8eaa-485b-a414-945fcd84a3c6 nova - - - -] Call unlock server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:41.080 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task 'EvacuateInstancesTask' (e49cc088-a9d7-4dd8-a6db-75339c6eaa4d) transitioned into state 'FAILURE' from state 'RUNNING' 4 predecessors (most recent first): Flow 'post_tasks' |__Flow 'main_tasks' |__Flow 'pre_tasks' |__Flow 'instance_evacuate_engine': masakari.exception.HostRecoveryFailureException: Failed to evacuate instances '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' from host 'ctl02-hml.amt.net.br' 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver Traceback (most recent call last): 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver File "/var/lib/kolla/venv/lib/python3.8/site-packages/taskflow/engines/action_engine/executor.py", line 53, in _execute_task 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver result = task.execute(**arguments) 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver File "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", line 396, in execute 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver _do_evacuate(self.context, host_name, instance_list) 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver File "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", line 376, in _do_evacuate 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver raise exception.HostRecoveryFailureException( 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver masakari.exception.HostRecoveryFailureException: Failed to evacuate instances '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' from host 'ctl02-hml.amt.net.br' 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver 2021-12-06 20:05:41.088 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task 'EvacuateInstancesTask' (e49cc088-a9d7-4dd8-a6db-75339c6eaa4d) transitioned into state 'REVERTED' from state 'REVERTING' 2021-12-06 20:05:41.091 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task 'PrepareHAEnabledInstancesTask' (21c32e80-7521-44bb-bd28-53f88c3d13da) transitioned into state 'REVERTED' from state 'REVERTING' 2021-12-06 20:05:41.093 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task 'DisableComputeServiceTask' (b3522750-c7a6-4f71-ad8c-e2a61a40e2b8) transitioned into state 'REVERTED' from state 'REVERTING' 2021-12-06 20:05:41.095 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Flow 'instance_evacuate_engine' (3c8c5def-39b8-4956-8b2c-62a218e612ee) transitioned into state 'REVERTED' from state 'RUNNING' 2021-12-06 20:05:41.096 7 ERROR masakari.engine.manager [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Failed to process notification '2468563d-70c5-41ef-ad04-a51b4ad3dd4d'. Reason: Failed to evacuate instances '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' from host 'ctl02-hml.amt.net.br': masakari.exception.HostRecoveryFailureException: Failed to evacuate instances '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' from host 'ctl02-hml.amt.net.br' 2021-12-06 20:05:41.099 7 INFO masakari.engine.manager [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Notification 2468563d-70c5-41ef-ad04-a51b4ad3dd4d exits with status: error. 2021-12-06 20:07:08.803 7 INFO masakari.engine.manager [req-d2c18313-253f-4bb2-8abb-5ca5d8ac248a nova - - - -] Processing notification 2468563d-70c5-41ef-ad04-a51b4ad3dd4d of type: COMPUTE_HOST 2021-12-06 20:07:09.398 7 INFO masakari.compute.nova [req-5b595769-209f-422b-a71f-ef4507187a61 nova - - - -] Disable nova-compute on ctl02-hml.amt.net.br 2021-12-06 20:07:09.467 7 INFO masakari.engine.drivers.taskflow.host_failure [req-5b595769-209f-422b-a71f-ef4507187a61 nova - - - -] Sleeping 180 sec before starting recovery thread until nova recognizes the node down.2021-12-06 20:02:33.502 7 INFO masakari.engine.manager [req-034e0a9e-1b34-4db3-a2e0-6d61a369ba6c 6c128366b66346d38fb5493adf0cf666 e39a7ea7d17046b5b97b2253bf8195bc - - -] Processing notification 2468563d-70c5-41ef-ad04-a51b4ad3dd4d of type: COMPUTE_HOST 2021-12-06 20:02:34.052 7 INFO masakari.compute.nova [req-882d4151-0549-4cb2-acbb-90509073179c nova - - - -] Disable nova-compute on ctl02-hml.amt.net.br 2021-12-06 20:02:34.113 7 INFO masakari.engine.drivers.taskflow.host_failure [req-882d4151-0549-4cb2-acbb-90509073179c nova - - - -] Sleeping 180 sec before starting recovery thread until nova recognizes the node down. 2021-12-06 20:05:34.130 7 INFO masakari.compute.nova [req-c570bdef-0786-47bd-94fc-e6c1396eabb1 nova - - - -] Fetch Server list on ctl02-hml.amt.net.br 2021-12-06 20:05:35.166 7 INFO masakari.compute.nova [req-adec6e54-15ff-4a28-98c3-01070ba2cf87 nova - - - -] Call get server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:35.949 7 INFO masakari.compute.nova [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Call get server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:35.955 7 INFO masakari.compute.nova [req-e548525d-74f7-42c4-8045-2f63d645f76f nova - - - -] Call get server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:36.739 7 INFO masakari.compute.nova [req-9cea2ff2-0d32-42a1-8605-b889d95cadc0 nova - - - -] Call lock server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:36.747 7 INFO masakari.compute.nova [req-bc3bdea3-e923-4c08-bf84-bea711465dce nova - - - -] Call get server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:37.391 7 INFO masakari.compute.nova [req-9cea2ff2-0d32-42a1-8605-b889d95cadc0 nova - - - -] Call evacuate command for instance 618e44e8-248f-4f50-a760-581972352af8 on host None 2021-12-06 20:05:37.478 7 INFO masakari.compute.nova [req-84d38844-ae87-4767-b822-0099bc40aa16 nova - - - -] Call lock server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:38.059 7 INFO masakari.compute.nova [req-d57bd685-2e42-48e0-bf7d-9978ac516451 nova - - - -] Call get server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:38.102 7 INFO masakari.compute.nova [req-84d38844-ae87-4767-b822-0099bc40aa16 nova - - - -] Call evacuate command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 on host None 2021-12-06 20:05:38.734 7 INFO masakari.compute.nova [req-25b4bb9c-2e05-409a-bae8-fd7c8348ba5b nova - - - -] Call get server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:39.062 7 INFO masakari.compute.nova [req-3d69363b-aa6c-4752-be52-285bc36f25c2 nova - - - -] Call get server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:39.732 7 INFO masakari.compute.nova [req-425752db-0599-42a8-bf2a-0229148410cc nova - - - -] Call get server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall [req-3d69363b-aa6c-4752-be52-285bc36f25c2 nova - - - -] Fixed interval looping call 'masakari.engine.drivers.taskflow.host_failure.EvacuateInstancesTask._evacuate_and_confirm.._wait_for_evacuation_confirmation' failed: masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall Traceback (most recent call last): 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_service/loopingcall.py", line 150, in _run_loop 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall result = func(*self.args, **self.kw) 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall File "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", line 207, in _wait_for_evacuation_confirmation 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall raise exception.InstanceEvacuateFailed( 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall 2021-12-06 20:05:39.790 7 WARNING masakari.engine.drivers.taskflow.host_failure [req-d537ed5e-f5c8-4ba5-b559-0156883ca02b nova - - - -] Failed to evacuate instance 618e44e8-248f-4f50-a760-581972352af8: masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:39.793 7 INFO masakari.compute.nova [req-c02a47f3-fc23-4d5b-9167-cf21bf8923db nova - - - -] Call unlock server command for instance 618e44e8-248f-4f50-a760-581972352af8 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall [req-425752db-0599-42a8-bf2a-0229148410cc nova - - - -] Fixed interval looping call 'masakari.engine.drivers.taskflow.host_failure.EvacuateInstancesTask._evacuate_and_confirm.._wait_for_evacuation_confirmation' failed: masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall Traceback (most recent call last): 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall File "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_service/loopingcall.py", line 150, in _run_loop 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall result = func(*self.args, **self.kw) 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall File "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", line 207, in _wait_for_evacuation_confirmation 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall raise exception.InstanceEvacuateFailed( 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall 2021-12-06 20:05:40.423 7 WARNING masakari.engine.drivers.taskflow.host_failure [req-f8061ba7-353b-487c-8813-30f85216335f nova - - - -] Failed to evacuate instance 746178b2-14ce-4ce2-83f6-2bf9d613a887: masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:40.425 7 INFO masakari.compute.nova [req-7f0c16fe-8eaa-485b-a414-945fcd84a3c6 nova - - - -] Call unlock server command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 2021-12-06 20:05:41.080 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task 'EvacuateInstancesTask' (e49cc088-a9d7-4dd8-a6db-75339c6eaa4d) transitioned into state 'FAILURE' from state 'RUNNING' 4 predecessors (most recent first): Flow 'post_tasks' |__Flow 'main_tasks' |__Flow 'pre_tasks' |__Flow 'instance_evacuate_engine': masakari.exception.HostRecoveryFailureException: Failed to evacuate instances '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' from host 'ctl02-hml.amt.net.br' 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver Traceback (most recent call last): 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver File "/var/lib/kolla/venv/lib/python3.8/site-packages/taskflow/engines/action_engine/executor.py", line 53, in _execute_task 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver result = task.execute(**arguments) 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver File "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", line 396, in execute 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver _do_evacuate(self.context, host_name, instance_list) 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver File "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", line 376, in _do_evacuate 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver raise exception.HostRecoveryFailureException( 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver masakari.exception.HostRecoveryFailureException: Failed to evacuate instances '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' from host 'ctl02-hml.amt.net.br' 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver 2021-12-06 20:05:41.088 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task 'EvacuateInstancesTask' (e49cc088-a9d7-4dd8-a6db-75339c6eaa4d) transitioned into state 'REVERTED' from state 'REVERTING' 2021-12-06 20:05:41.091 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task 'PrepareHAEnabledInstancesTask' (21c32e80-7521-44bb-bd28-53f88c3d13da) transitioned into state 'REVERTED' from state 'REVERTING' 2021-12-06 20:05:41.093 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task 'DisableComputeServiceTask' (b3522750-c7a6-4f71-ad8c-e2a61a40e2b8) transitioned into state 'REVERTED' from state 'REVERTING' 2021-12-06 20:05:41.095 7 WARNING masakari.engine.drivers.taskflow.driver [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Flow 'instance_evacuate_engine' (3c8c5def-39b8-4956-8b2c-62a218e612ee) transitioned into state 'REVERTED' from state 'RUNNING' 2021-12-06 20:05:41.096 7 ERROR masakari.engine.manager [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Failed to process notification '2468563d-70c5-41ef-ad04-a51b4ad3dd4d'. Reason: Failed to evacuate instances '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' from host 'ctl02-hml.amt.net.br': masakari.exception.HostRecoveryFailureException: Failed to evacuate instances '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' from host 'ctl02-hml.amt.net.br' 2021-12-06 20:05:41.099 7 INFO masakari.engine.manager [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Notification 2468563d-70c5-41ef-ad04-a51b4ad3dd4d exits with status: error. 2021-12-06 20:07:08.803 7 INFO masakari.engine.manager [req-d2c18313-253f-4bb2-8abb-5ca5d8ac248a nova - - - -] Processing notification 2468563d-70c5-41ef-ad04-a51b4ad3dd4d of type: COMPUTE_HOST 2021-12-06 20:07:09.398 7 INFO masakari.compute.nova [req-5b595769-209f-422b-a71f-ef4507187a61 nova - - - -] Disable nova-compute on ctl02-hml.amt.net.br 2021-12-06 20:07:09.467 7 INFO masakari.engine.drivers.taskflow.host_failure [req-5b595769-209f-422b-a71f-ef4507187a61 nova - - - -] Sleeping 180 sec before starting recovery thread until nova recognizes the node down. I tested forcing a kernel panic on target node Em s?b., 4 de dez. de 2021 ?s 08:48, Rados?aw Piliszek < radoslaw.piliszek at gmail.com> escreveu: > What do the Masakari API and Masakari Engine logs show? > > -yoctozepto > > On Fri, 3 Dec 2021 at 19:57, Rodrigo Lima > wrote: > >> I'm working on upgrading an openstack farm from victoria to wallaby. >> After successful upgrade, I would like to enable hacluster and masakari to >> test HA between failing compute nodes. Everything seems to be running >> (pacemaker with the remote nodes OK, corosync without errors, >> masakari-monitors detects the 2 compute nodes online), but... when I >> simulate the failure of a node with shutdown, the failure and notification >> appears in the hostmonitor log, but the instances that were on the failed >> node don't evacuate, and I couldn't find documentation that explains how to >> do this specific configuration, if necessary. Does anyone have any ideas? >> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: From radoslaw.piliszek at gmail.com Mon Dec 6 20:41:03 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Mon, 6 Dec 2021 21:41:03 +0100 Subject: [kolla-ansible] Default masakari configuration should evacuate instances across nodes in wallaby? In-Reply-To: References: Message-ID: Then I suggest looking at nova api and nova compute logs correlated with these ERROR timestamps. -yoctozepto On Mon, 6 Dec 2021 at 21:11, Rodrigo Lima wrote: > Hi! > > It?s a lot of information. From engine: > 2021-12-06 20:02:33.502 7 INFO masakari.engine.manager > [req-034e0a9e-1b34-4db3-a2e0-6d61a369ba6c 6c128366b66346d38fb5493adf0cf666 > e39a7ea7d17046b5b97b2253bf8195bc - - -] Processing notification > 2468563d-70c5-41ef-ad04-a51b4ad3dd4d of type: COMPUTE_HOST > 2021-12-06 20:02:34.052 7 INFO masakari.compute.nova > [req-882d4151-0549-4cb2-acbb-90509073179c nova - - - -] Disable > nova-compute on ctl02-hml.amt.net.br > 2021-12-06 20:02:34.113 7 INFO > masakari.engine.drivers.taskflow.host_failure > [req-882d4151-0549-4cb2-acbb-90509073179c nova - - - -] Sleeping 180 sec > before starting recovery thread until nova recognizes the node down. > 2021-12-06 20:05:34.130 7 INFO masakari.compute.nova > [req-c570bdef-0786-47bd-94fc-e6c1396eabb1 nova - - - -] Fetch Server list > on ctl02-hml.amt.net.br > 2021-12-06 20:05:35.166 7 INFO masakari.compute.nova > [req-adec6e54-15ff-4a28-98c3-01070ba2cf87 nova - - - -] Call get server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:35.949 7 INFO masakari.compute.nova > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Call get server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:35.955 7 INFO masakari.compute.nova > [req-e548525d-74f7-42c4-8045-2f63d645f76f nova - - - -] Call get server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:36.739 7 INFO masakari.compute.nova > [req-9cea2ff2-0d32-42a1-8605-b889d95cadc0 nova - - - -] Call lock server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:36.747 7 INFO masakari.compute.nova > [req-bc3bdea3-e923-4c08-bf84-bea711465dce nova - - - -] Call get server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:37.391 7 INFO masakari.compute.nova > [req-9cea2ff2-0d32-42a1-8605-b889d95cadc0 nova - - - -] Call evacuate > command for instance 618e44e8-248f-4f50-a760-581972352af8 on host None > 2021-12-06 20:05:37.478 7 INFO masakari.compute.nova > [req-84d38844-ae87-4767-b822-0099bc40aa16 nova - - - -] Call lock server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:38.059 7 INFO masakari.compute.nova > [req-d57bd685-2e42-48e0-bf7d-9978ac516451 nova - - - -] Call get server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:38.102 7 INFO masakari.compute.nova > [req-84d38844-ae87-4767-b822-0099bc40aa16 nova - - - -] Call evacuate > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 on host None > 2021-12-06 20:05:38.734 7 INFO masakari.compute.nova > [req-25b4bb9c-2e05-409a-bae8-fd7c8348ba5b nova - - - -] Call get server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:39.062 7 INFO masakari.compute.nova > [req-3d69363b-aa6c-4752-be52-285bc36f25c2 nova - - - -] Call get server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:39.732 7 INFO masakari.compute.nova > [req-425752db-0599-42a8-bf2a-0229148410cc nova - - - -] Call get server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall > [req-3d69363b-aa6c-4752-be52-285bc36f25c2 nova - - - -] Fixed interval > looping call > 'masakari.engine.drivers.taskflow.host_failure.EvacuateInstancesTask._evacuate_and_confirm.._wait_for_evacuation_confirmation' > failed: masakari.exception.InstanceEvacuateFailed: Failed to evacuate > instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall Traceback (most > recent call last): > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_service/loopingcall.py", > line 150, in _run_loop > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall result = > func(*self.args, **self.kw) > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall File > "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", > line 207, in _wait_for_evacuation_confirmation > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall raise > exception.InstanceEvacuateFailed( > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall > masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance > 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall > 2021-12-06 20:05:39.790 7 WARNING > masakari.engine.drivers.taskflow.host_failure > [req-d537ed5e-f5c8-4ba5-b559-0156883ca02b nova - - - -] Failed to evacuate > instance 618e44e8-248f-4f50-a760-581972352af8: > masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance > 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:39.793 7 INFO masakari.compute.nova > [req-c02a47f3-fc23-4d5b-9167-cf21bf8923db nova - - - -] Call unlock server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall > [req-425752db-0599-42a8-bf2a-0229148410cc nova - - - -] Fixed interval > looping call > 'masakari.engine.drivers.taskflow.host_failure.EvacuateInstancesTask._evacuate_and_confirm.._wait_for_evacuation_confirmation' > failed: masakari.exception.InstanceEvacuateFailed: Failed to evacuate > instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall Traceback (most > recent call last): > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_service/loopingcall.py", > line 150, in _run_loop > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall result = > func(*self.args, **self.kw) > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall File > "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", > line 207, in _wait_for_evacuation_confirmation > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall raise > exception.InstanceEvacuateFailed( > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall > masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance > 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall > 2021-12-06 20:05:40.423 7 WARNING > masakari.engine.drivers.taskflow.host_failure > [req-f8061ba7-353b-487c-8813-30f85216335f nova - - - -] Failed to evacuate > instance 746178b2-14ce-4ce2-83f6-2bf9d613a887: > masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance > 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:40.425 7 INFO masakari.compute.nova > [req-7f0c16fe-8eaa-485b-a414-945fcd84a3c6 nova - - - -] Call unlock server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:41.080 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task > 'EvacuateInstancesTask' (e49cc088-a9d7-4dd8-a6db-75339c6eaa4d) transitioned > into state 'FAILURE' from state 'RUNNING' > 4 predecessors (most recent first): > Flow 'post_tasks' > |__Flow 'main_tasks' > |__Flow 'pre_tasks' > |__Flow 'instance_evacuate_engine': > masakari.exception.HostRecoveryFailureException: Failed to evacuate > instances > '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' > from host 'ctl02-hml.amt.net.br' > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > Traceback (most recent call last): > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/taskflow/engines/action_engine/executor.py", > line 53, in _execute_task > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > result = task.execute(**arguments) > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", > line 396, in execute > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > _do_evacuate(self.context, host_name, instance_list) > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", > line 376, in _do_evacuate > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > raise exception.HostRecoveryFailureException( > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > masakari.exception.HostRecoveryFailureException: Failed to evacuate > instances > '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' > from host 'ctl02-hml.amt.net.br' > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > 2021-12-06 20:05:41.088 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task > 'EvacuateInstancesTask' (e49cc088-a9d7-4dd8-a6db-75339c6eaa4d) transitioned > into state 'REVERTED' from state 'REVERTING' > 2021-12-06 20:05:41.091 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task > 'PrepareHAEnabledInstancesTask' (21c32e80-7521-44bb-bd28-53f88c3d13da) > transitioned into state 'REVERTED' from state 'REVERTING' > 2021-12-06 20:05:41.093 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task > 'DisableComputeServiceTask' (b3522750-c7a6-4f71-ad8c-e2a61a40e2b8) > transitioned into state 'REVERTED' from state 'REVERTING' > 2021-12-06 20:05:41.095 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Flow > 'instance_evacuate_engine' (3c8c5def-39b8-4956-8b2c-62a218e612ee) > transitioned into state 'REVERTED' from state 'RUNNING' > 2021-12-06 20:05:41.096 7 ERROR masakari.engine.manager > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Failed to process > notification '2468563d-70c5-41ef-ad04-a51b4ad3dd4d'. Reason: Failed to > evacuate instances > '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' > from host 'ctl02-hml.amt.net.br': > masakari.exception.HostRecoveryFailureException: Failed to evacuate > instances > '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' > from host 'ctl02-hml.amt.net.br' > 2021-12-06 20:05:41.099 7 INFO masakari.engine.manager > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Notification > 2468563d-70c5-41ef-ad04-a51b4ad3dd4d exits with status: error. > 2021-12-06 20:07:08.803 7 INFO masakari.engine.manager > [req-d2c18313-253f-4bb2-8abb-5ca5d8ac248a nova - - - -] Processing > notification 2468563d-70c5-41ef-ad04-a51b4ad3dd4d of type: COMPUTE_HOST > 2021-12-06 20:07:09.398 7 INFO masakari.compute.nova > [req-5b595769-209f-422b-a71f-ef4507187a61 nova - - - -] Disable > nova-compute on ctl02-hml.amt.net.br > 2021-12-06 20:07:09.467 7 INFO > masakari.engine.drivers.taskflow.host_failure > [req-5b595769-209f-422b-a71f-ef4507187a61 nova - - - -] Sleeping 180 sec > before starting recovery thread until nova recognizes the node > down.2021-12-06 20:02:33.502 7 INFO masakari.engine.manager > [req-034e0a9e-1b34-4db3-a2e0-6d61a369ba6c 6c128366b66346d38fb5493adf0cf666 > e39a7ea7d17046b5b97b2253bf8195bc - - -] Processing notification > 2468563d-70c5-41ef-ad04-a51b4ad3dd4d of type: COMPUTE_HOST > 2021-12-06 20:02:34.052 7 INFO masakari.compute.nova > [req-882d4151-0549-4cb2-acbb-90509073179c nova - - - -] Disable > nova-compute on ctl02-hml.amt.net.br > 2021-12-06 20:02:34.113 7 INFO > masakari.engine.drivers.taskflow.host_failure > [req-882d4151-0549-4cb2-acbb-90509073179c nova - - - -] Sleeping 180 sec > before starting recovery thread until nova recognizes the node down. > 2021-12-06 20:05:34.130 7 INFO masakari.compute.nova > [req-c570bdef-0786-47bd-94fc-e6c1396eabb1 nova - - - -] Fetch Server list > on ctl02-hml.amt.net.br > 2021-12-06 20:05:35.166 7 INFO masakari.compute.nova > [req-adec6e54-15ff-4a28-98c3-01070ba2cf87 nova - - - -] Call get server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:35.949 7 INFO masakari.compute.nova > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Call get server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:35.955 7 INFO masakari.compute.nova > [req-e548525d-74f7-42c4-8045-2f63d645f76f nova - - - -] Call get server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:36.739 7 INFO masakari.compute.nova > [req-9cea2ff2-0d32-42a1-8605-b889d95cadc0 nova - - - -] Call lock server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:36.747 7 INFO masakari.compute.nova > [req-bc3bdea3-e923-4c08-bf84-bea711465dce nova - - - -] Call get server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:37.391 7 INFO masakari.compute.nova > [req-9cea2ff2-0d32-42a1-8605-b889d95cadc0 nova - - - -] Call evacuate > command for instance 618e44e8-248f-4f50-a760-581972352af8 on host None > 2021-12-06 20:05:37.478 7 INFO masakari.compute.nova > [req-84d38844-ae87-4767-b822-0099bc40aa16 nova - - - -] Call lock server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:38.059 7 INFO masakari.compute.nova > [req-d57bd685-2e42-48e0-bf7d-9978ac516451 nova - - - -] Call get server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:38.102 7 INFO masakari.compute.nova > [req-84d38844-ae87-4767-b822-0099bc40aa16 nova - - - -] Call evacuate > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 on host None > 2021-12-06 20:05:38.734 7 INFO masakari.compute.nova > [req-25b4bb9c-2e05-409a-bae8-fd7c8348ba5b nova - - - -] Call get server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:39.062 7 INFO masakari.compute.nova > [req-3d69363b-aa6c-4752-be52-285bc36f25c2 nova - - - -] Call get server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:39.732 7 INFO masakari.compute.nova > [req-425752db-0599-42a8-bf2a-0229148410cc nova - - - -] Call get server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall > [req-3d69363b-aa6c-4752-be52-285bc36f25c2 nova - - - -] Fixed interval > looping call > 'masakari.engine.drivers.taskflow.host_failure.EvacuateInstancesTask._evacuate_and_confirm.._wait_for_evacuation_confirmation' > failed: masakari.exception.InstanceEvacuateFailed: Failed to evacuate > instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall Traceback (most > recent call last): > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_service/loopingcall.py", > line 150, in _run_loop > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall result = > func(*self.args, **self.kw) > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall File > "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", > line 207, in _wait_for_evacuation_confirmation > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall raise > exception.InstanceEvacuateFailed( > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall > masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance > 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:39.778 7 ERROR oslo.service.loopingcall > 2021-12-06 20:05:39.790 7 WARNING > masakari.engine.drivers.taskflow.host_failure > [req-d537ed5e-f5c8-4ba5-b559-0156883ca02b nova - - - -] Failed to evacuate > instance 618e44e8-248f-4f50-a760-581972352af8: > masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance > 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:39.793 7 INFO masakari.compute.nova > [req-c02a47f3-fc23-4d5b-9167-cf21bf8923db nova - - - -] Call unlock server > command for instance 618e44e8-248f-4f50-a760-581972352af8 > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall > [req-425752db-0599-42a8-bf2a-0229148410cc nova - - - -] Fixed interval > looping call > 'masakari.engine.drivers.taskflow.host_failure.EvacuateInstancesTask._evacuate_and_confirm.._wait_for_evacuation_confirmation' > failed: masakari.exception.InstanceEvacuateFailed: Failed to evacuate > instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall Traceback (most > recent call last): > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall File > "/var/lib/kolla/venv/lib/python3.8/site-packages/oslo_service/loopingcall.py", > line 150, in _run_loop > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall result = > func(*self.args, **self.kw) > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall File > "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", > line 207, in _wait_for_evacuation_confirmation > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall raise > exception.InstanceEvacuateFailed( > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall > masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance > 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:40.422 7 ERROR oslo.service.loopingcall > 2021-12-06 20:05:40.423 7 WARNING > masakari.engine.drivers.taskflow.host_failure > [req-f8061ba7-353b-487c-8813-30f85216335f nova - - - -] Failed to evacuate > instance 746178b2-14ce-4ce2-83f6-2bf9d613a887: > masakari.exception.InstanceEvacuateFailed: Failed to evacuate instance > 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:40.425 7 INFO masakari.compute.nova > [req-7f0c16fe-8eaa-485b-a414-945fcd84a3c6 nova - - - -] Call unlock server > command for instance 746178b2-14ce-4ce2-83f6-2bf9d613a887 > 2021-12-06 20:05:41.080 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task > 'EvacuateInstancesTask' (e49cc088-a9d7-4dd8-a6db-75339c6eaa4d) transitioned > into state 'FAILURE' from state 'RUNNING' > 4 predecessors (most recent first): > Flow 'post_tasks' > |__Flow 'main_tasks' > |__Flow 'pre_tasks' > |__Flow 'instance_evacuate_engine': > masakari.exception.HostRecoveryFailureException: Failed to evacuate > instances > '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' > from host 'ctl02-hml.amt.net.br' > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > Traceback (most recent call last): > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/taskflow/engines/action_engine/executor.py", > line 53, in _execute_task > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > result = task.execute(**arguments) > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", > line 396, in execute > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > _do_evacuate(self.context, host_name, instance_list) > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > File > "/var/lib/kolla/venv/lib/python3.8/site-packages/masakari/engine/drivers/taskflow/host_failure.py", > line 376, in _do_evacuate > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > raise exception.HostRecoveryFailureException( > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > masakari.exception.HostRecoveryFailureException: Failed to evacuate > instances > '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' > from host 'ctl02-hml.amt.net.br' > 2021-12-06 20:05:41.080 7 ERROR masakari.engine.drivers.taskflow.driver > 2021-12-06 20:05:41.088 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task > 'EvacuateInstancesTask' (e49cc088-a9d7-4dd8-a6db-75339c6eaa4d) transitioned > into state 'REVERTED' from state 'REVERTING' > 2021-12-06 20:05:41.091 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task > 'PrepareHAEnabledInstancesTask' (21c32e80-7521-44bb-bd28-53f88c3d13da) > transitioned into state 'REVERTED' from state 'REVERTING' > 2021-12-06 20:05:41.093 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Task > 'DisableComputeServiceTask' (b3522750-c7a6-4f71-ad8c-e2a61a40e2b8) > transitioned into state 'REVERTED' from state 'REVERTING' > 2021-12-06 20:05:41.095 7 WARNING masakari.engine.drivers.taskflow.driver > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Flow > 'instance_evacuate_engine' (3c8c5def-39b8-4956-8b2c-62a218e612ee) > transitioned into state 'REVERTED' from state 'RUNNING' > 2021-12-06 20:05:41.096 7 ERROR masakari.engine.manager > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Failed to process > notification '2468563d-70c5-41ef-ad04-a51b4ad3dd4d'. Reason: Failed to > evacuate instances > '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' > from host 'ctl02-hml.amt.net.br': > masakari.exception.HostRecoveryFailureException: Failed to evacuate > instances > '618e44e8-248f-4f50-a760-581972352af8,746178b2-14ce-4ce2-83f6-2bf9d613a887' > from host 'ctl02-hml.amt.net.br' > 2021-12-06 20:05:41.099 7 INFO masakari.engine.manager > [req-49414eed-fdcf-4fd0-b803-e5cc5f2f9227 nova - - - -] Notification > 2468563d-70c5-41ef-ad04-a51b4ad3dd4d exits with status: error. > 2021-12-06 20:07:08.803 7 INFO masakari.engine.manager > [req-d2c18313-253f-4bb2-8abb-5ca5d8ac248a nova - - - -] Processing > notification 2468563d-70c5-41ef-ad04-a51b4ad3dd4d of type: COMPUTE_HOST > 2021-12-06 20:07:09.398 7 INFO masakari.compute.nova > [req-5b595769-209f-422b-a71f-ef4507187a61 nova - - - -] Disable > nova-compute on ctl02-hml.amt.net.br > 2021-12-06 20:07:09.467 7 INFO > masakari.engine.drivers.taskflow.host_failure > [req-5b595769-209f-422b-a71f-ef4507187a61 nova - - - -] Sleeping 180 sec > before starting recovery thread until nova recognizes the node down. > > I tested forcing a kernel panic on target node > > > > > Em s?b., 4 de dez. de 2021 ?s 08:48, Rados?aw Piliszek < > radoslaw.piliszek at gmail.com> escreveu: > >> What do the Masakari API and Masakari Engine logs show? >> >> -yoctozepto >> >> On Fri, 3 Dec 2021 at 19:57, Rodrigo Lima >> wrote: >> >>> I'm working on upgrading an openstack farm from victoria to wallaby. >>> After successful upgrade, I would like to enable hacluster and masakari to >>> test HA between failing compute nodes. Everything seems to be running >>> (pacemaker with the remote nodes OK, corosync without errors, >>> masakari-monitors detects the 2 compute nodes online), but... when I >>> simulate the failure of a node with shutdown, the failure and notification >>> appears in the hostmonitor log, but the instances that were on the failed >>> node don't evacuate, and I couldn't find documentation that explains how to >>> do this specific configuration, if necessary. Does anyone have any ideas? >>> >>> -------------- next part -------------- An HTML attachment was scrubbed... URL: From DHilsbos at performair.com Mon Dec 6 22:53:18 2021 From: DHilsbos at performair.com (DHilsbos at performair.com) Date: Mon, 6 Dec 2021 22:53:18 +0000 Subject: [ops]RabbitMQ High Availability In-Reply-To: References: <0670B960225633449A24709C291A525251D511CD@COM03.performair.local> <4e37d1a7-ca17-b50f-ba6c-96229d85a75e@redhat.com> Message-ID: <0670B960225633449A24709C291A525251DB698C@COM03.performair.local> Arnaud; Thank you for putting this together, and thank you for sending it over to me. Reading it now. Dominic L. Hilsbos, MBA Vice President - Information Technology Perform Air International Inc. DHilsbos at PerformAir.com www.PerformAir.com -----Original Message----- From: Arnaud Morin [mailto:arnaud.morin at gmail.com] Sent: Monday, November 29, 2021 12:04 PM To: Bogdan Dobrelya Cc: Dominic Hilsbos; openstack-discuss at lists.openstack.org Subject: Re: [ops]RabbitMQ High Availability Hi, After a talk on this ml (which starts at [1]), we endup building a documentation with Large Scale group. The doc is accessible at [2]. Hope this will help. [1] http://lists.openstack.org/pipermail/openstack-discuss/2020-August/016362.html [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit On 24.11.21 - 11:31, Bogdan Dobrelya wrote: > On 11/24/21 12:34 AM, DHilsbos at performair.com wrote: > > All; > > > > In the time I've been part of this mailing list, the subject of RabbitMQ high availability has come up several times, and each time specific recommendations for both Rabbit and Open Stack are provided. I remember it being an A or B kind of recommendation (i.e. configure Rabbit like A1, and Open Stack like A2, OR configure Rabbit like B1, and Open Stack like B2). > > There is no special recommendations for rabbitmq setup for openstack, > but probably a few, like instead of putting it behind a haproxy, or the > like, list the rabbit cluster nodes in the oslo messaging config > settings directly. Also, it seems that durable queues makes a very > little sense for highly ephemeral RPC calls, just by design. I would > also add that the raft quorum queues feature of rabbitmq >=3.18 does > neither fit well into the oslo messaging design for RPC calls. > > A discussable and highly opinionated thing is also configuring > ha/mirror queue policy params for queues used for RPC calls vs > broad-casted notifications. > > And my biased personal humble recommendation is: use the upstream OCF RA > [0][1], if configuring rabbitmq cluster by pacemaker. > > [0] https://www.rabbitmq.com/pacemaker.html#auto-pacemaker > > [1] > https://github.com/ClusterLabs/resource-agents/blob/master/heartbeat/rabbitmq-server-ha > > > > > Unfortunately, I can't find the previous threads on this topic. > > > > Does anyone have this information, that they would care to share with me? > > > > Thank you, > > > > Dominic L. Hilsbos, MBA > > Vice President - Information Technology > > Perform Air International Inc. > > DHilsbos at PerformAir.com > > www.PerformAir.com > > > > > > > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando > > From gmann at ghanshyammann.com Mon Dec 6 23:43:09 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Mon, 06 Dec 2021 17:43:09 -0600 Subject: [all][tc] Technical Committee next weekly meeting on Dec 9th at 1500 UTC Message-ID: <17d9221334f.11a9bcb48483928.5563521669059290320@ghanshyammann.com> Hello Everyone, Technical Committee's next weekly meeting is scheduled for Dec 9th at 1500 UTC. If you would like to add topics for discussion, please add them to the below wiki page by Wednesday, Dec 8th, at 2100 UTC. https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting -gmann From gmann at ghanshyammann.com Mon Dec 6 23:53:21 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Mon, 06 Dec 2021 17:53:21 -0600 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <20211022224028.blm3djon2rrgbguk@yuggoth.org> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> Message-ID: <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> ---- On Fri, 22 Oct 2021 17:40:29 -0500 Jeremy Stanley wrote ---- > On 2021-10-23 00:06:02 +0200 (+0200), Thomas Goirand wrote: > [...] > > This is essentially bullshit: the replacement method proposed in this > > page is supposed to be "pip install", which does dependency resolving > > and download from the internet, which is not useful (and even forbidden > > at package build time) for downstream distributions. > > > > If using pip is the only thing that upstream Python people are proposing > > to distributions, without any distro-specific options (like the > > --install-layout=deb option of setuptools, for example), then something > > really wrong is going on. > > > > Unless I missed something and upstream Python likes us from now on?!? > [...] > > Their new answer to "how should we be building binary packages from > Python source distributions" is this (already packaged in Debian): > > https://packages.debian.org/python3-build > > The idea is that a source distribution may have a variety of > different possible build backends per PEP 517, but as long as you > run `python3 -m build` (and the relevant build dependencies are made > available somehow) then it shouldn't matter. Calling directly into > setup.py is only going to work for projects which use Setuptools as > their build backend and provide a setup.py, of course, probably not > for any other build backend. > > > IMO skyline needs to conform with everything that's made elsewhere > > in OpenStack, and adopt the same standards. > [...] > > I agree, and I attempted to leave some detailed notes on > https://review.opendev.org/814037 highlighting the differences I > spotted, which I think they're going to want to address for > consistency with the rest of OpenStack. Hello Skyline team, We have not heard from you on the open questions (mentioned in https://review.opendev.org/c/openstack/governance/+/814037 too). It's been ~ 21 days TC has voted -1 in Gerrit with required things to fix/or tell plan so that TC can merge the new project application. Please respond on the Gerrit to proceed next. -gmann > -- > Jeremy Stanley > From ricolin at ricolky.com Tue Dec 7 08:39:26 2021 From: ricolin at ricolky.com (Rico Lin) Date: Tue, 7 Dec 2021 16:39:26 +0800 Subject: [tc][all] Follow up pain point discussion In-Reply-To: References: Message-ID: Dear all For pain point discussion, Let's schedule the video meeting for *Dec. 8th, 15:00 UTC time* Location: https://meet.google.com/xsx-inbw-mos *Rico Lin* OIF Individual Board of directors, OpenStack TC, Multi-arch SIG chair, Heat PTL, Senior Software Engineer at EasyStack On Fri, Dec 3, 2021 at 5:37 PM Rico Lin wrote: > Dear all > > The pain point collection etherpad [1] seems to get good attention from > teams. > We plan to host a one-hour video meeting to keep the discussion going. > Please fill in the doodle to select your preferred time if you can join us: > https://doodle.com/poll/kur2mfywceerqcvh?utm_source=poll&utm_medium=link > The discussion will be hosted in video format. > > Here are some items (but not limited to) I expect we will discuss and > hopefully put some actions/efforts on. > > - General Pain point: RabbitMQ > - General Pain point: OpenStackClient support > - General project pain point discussion: How to encourage teams to > work/reply on pain points. > > Meanwhile, we encourage all to check the etherpad and provide your > feedback. > Also if teams can help to answer questions/concerns on etherpad (like Nova > team does) will be greatly appreciated. > > [1] https://etherpad.opendev.org/p/pain-point-elimination > > *Rico Lin* > -------------- next part -------------- An HTML attachment was scrubbed... URL: From skaplons at redhat.com Tue Dec 7 10:18:41 2021 From: skaplons at redhat.com (Slawek Kaplonski) Date: Tue, 07 Dec 2021 11:18:41 +0100 Subject: [neutron][CI] How to reduce number of rechecks - brainstorming In-Reply-To: <2165480.iZASKD2KPV@p1> References: <2165480.iZASKD2KPV@p1> Message-ID: <4362100.LvFx2qVVIh@p1> Hi, Thank You all for all the ideas and discussion in this thread. With Yatin we prepared summary at https://etherpad.opendev.org/p/neutron-ci-improvements and we want to go over and discuss them on the today's CI meeting. It will be both video and irc meeting. We will talk on https:// meetpad.opendev.org/neutron-ci-meetings Agenda for the meeting is at https://etherpad.opendev.org/p/neutron-ci-meetings Everyone is welcome to join the discussion :) On ?roda, 17 listopada 2021 09:13:34 CET Slawek Kaplonski wrote: > Hi, > > Recently I spent some time to check how many rechecks we need in Neutron to > get patch merged and I compared it to some other OpenStack projects (see [1] > for details). > TL;DR - results aren't good for us and I think we really need to do something > with that. > > Of course "easiest" thing to say is that we should fix issues which we are > hitting in the CI to make jobs more stable. But it's not that easy. We are > struggling with those jobs for very long time. We have CI related meeting > every week and we are fixing what we can there. > Unfortunately there is still bunch of issues which we can't fix so far > because they are intermittent and hard to reproduce locally or in some cases > the issues aren't realy related to the Neutron or there are new bugs which > we need to investigate and fix :) > So this is never ending battle for us. The problem is that we have to test > various backends, drivers, etc. so as a result we have many jobs running on > each patch - excluding UT, pep8 and docs jobs we have around 19 jobs in check > and 14 jobs in gate queue. > > In the past we made a lot of improvements, like e.g. we improved irrelevant > files lists for jobs to run less jobs on some of the patches, together with > QA team we did "integrated-networking" template to run only Neutron and Nova > related scenario tests in the Neutron queues, we removed and consolidated > some of the jobs (there is still one patch in progress for that but it > should just remove around 2 jobs from the check queue). All of that are good > improvements but still not enough to make our CI really stable :/ > > Because of all of that, I would like to ask community about any other ideas > how we can improve that. If You have any ideas, please send it in this email > thread or reach out to me directly on irc. > We want to discuss about them in the next video CI meeting which will be on > November 30th. If You would have any idea and would like to join that > discussion, You are more than welcome in that meeting of course :) > > [1] http://lists.openstack.org/pipermail/openstack-discuss/2021-November/ > 025759.html -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From kamil.madac at slovenskoit.sk Tue Dec 7 10:20:27 2021 From: kamil.madac at slovenskoit.sk (=?iso-8859-2?Q?Kamil_Mad=E1=E8?=) Date: Tue, 7 Dec 2021 10:20:27 +0000 Subject: [nova] How to update OS-EXT-SRV-ATTR:hostname Message-ID: Hello, I would like to ask, how to update OS-EXT-SRV-ATTR:hostname via openstack CLI? We changed hostnames on bunch of instances, but metadata service still serves old hostname which causes that every future rebuild of instance reverts hostname (`hostname` command in instance console) back to old one. Kamil Mad?? Slovensko IT a.s. -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Tue Dec 7 10:53:17 2021 From: zigo at debian.org (Thomas Goirand) Date: Tue, 7 Dec 2021 11:53:17 +0100 Subject: [nova] How to update OS-EXT-SRV-ATTR:hostname In-Reply-To: References: Message-ID: On 12/7/21 11:20 AM, Kamil Mad?? wrote: > Hello, > > I would like to ask, how to update OS-EXT-SRV-ATTR:hostname via > openstack CLI? We changed hostnames on bunch of instances, but metadata > service still serves old hostname which causes that every future rebuild > of instance reverts hostname (`hostname` command in instance console) > back to old one. > > Kamil Mad?? > /Slovensko IT a.s./ Looks like what you want to do is disable cloud-init editing of your VM hostname. This is done simply by cloud-init to stop doing this. Simply edit /etc/cloud/cloud.cfg and set: preserve_hostname: true or remove any cloud-init module from the "cloud_init_modules:" list. Simply remove the lines containing: - set_hostname - update_hostname - update_etc_hosts I hope this helps, Cheers, Thomas Goirand (zigo) From stephenfin at redhat.com Tue Dec 7 11:30:42 2021 From: stephenfin at redhat.com (Stephen Finucane) Date: Tue, 07 Dec 2021 11:30:42 +0000 Subject: [nova] How to update OS-EXT-SRV-ATTR:hostname In-Reply-To: References: Message-ID: On Tue, 2021-12-07 at 11:53 +0100, Thomas Goirand wrote: > On 12/7/21 11:20 AM, Kamil Mad?? wrote: > > Hello, > > > > I would like to ask, how to update OS-EXT-SRV-ATTR:hostname via > > openstack CLI? We changed hostnames on bunch of instances, but metadata > > service still serves old hostname which causes that every future rebuild > > of instance reverts hostname (`hostname` command in instance console) > > back to old one. > > > > Kamil Mad?? > > /Slovensko IT a.s./ > > Looks like what you want to do is disable cloud-init editing of your VM > hostname. This is done simply by cloud-init to stop doing this. Simply > edit /etc/cloud/cloud.cfg and set: > > preserve_hostname: true > > or remove any cloud-init module from the "cloud_init_modules:" list. > Simply remove the lines containing: > - set_hostname > - update_hostname > - update_etc_hosts > > I hope this helps, > > Cheers, > > Thomas Goirand (zigo) To add to this, there's no way to change the hostname embedded in the instance metadata until Xena, when we added the ability to provide a 'hostname' parameter when creating, updating or rebuilding servers (microversion 2.90). See [1] and [2] for more details. Stephen [1]https://docs.openstack.org/nova/latest/reference/api-microversion-history.html#maximum-in-xena [2]https://specs.openstack.org/openstack/nova-specs/specs/xena/implemented/configurable-instance-hostnames.html From lokendrarathour at gmail.com Tue Dec 7 12:26:44 2021 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Tue, 7 Dec 2021 17:56:44 +0530 Subject: [Triple0] [Undercloud] Overcloud deployment getting failed at Introspection Message-ID: Hi Team, we are trying to deploy the triple0 Stein release on Centos 7. Undercloud is getting deployed successfully but the introspection of the overcloud node is getting failed with the reason as below: *on the DNS podman logs:* ()[root at undercloud dhcp-hostsdir]# tail -f /var/log/ironic-inspector/dnsmasq.log Dec 7 11:21:08 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) f8:f2:1e:9e:82:51 ignored Dec 7 11:21:08 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) f8:f2:1e:9e:82:50 ignored Dec 7 11:21:12 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) f8:f2:1e:9e:82:51 ignored *Checking the TCPDUMP at the same time gives:* 11:38:38.308660 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, Request from f8:f2:1e:9e:82:51 (oui Unknown), length 300 11:38:38.435173 IP6 fe80::f53a:6448:4a29:e051 > ff02::16: HBH ICMP6, multicast listener report v2, 1 group record(s), length 28 11:38:38.435184 IP6 fe80::f53a:6448:4a29:e051 > ff02::16: HBH ICMP6, multicast listener report v2, 1 group record(s), length 28 11:38:41.155784 IP6 fe80::322d:3253:e3b2:316e > ff02::2: ICMP6, router solicitation, length 8 11:38:41.394304 IP6 fe80::f53a:6448:4a29:e051 > ff02::2: ICMP6, router solicitation, length 8 11:38:42.563182 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, Request from f8:f2:1e:9e:82:50 (oui Unknown), length 300 11:38:42.611973 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, Request from f8:f2:1e:9e:82:51 (oui Unknown), length 300 *We observed that DHCP Server is not offering the IP to the baremetal. * *DHCP discovery comes at br-ctlplane but no ip is offered.* Any input would be really helpful. Best Regards, Lokendra -------------- next part -------------- An HTML attachment was scrubbed... URL: From cyril at redhat.com Tue Dec 7 15:44:30 2021 From: cyril at redhat.com (Cyril Roelandt) Date: Tue, 7 Dec 2021 16:44:30 +0100 Subject: Glance review party - 2021-12-10T14:00:00Z Message-ID: Hello, On Friday (2021-12-10), at 14:00 UTC, we will have a "review party" for Glance. The party should last 2 to 3 hours, and the goal will be to clean the backlogs of glance_store and glanceclient, and try to work on the glance backlog as well if there is enough time. If some of your patches are currently in "merge conflict", please rebase them and make sure the tests pass, which will allow us to review them. If you think some of your *Glance* patches are a top priority for the project, please answer this email with a link to them, and we will try to make time to review them on Friday. We will be using https://bluejeans.com/u/croeland to coordinate. Happy Hacking! Cyril Roelandt From openstack at a.spamming.party Tue Dec 7 16:36:31 2021 From: openstack at a.spamming.party (Jean-Philippe Evrard) Date: Tue, 07 Dec 2021 17:36:31 +0100 Subject: [all][tc] Relmgt team position on release cadence In-Reply-To: References: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> <8301ef05-4011-43e9-beec-66f6aeeb00f9@www.fastmail.com> <20211129130957.3poysdeqsufkp3pg@yuggoth.org> <17d72ae5ee2.b80bfd38119150.5945079300853303270@ghanshyammann.com> <20211130220913.ydk5zhyvkdl7g7zc@yuggoth.org> <6536dfae-cb75-4263-856d-f4dbdfa2db01@www.fastmail.com> Message-ID: Hello, On Sat, Dec 4, 2021, at 15:56, Thierry Carrez wrote: > Jean-Philippe Evrard wrote: >> [...] >> Here is what I read: >> 1) Many want more releases, not less. I haven't seen a complaint about tagging more releases. >> 2) More than one person is proposing to abandon the integrated release, and nobody has complained about it. > > Here are a few drawbacks of abandoning the integrated release (which is > really a "synchronized release"): > > - You would no longer have a baseline of components that are heavily > tested together which we do community QA on and (collectively) clearly > sign off on, so downstream integrators are on their own and have much > more work to do I don't think so. We can still have that without the synchronized release. We decide what we test together. (Maybe the TC can define that?). For example, with Zuul, we can test the master branch of all projects together to ensure the latest branch always work according to the set criteria. (No change there!) Then, what becomes "openstack version x" is just about a manifest of the SHAs or the tags of the projects, tested together. (No change again). (Let's call this proposition 1) We don't have multiple "synchronized" releases of stable branches, so we don't even have to bring any "openstack version x.y". Still, it doesn't prevent us to define any "openstack version x.y" if we want to, with an updated manifest compared to "version x.0" Technically proposition 1 looks like we run a deploy tool + tempest before release, based on the manifest. It's literally no new technology. I am relatively sure many (if not all) deployment tools can do that. What makes more sense for me is that we define "openstack version x" in terms of APIs, and that we have the test tooling to ensure that software wanted to be integrated into "version x" are passing said tests. (Let's call this proposition 2) It allows alternative implementation of the software, if someone is crazy enough to do that. I agree it might look like "We are abandonning the branches, what will we do for people deploying regularily on a single branch?". Well it doesn't change much: Those are currently consuming the manifests of releases, bumping the SHAs from a "stable" branch of a repo. If a project/repo _decides to branch_ for a very critical reason, then it can still happen. If there is no reason to branch, then you would still bump the latest branch available. Still no issue there. > - You would no longer have clearly-comparable points between > distributions. Everyone knows what "running Ussuri" means, which > facilitates communication and bug report handling in the community. I am not sure what "running Ussuri" means. Is that when the branch was cut, or the latest sha of all the projects' Ussuri branches? Having "OpenStack version Ussuri 1" corresponding to a manifest of project SHAs and or APIs versions is far more clear to me. > - You would no longer have clear community support commitments. We > currently maintain and fix bug reports from people running vanilla > "Ussuri"... but do we want to care about every combination of components > under the sun? (maybe we do already) We don't stop maintainers for contributing by being clearer in what we release and how we branch... If people want to maintain some old version and _need to branch a project_, I feel it's still possible. But it's now in the power of the project to decide whether it makes sense to do so, instead of being forced to manage something that might be stretching the teams thin. > - You would no longer have "OpenStack" released, so you miss the regular > marketing opportunity to remind the rest of the world that it still > exists. The OpenStack brand fades, and it gets more complicated to get > development resources to work on it. Again, it's wording. Please see my proposal above. I understand why it's a concern for the foundation however ;) > Without the synchronized release, OpenStack essentially becomes a > rolling distribution of cloud components on which we make very limited > guarantees. I guess it is suitable to build maintained distributions on, > but it really is no longer directly usable beyond development. Is that > what we want? We would indeed be more "rolling", but it doesn't prevent tagging/point in time testing and quality assurance. Also, if we have integrated testing during the cycle and before the tagging, then it doesn't remove any guarantee, does it? I disagree it's not usable beyond development. Here is my reasoning: 1) I don't see it as removing any guarantee if we test things out :) 2) Users of openstack are most likely using a deployment tooling (based on configuration management tool of choice) to deploy their clouds, not a manual deployment. This seems to be confirmed by the user survey. Do you mean that a change in the model of branching would irreversibly break all those tools, and make the ecosystem "no longer directly usable beyond development"? Keep in mind I see those deploy toolings as part of openstack. Not only because they are part of the ecosystem, but because some are under OpenStack governance (OpenStack charms, openstack-chef, puppet openstack, osa, openstack-helm, kolla, tripleO). Hence I consider that OpenStack would still be usable beyond development. I know at least a company that would continue to believe in OpenStack ;) >> 3) Many people seem eager to carry "stable branches" for "critical patches", but no new definition of such criticality was done. > > Note that a common stable branch cut is *exactly* the same thing as a > synchronized release... So I see 2 and 3 as being opposed views. I agree with you there on common stable branch = synchronized release, especially if "criticality" = "Whenever we decide to cut a release". I still wanted to mention what was said, for the reader. It doesn't mean that everyone agree on the definition of criticality yet, or maybe I am wrong? ;) Next to this, I have questions: A) Am I the only one wanting to act on this? B) Am I the only one caring? C) What should be the next steps if I want to change this model? D) Should I propose a change in governance, sync with release management team? As this goes deep into the foundation's model, I feel like we need a proper coordination to make this happen. E) Do you consider all the energy spent to change things does not bring enough positive reward for the whole ecosystem? I see people talking about changing releases for years now, I haven't seen a single change in our behaviour. (or maybe I missed something?). Is that stockholm syndrome? ;) JP From katonalala at gmail.com Tue Dec 7 10:32:08 2021 From: katonalala at gmail.com (Lajos Katona) Date: Tue, 7 Dec 2021 11:32:08 +0100 Subject: [neutron] Instances can't get IP from the DHCP server in OpenStack In-Reply-To: References: Message-ID: Hi, I have more questions than ready answers. If I understand well you have a single host all-in-one setup, am I right? You mentioned flat network, for that please check this document to be sure that you have everything set correctly: https://docs.openstack.org/neutron/latest/admin/deploy-ovs-provider.html Regards Lajos Katona (lajoskatona) ??? ezt ?rta (id?pont: 2021. dec. 6., H, 17:19): > Hi, > > I'm trying to create instance in OpenStack Victoria. But, I am facing a > issues about dhcp of neutron. > > I have OpenStack Victoria running in one baremetl server with uniontech > os(a downstream of centos 8). I have a Flat network created in the range of > 10.12.21.190-10.12.21.195. I selected to have a DHCP. The instance ran, > and neutron-dhcp-agent.service had allocated a IP to it. > # openstack server list > +--------------------------------------+-------+--------+-----------------------+----------------+--------+ > | ID | Name | Status | Networks | Image | Flavor | > +--------------------------------------+-------+--------+-----------------------+----------------+--------+ > | 4ccae37e-fbfe-4acb-a109-e1bc9175c2e0 | inst1 | ACTIVE | > provider=10.12.21.192 | centos8-server | h2 | > +--------------------------------------+-------+--------+-----------------------+----------------+--------+ > But the instance can't get a response from DHCP. No any error in log. And > if I setup the ip manually in the instance I can get access to the gateway, > dhcp port and external network. > > I can catpure the DHCP request in any network device in the server, > include dhcp port(10.12.21.190). But no response. Dnsmasq process was > running and had > # openstack port list > +--------------------------------------+------+-------------------+-----------------------------------------------------------------------------+--------+ > | ID | Name | MAC Address | Fixed IP Addresses | Status | > +--------------------------------------+------+-------------------+-----------------------------------------------------------------------------+--------+ > | 5ab32ca8-2d1d-47cf-9a62-7c01d21abaf0 | | fa:16:3e:1d:48:45 | ip_address= > '10.12.21.192', subnet_id='3daf5a55-e76a-4093-8533-78d9464b1beb' | ACTIVE > | | 63fb2306-7759-4933-8d93-590e3a56f315 | | fa:16:3e:c5:59:9a | ip_address= > '10.12.21.190', subnet_id='3daf5a55-e76a-4093-8533-78d9464b1beb' | ACTIVE > | > +--------------------------------------+------+-------------------+-----------------------------------------------------------------------------+--------+ > # ip netns qdhcp-ec2c4d9d-888b-4312-b0af-ab2127b76e0e (id: 0) # sudo ip > netns exec qdhcp-ec2c4d9d-888b-4312-b0af-ab2127b76e0e sudo tcpdump -n -S -i > tap63fb2306-77|grep DHCP > ## notes: *fa:16:3e:1d:48:45 is the mac of nic of inst1* ? > dropped privs to tcpdump tcpdump: verbose output suppressed, use -v or -vv > for full protocol decode listening on tap63fb2306-77, link-type EN10MB > (Ethernet), capture size 262144 bytes 16:47:18.194028 IP 0.0.0.0.bootpc > > 255.255.255.255.bootps: BOOTP/DHCP, Request from fa:16:3e:1d:48:45, length > 276 16:47:21.146736 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, > Request from fa:16:3e:1d:48:45, length 276 16:47:25.226387 IP > 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, Request from > fa:16:3e:1d:48:45, length 276 > > more detail informations: > # ps -ef|grep dnsmasq > dnsmasq 7441 1 0 12?03 ? 00:00:00 dnsmasq --no-hosts --no-resolv > --pid-file=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/pid > --dhcp-hostsfile=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/host > --addn-hosts=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/addn_hosts > --dhcp-optsfile=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/opts > --dhcp-leasefile=/var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/leases > --dhcp-match=set:ipxe,175 --dhcp-userclass=set:ipxe6,iPXE --local-service > --bind-dynamic > --dhcp-range=set:subnet-3daf5a55-e76a-4093-8533-78d9464b1beb,10.12.21.0,static,255.255.255.0,86400s > --dhcp-option-force=option:mtu,1500 --dhcp-lease-max=256 --conf-file= > --domain=openstacklocal > dnsmasq 317868 1 0 14:33 ? 00:00:00 /usr/sbin/dnsmasq > --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro > --dhcp-script=/usr/libexec/libvirt_leaseshelper > root 317869 317868 0 14:33 ? 00:00:00 /usr/sbin/dnsmasq > --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro > --dhcp-script=/usr/libexec/libvirt_leaseshelper > root 340315 317310 0 18:43 pts/2 00:00:00 grep --color=auto dnsmasq > [root at a16 ~]# cat > /var/lib/neutron/dhcp/ec2c4d9d-888b-4312-b0af-ab2127b76e0e/host > fa:16:3e:c5:59:9a,host-10-12-21-190.openstacklocal,10.12.21.190 > fa:16:3e:1d:48:45,host-10-12-21-192.openstacklocal,10.12.21.192 > # ovs-vsctl show adbd4b0c-cb78-4b8f-b8a6-197c5948312c Manager > "ptcp:6640:127.0.0.1" is_connected: true Bridge br0 Controller "tcp: > 127.0.0.1:6633" is_connected: true fail_mode: secure datapath_type: > system Port enp5s0 Interface enp5s0 Port phy-br0 Interface phy-br0 type: > patch options: {peer=int-br0} Port br0 Interface br0 type: internal > Bridge br-tun Controller "tcp:127.0.0.1:6633" is_connected: true > fail_mode: secure datapath_type: system Port br-tun Interface br-tun type: > internal Port patch-int Interface patch-int type: patch options: > {peer=patch-tun} Bridge br-int Controller "tcp:127.0.0.1:6633" > is_connected: true fail_mode: secure datapath_type: system Port int-br0 > Interface int-br0 type: patch options: {peer=phy-br0} Port br-int > Interface br-int type: internal Port tap63fb2306-77 tag: 1 Interface > tap63fb2306-77 type: internal Port patch-tun Interface patch-tun type: > patch options: {peer=patch-int} Port qvo5ab32ca8-2d tag: 1 Interface > qvo5ab32ca8-2d ovs_version: "2.13.0". # ip a 1: lo: > mtu 65536 qdisc noqueue state UNKNOWN group default > qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet > 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 > ::1/128 scope host valid_lft forever preferred_lft forever 2: enp5s0: > mtu 1500 qdisc fq_codel master ovs-system > state UP group default qlen 1000 link/ether 1c:69:7a:92:be:30 brd > ff:ff:ff:ff:ff:ff 3: ovs-system: mtu 1500 qdisc noop > state DOWN group default qlen 1000 link/ether ea:f5:ce:24:93:12 brd > ff:ff:ff:ff:ff:ff 4: br0: mtu 1500 qdisc > noqueue state UNKNOWN group default qlen 1000 link/ether 1c:69:7a:92:be:30 > brd ff:ff:ff:ff:ff:ff inet 10.12.21.142/24 scope global br0 valid_lft > forever preferred_lft forever inet6 fe80::1e69:7aff:fe92:be30/64 scope link > valid_lft forever preferred_lft forever 5: br-int: > mtu 1500 qdisc noqueue state UNKNOWN > group default qlen 1000 link/ether ba:5d:16:cf:89:40 brd ff:ff:ff:ff:ff:ff > inet6 fe80::b85d:16ff:fecf:8940/64 scope link valid_lft forever > preferred_lft forever 9: br-tun: mtu 1500 qdisc noop > state DOWN group default qlen 1000 link/ether 86:bb:24:69:7e:43 brd > ff:ff:ff:ff:ff:ff 11: virbr0: mtu 1500 > qdisc noqueue state DOWN group default qlen 1000 link/ether > 52:54:00:4d:96:e4 brd ff:ff:ff:ff:ff:ff inet 192.168.122.1/24 brd > 192.168.122.255 scope global virbr0 valid_lft forever preferred_lft forever > 12: virbr0-nic: mtu 1500 qdisc fq_codel master virbr0 > state DOWN group default qlen 1000 link/ether 52:54:00:4d:96:e4 brd > ff:ff:ff:ff:ff:ff 13: qbr5ab32ca8-2d: mtu > 1500 qdisc noqueue state UP group default qlen 1000 link/ether > 76:0e:be:40:a6:ee brd ff:ff:ff:ff:ff:ff 14: qvo5ab32ca8-2d at qvb5ab32ca8-2d: > mtu 1500 qdisc noqueue master ovs-system > state UP group default qlen 1000 link/ether 12:90:d5:07:2a:a2 brd > ff:ff:ff:ff:ff:ff inet6 fe80::1090:d5ff:fe07:2aa2/64 scope link valid_lft > forever preferred_lft forever 15: qvb5ab32ca8-2d at qvo5ab32ca8-2d: > mtu 1500 qdisc noqueue master > qbr5ab32ca8-2d state UP group default qlen 1000 link/ether > 76:0e:be:40:a6:ee brd ff:ff:ff:ff:ff:ff inet6 fe80::740e:beff:fe40:a6ee/64 > scope link valid_lft forever preferred_lft forever 16: tap5ab32ca8-2d: > mtu 1500 qdisc fq_codel master > qbr5ab32ca8-2d state UNKNOWN group default qlen 1000 link/ether > fe:16:3e:1d:48:45 brd ff:ff:ff:ff:ff:ff inet6 fe80::fc16:3eff:fe1d:4845/64 > scope link valid_lft forever preferred_lft forever # cat > /etc/neutron/plugins/ml2/openvswitch_agent.ini [ovs] local_ip = > 10.12.21.142 datapath_type = system bridge_mappings = provider:br0 [vxlan] > enable_vxlan = true local_ip = 10.12.21.142 l2_population = true > prevent_arp_spoofing = True [securitygroup] enable_security_group = True > firewall_driver = > neutron.agent.linux.iptables_firewall.OVSHybridIptablesFirewallDriver # > cat /etc/neutron/dhcp_agent.ini [DEFAULT] interface_driver = > neutron.agent.linux.interface.OVSInterfaceDriver dhcp_driver = > neutron.agent.linux.dhcp.Dnsmasq enable_isolated_metadata = true > > I would very very appreciate any kind of guidance or help. > > Thanks, > Han Guangyu > -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: 3008DC07 at 74C0114D.D1F8AD61 Type: application/octet-stream Size: 31847 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: 3307DD07 at EEDEB41E.D1F8AD61 Type: application/octet-stream Size: 8758 bytes Desc: not available URL: From hanguangyu at uniontech.com Tue Dec 7 13:05:25 2021 From: hanguangyu at uniontech.com (=?utf-8?B?6Z+p5YWJ5a6H?=) Date: Tue, 7 Dec 2021 21:05:25 +0800 Subject: [neutron] Instances can't get IP from the DHCP server in OpenStack In-Reply-To: References: Message-ID: Hi , Thank you very very much, you can ask any questions. > If I understand well you have a single host all-in-one setup, am I right?< Yes, you are right. > You mentioned flat network, for that please check this document to be sure that you have everything set correctly: > https://docs.openstack.org/neutron/latest/admin/deploy-ovs-provider.html I have configured and recreated the network and instances according to this document. But the network behaves the same and still has no DHCP response. All settings is in line with this document, except the single host only have one nic, so I put all network in it, such as Management network adn Provider network. If you need any information, I will provider it soon. paste more details information in here: https://paste.opendev.org/show/811513/ Regards Han Guangyu ----------------- Original ------------------From:  "Lajos Katona" -------------- next part -------------- A non-text attachment was scrubbed... Name: D4C14173 at 2703EC48.155CAF61 Type: application/octet-stream Size: 31847 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: 489E9CE5 at 068B6A33.155CAF61 Type: application/octet-stream Size: 8758 bytes Desc: not available URL: From hanguangyu at uniontech.com Tue Dec 7 13:07:30 2021 From: hanguangyu at uniontech.com (=?utf-8?B?6Z+p5YWJ5a6H?=) Date: Tue, 7 Dec 2021 21:07:30 +0800 Subject: [neutron] Instances can't get IP from the DHCP server in OpenStack In-Reply-To: References: Message-ID: Hi ,  Thank you very very much, you can ask any questions.  > If I understand well you have a single host all-in-one setup, am I right?  Yes, you are right.  > You mentioned flat network, for that please check this document to be sure that you have everything set correctly: > https://docs.openstack.org/neutron/latest/admin/deploy-ovs-provider.html  I have configured and recreated the network and instances according  to this document. But the network behaves the same and still has no  DHCP response. All settings is in line with this document, except the  single host only have one nic, so I put all network in it, such as  Management network adn Provider network.  If you need any information, I will provider it soon.  paste more details information in here:  https://paste.opendev.org/show/811513/  Regards Han Guangyu      ------------------ Original ------------------ From:  "Lajos Katona" -------------- next part -------------- A non-text attachment was scrubbed... Name: D4C14173 at AEB01124.925CAF61 Type: application/octet-stream Size: 31847 bytes Desc: not available URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: 489E9CE5 at DB4EA60E.925CAF61 Type: application/octet-stream Size: 8758 bytes Desc: not available URL: From gmann at ghanshyammann.com Wed Dec 8 01:26:16 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Tue, 07 Dec 2021 19:26:16 -0600 Subject: [all][ptl][tc] Updates on Community wide goal: Current Active goals Message-ID: <17d97a5f9d5.d54b36e9564861.7457402543417580254@ghanshyammann.com> Hello Everyone, You might know, TC has de-coupled the community-wide goals from the release cycle[1]. As per the new structure, community-wide goals are tied or have to be finished within the release cycle. Instead, it can have target dates which can be a particular cycle release date or multi-cycle or even in between a release cycle. Target dates will be selected by considering the amount of work needed and the community bandwidth. This will help to make sure the goal is ready to start and also continue working on the goal until work is completed. 'Secure RBAC' goal is one example of a new structure that has different milestones targeting the different set of work[2]. I think this will be helpful for PTLs and project teams to plan the active goals for their project. Current Selected (Active) Goals: ------------------------------------- We have two Selected/Active goals. The project team can start/continue working on: 1. Migrate from oslo.rootwrap to oslo.privsep [3]. This is contouring from the previous cycle. 2. Consistent and Secure Default RBAC [4]. New design/implementation details are finalized now, please read those and for any queries, plan to attend the policy popup team biweekly meeting[5]. [1] https://review.opendev.org/c/openstack/governance/+/816387 [2] https://governance.openstack.org/tc/goals/selected/consistent-and-secure-rbac.html#completion-date-criteria [3] https://governance.openstack.org/tc/goals/selected/migrate-to-privsep.html [4] https://governance.openstack.org/tc/goals/selected/consistent-and-secure-rbac.html [5] https://wiki.openstack.org/wiki/Consistent_and_Secure_Default_Policies_Popup_Team#Meeting -gmann From manchandavishal143 at gmail.com Wed Dec 8 06:35:15 2021 From: manchandavishal143 at gmail.com (vishal manchanda) Date: Wed, 8 Dec 2021 12:05:15 +0530 Subject: [horizon] No Weekly meeting today Message-ID: Hello Team, Since there are no agenda items [1] to discuss for today's horizon weekly meeting. So let's cancel today's weekly meeting. Next weekly meeting will be on 15 December. Thanks & Regards, Vishal Manchanda [1] https://etherpad.opendev.org/p/horizon-release-priorities -------------- next part -------------- An HTML attachment was scrubbed... URL: From norman.kern at gmx.com Wed Dec 8 09:56:36 2021 From: norman.kern at gmx.com (norman.kern) Date: Wed, 8 Dec 2021 17:56:36 +0800 Subject: How to config vGPU in cyborg Message-ID: <188f229d-e777-9bde-5475-0c77e51577b9@gmx.com> Hi guys, I want to config cyborg with kolla in openstack wallaby cluster, but I found that few docs can be found to config vGPU, Where can I find the details in configuration from openstack website?? And I found another question, Cyborg can not config different types of vGPU for a pGPU. Waiting for you replies. Thanks. From bdobreli at redhat.com Wed Dec 8 10:46:46 2021 From: bdobreli at redhat.com (Bogdan Dobrelya) Date: Wed, 8 Dec 2021 11:46:46 +0100 Subject: [ops] [kolla] RabbitMQ High Availability Message-ID: <5dd6d28f-9955-7ca5-0ab8-0c5ce11ceb7e@redhat.com> Please see inline >> I read this with great interest because we are seeing this issue. Questions: >> >> 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? >> 2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? >> 3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? > > Note that even having rabbit HA policies adjusted like that and its HA > replication factor [0] decreased (e.g. to a 2), there still might be > high churn caused by a large enough number of replicated durable RPC > topic queues. And that might cripple the cloud down with the incurred > I/O overhead because a durable queue requires all messages in it to be > persisted to a disk (for all the messaging cluster replicas) before they > are ack'ed by the broker. > > Given that said, Oslo messaging would likely require a more granular > control for topic exchanges and the durable queues flag - to tell it to > declare as durable only the most critical paths of a service. A single > config setting and a single control exchange per a service might be not > enough. Also note that therefore, amqp_durable_queue=True requires dedicated control exchanges configured for each service. Those that use 'openstack' as a default cannot turn the feature ON. Changing it to a service specific might also cause upgrade impact, as described in the topic [3]. [3] https://review.opendev.org/q/topic:scope-config-opts > > There are also race conditions with durable queues enabled, like [1]. A > solution could be where each service declare its own dedicated control > exchange with its own configuration. > > Finally, openstack components should add perhaps a *.next CI job to test > it with durable queues, like [2] > > [0] https://www.rabbitmq.com/ha.html#replication-factor > > [1] > https://zuul.opendev.org/t/openstack/build/aa514dd788f34cc1be3800e6d7dba0e8/log/controller/logs/screen-n-cpu.txt > > [2] https://review.opendev.org/c/openstack/nova/+/820523 > >> >> Does anyone have a sample set of RMQ config files that they can share? >> >> It looks like my Outlook has ruined the link; reposting: >> [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando -- Best regards, Bogdan Dobrelya, Irc #bogdando From thierry at openstack.org Wed Dec 8 11:16:22 2021 From: thierry at openstack.org (Thierry Carrez) Date: Wed, 8 Dec 2021 12:16:22 +0100 Subject: [all][tc] Relmgt team position on release cadence In-Reply-To: References: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> <8301ef05-4011-43e9-beec-66f6aeeb00f9@www.fastmail.com> <20211129130957.3poysdeqsufkp3pg@yuggoth.org> <17d72ae5ee2.b80bfd38119150.5945079300853303270@ghanshyammann.com> <20211130220913.ydk5zhyvkdl7g7zc@yuggoth.org> <6536dfae-cb75-4263-856d-f4dbdfa2db01@www.fastmail.com> Message-ID: <13d6c4a5-b508-c9ed-8392-b6aaaeb70774@openstack.org> Jean-Philippe Evrard wrote: > On Sat, Dec 4, 2021, at 15:56, Thierry Carrez wrote: >> Jean-Philippe Evrard wrote: >>> [...] >>> Here is what I read: >>> 1) Many want more releases, not less. I haven't seen a complaint about tagging more releases. >>> 2) More than one person is proposing to abandon the integrated release, and nobody has complained about it. >> >> Here are a few drawbacks of abandoning the integrated release (which is >> really a "synchronized release"): >> >> - You would no longer have a baseline of components that are heavily >> tested together which we do community QA on and (collectively) clearly >> sign off on, so downstream integrators are on their own and have much >> more work to do > > I don't think so. We can still have that without the synchronized release. We decide what we test together. > (Maybe the TC can define that?). For example, with Zuul, we can test the master branch of all projects together to ensure the latest branch always work according to the set criteria. (No change there!) > > Then, what becomes "openstack version x" is just about a manifest of the SHAs or the tags of the projects, tested together. (No change again). (Let's call this proposition 1) I'm still wrapping my head around your proposal... It appears you want to drop the "stable branch" part of the release rather than the regular tagging and tracking of what has been tested together (the "integrated release"). It definitely sounds possible to me to: - have all components use an intermediary-released model (release as needed, no common feature freeze, no RCs) - have regular points in time where we communicate combinations of components that work together - not create stable branches for those components, not backport any bugfix and just roll forward (single branch development) Then you would still have a comparison baseline ("I run X"), and you would still have "openstack releases" that you can communicate and generate excitement around. And it would definitely reduce the backporting work happening upstream. However I am not sure I see how that proposal solves the upgrade pressure, or make downstream distribution work any easier... which are the two major reasons people ask for a longer cycle in the current system. If anything, that would make both worse, no? -- Thierry Carrez (ttx) From ekuvaja at redhat.com Wed Dec 8 12:44:39 2021 From: ekuvaja at redhat.com (Erno Kuvaja) Date: Wed, 8 Dec 2021 12:44:39 +0000 Subject: [all][tc] Relmgt team position on release cadence In-Reply-To: References: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> <8301ef05-4011-43e9-beec-66f6aeeb00f9@www.fastmail.com> <20211129130957.3poysdeqsufkp3pg@yuggoth.org> <17d72ae5ee2.b80bfd38119150.5945079300853303270@ghanshyammann.com> <20211130220913.ydk5zhyvkdl7g7zc@yuggoth.org> <6536dfae-cb75-4263-856d-f4dbdfa2db01@www.fastmail.com> Message-ID: On Tue, Dec 7, 2021 at 4:42 PM Jean-Philippe Evrard wrote: > Hello, > > On Sat, Dec 4, 2021, at 15:56, Thierry Carrez wrote: > > Jean-Philippe Evrard wrote: > >> [...] > >> Here is what I read: > >> 1) Many want more releases, not less. I haven't seen a complaint about > tagging more releases. > >> 2) More than one person is proposing to abandon the integrated release, > and nobody has complained about it. > > > > Here are a few drawbacks of abandoning the integrated release (which is > > really a "synchronized release"): > > > > - You would no longer have a baseline of components that are heavily > > tested together which we do community QA on and (collectively) clearly > > sign off on, so downstream integrators are on their own and have much > > more work to do > > I don't think so. We can still have that without the synchronized release. > We decide what we test together. > (Maybe the TC can define that?). For example, with Zuul, we can test the > master branch of all projects together to ensure the latest branch always > work according to the set criteria. (No change there!) > > Then, what becomes "openstack version x" is just about a manifest of the > SHAs or the tags of the projects, tested together. (No change again). > (Let's call this proposition 1) > > We don't have multiple "synchronized" releases of stable branches, so we > don't even have to bring any "openstack version x.y". Still, it doesn't > prevent us to define any "openstack version x.y" if we want to, with an > updated manifest compared to "version x.0" > > Technically proposition 1 looks like we run a deploy tool + tempest before > release, based on the manifest. > It's literally no new technology. I am relatively sure many (if not all) > deployment tools can do that. > > What makes more sense for me is that we define "openstack version x" in > terms of APIs, and that we have the test tooling to ensure that software > wanted to be integrated into "version x" are passing said tests. (Let's > call this proposition 2) > It allows alternative implementation of the software, if someone is crazy > enough to do that. > > I agree it might look like "We are abandonning the branches, what will we > do for people deploying regularily on a single branch?". Well it doesn't > change much: Those are currently consuming the manifests of releases, > bumping the SHAs from a "stable" branch of a repo. If a project/repo > _decides to branch_ for a very critical reason, then it can still happen. > If there is no reason to branch, then you would still bump the latest > branch available. Still no issue there. > > > - You would no longer have clearly-comparable points between > > distributions. Everyone knows what "running Ussuri" means, which > > facilitates communication and bug report handling in the community. > > I am not sure what "running Ussuri" means. Is that when the branch was > cut, or the latest sha of all the projects' Ussuri branches? Having > "OpenStack version Ussuri 1" corresponding to a manifest of project SHAs > and or APIs versions is far more clear to me. > > > - You would no longer have clear community support commitments. We > > currently maintain and fix bug reports from people running vanilla > > "Ussuri"... but do we want to care about every combination of components > > under the sun? (maybe we do already) > > We don't stop maintainers for contributing by being clearer in what we > release and how we branch... > If people want to maintain some old version and _need to branch a > project_, I feel it's still possible. But it's now in the power of the > project to decide whether it makes sense to do so, instead of being forced > to manage something that might be stretching the teams thin. > > > - You would no longer have "OpenStack" released, so you miss the regular > > marketing opportunity to remind the rest of the world that it still > > exists. The OpenStack brand fades, and it gets more complicated to get > > development resources to work on it. > > Again, it's wording. Please see my proposal above. > I understand why it's a concern for the foundation however ;) > > > Without the synchronized release, OpenStack essentially becomes a > > rolling distribution of cloud components on which we make very limited > > guarantees. I guess it is suitable to build maintained distributions on, > > but it really is no longer directly usable beyond development. Is that > > what we want? > > We would indeed be more "rolling", but it doesn't prevent tagging/point in > time testing and quality assurance. > Also, if we have integrated testing during the cycle and before the > tagging, then it doesn't remove any guarantee, does it? > > I disagree it's not usable beyond development. Here is my reasoning: > 1) I don't see it as removing any guarantee if we test things out :) > 2) Users of openstack are most likely using a deployment tooling (based on > configuration management tool of choice) to deploy their clouds, not a > manual deployment. This seems to be confirmed by the user survey. Do you > mean that a change in the model of branching would irreversibly break all > those tools, and make the ecosystem "no longer directly usable beyond > development"? > > Keep in mind I see those deploy toolings as part of openstack. Not only > because they are part of the ecosystem, but because some are under > OpenStack governance (OpenStack charms, openstack-chef, puppet openstack, > osa, openstack-helm, kolla, tripleO). > > Hence I consider that OpenStack would still be usable beyond development. > I know at least a company that would continue to believe in OpenStack ;) > > >> 3) Many people seem eager to carry "stable branches" for "critical > patches", but no new definition of such criticality was done. > > > > Note that a common stable branch cut is *exactly* the same thing as a > > synchronized release... So I see 2 and 3 as being opposed views. > > I agree with you there on common stable branch = synchronized release, > especially if "criticality" = "Whenever we decide to cut a release". I > still wanted to mention what was said, for the reader. > > It doesn't mean that everyone agree on the definition of criticality yet, > or maybe I am wrong? ;) > > Next to this, I have questions: > A) Am I the only one wanting to act on this? > Probably > B) Am I the only one caring? > Nope > C) What should be the next steps if I want to change this model? > D) Should I propose a change in governance, sync with release management > team? As this goes deep into the foundation's model, I feel like we need a > proper coordination to make this happen. > Likely a patches to gov repos are needed (ref C & D) > E) Do you consider all the energy spent to change things does not bring > enough positive reward for the whole ecosystem? > Absolutely, as someone who needs to think both upstream and downstream aspects of the release and has cared a lot about stable maintenance of OpenStack for years, what you're proposing here is literally pushing all our stable maintenance work to downstream and not even trying to share the efforts with the rest of the community. I also feel like this would turn really quickly to the point where if you're not following master on your deployment or using our product directly, I'd be happy to avice you to upgrade to the latest release and see if your problem still exists or go and talk to your deployment tool guys if they have seen the issue with whatever hashes they happen to deploy. I could have seen this as a tempting model 8-9 years ago when the competition was who is trailing master the closest and everyone was focusing on getting the next absolutely necessary feature set in to make OpenStack usable. Not now when we have matured considerably and are looking to provide stable production environments rather than rolling new code into production on a weekly basis. I'm not sure I even managed to grasp your proposal fully, but it really feels like half a step forward and mile, mile and half backwards. > > I see people talking about changing releases for years now, I haven't seen > a single change in our behaviour. (or maybe I missed something?). Is that > stockholm syndrome? ;) > > I see this as a matter that handful few want to bring up every few months (for years now, we've had this discussion probably close to dozen times) and I have a feeling that the majority of the community is just tired of copypasting the same arguments every round to avoid breaking what works and genuinely is waiting for the thread being buried for the next few months so they can get back to work. - Erno "jokke" Kuvaja > JP > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From dtantsur at redhat.com Wed Dec 8 13:11:42 2021 From: dtantsur at redhat.com (Dmitry Tantsur) Date: Wed, 8 Dec 2021 14:11:42 +0100 Subject: [ironic] [ops] RFC: deprecating instance network boot in Ironic Message-ID: Hi folks, As discussed at the PTG, I'd like to collect thoughts about this proposal. I would like to deprecate support for network-booting **normal instances**. This does not concern the deployment process, ramdisk deploy or booting from volume! Historically, Ironic has been able to boot the final instances from network (or ISO in case of virtual media). This is no longer the recommended mode, except for the ramdisk deploy or boot-from-volume. It adds quite a bit of complexity to deploy interfaces and has one serious security implication: there is no way to update kernel/ramdisk once the instance is deployed (without rebuilding it). For this reason, I don't think anybody should use instance network booting in production. I would like to announce the deprecation in the Yoga cycle with the intention of removing the feature in Z. Then we can also get rid of the boot_option capability (which is not really a capability). Am I missing something? Any objections? Dmitry -- Red Hat GmbH, https://de.redhat.com/ , Registered seat: Grasbrunn, Commercial register: Amtsgericht Muenchen, HRB 153243, Managing Directors: Charles Cachera, Brian Klemm, Laurie Krebs, Michael O'Neill -------------- next part -------------- An HTML attachment was scrubbed... URL: From senrique at redhat.com Wed Dec 8 13:42:52 2021 From: senrique at redhat.com (Sofia Enriquez) Date: Wed, 8 Dec 2021 10:42:52 -0300 Subject: [cinder] Bug deputy report for week of 12-08-2021 Message-ID: This is a bug report from 11-24-2021 to 12-08-2021. Agenda: https://etherpad.opendev.org/p/cinder-bug-squad-meeting ----------------------------------------------------------------------------------------- Medium - https://bugs.launchpad.net/cinder/+bug/1953168 ``Argument error in log message capacity filter". Assigned to caiquemello and Fix proposed. - https://bugs.launchpad.net/cinder/+bug/1952804 "cinder schedules backups on disabled backup services". Unassigned. - https://bugs.launchpad.net/cinder/+bug/1953185 "[IBM Storwize] Restore VM snapshot issue with replication volumes". Unassigned. - https://bugs.launchpad.net/cinder/+bug/1952927 "HPE 3PAR: migrating the vm attached with volume has been retyped will cause old volume to attach back again". Unassigned. - https://bugs.launchpad.net/cinder/+bug/1952805 "[posix] cinder schedules incremental backups on the wrong node". Unassigned. - https://bugs.launchpad.net/cinder/+bug/1952355 "[powerflex] Allocated capacity calculation doesn't work with powerflex driver". Unassigned. Low - https://bugs.launchpad.net/cinder/+bug/1953050 "A service should use a service specific control exchange for rpc". Fix proposed on master. *Quotas and more:* - https://bugs.launchpad.net/cinder/+bug/1952420 "Quota shows warnings on backup creation". Assigned to Gorka and Fix proposed. - https://bugs.launchpad.net/cinder/+bug/1952635 "When changing no_snapshot_gb_quota the gigabytes are wrong". Assigned to Gorka and Fix proposed. - https://bugs.launchpad.net/cinder/+bug/1952463 " Quota defaults and limits are not deleted when a volume type is deleted". Assigned to Gorka and Fix proposed. - https://bugs.launchpad.net/cinder/+bug/1952443 "Slow listings with many projects and deleted resources". Assigned to Gorka and Fix proposed. - https://bugs.launchpad.net/cinder/+bug/1952456 " Incorrect limit for private type volumes". Assigned to Gorka and Fix proposed. - https://bugs.launchpad.net/cinder/+bug/1952683 "Documentation typos". Fix proposed on master. Incomplete - https://bugs.launchpad.net/cinder/+bug/1952998 "[Ceph] Volume backup created with multiple snapshots preventing removal". Unassigned. Cheers -- Sof?a Enriquez she/her Software Engineer Red Hat PnT IRC: @enriquetaso @RedHat Red Hat Red Hat -------------- next part -------------- An HTML attachment was scrubbed... URL: From ekuvaja at redhat.com Wed Dec 8 13:50:46 2021 From: ekuvaja at redhat.com (Erno Kuvaja) Date: Wed, 8 Dec 2021 13:50:46 +0000 Subject: [all][tc] Relmgt team position on release cadence In-Reply-To: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> References: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> Message-ID: On Fri, Nov 5, 2021 at 2:39 PM Thierry Carrez wrote: > Hi everyone, > > The (long) document below reflects the current position of the release > management team on a popular question: should the OpenStack release > cadence be changed? Please note that we only address the release > management / stable branch management facet of the problem. There are > other dimensions to take into account (governance, feature deprecation, > supported distros...) to get a complete view of the debate. > > Introduction > ------------ > > The subject of how often OpenStack should be released has been regularly > debated in the OpenStack community. OpenStack started with a 3-month > release cycle, then switched to 6-month release cycle starting with > Diablo. It is often thought of a release management decision, but it is > actually a much larger topic: a release cadence is a trade-off between > pressure to release more often and pressure to release less often, > coming in from a lot of different stakeholders. In OpenStack, it is > ultimately a Technical Committee decision. But that decision is informed > by the position of a number of stakeholders. This document gives > historical context and describes the current release management team > position. > > The current trade-off > --------------------- > > The main pressure to release more often is to make features available to > users faster. Developers get a faster feedback loop, hardware vendors > ensure software is compatible with their latest products, and users get > exciting new features. "Release early, release often" is a best practice > in our industry -- we should generally aim at releasing as often as > possible. > > But that is counterbalanced by pressure to release less often. From a > development perspective, each release cycle comes with some process > overhead. On the integrators side, a new release means packaging and > validation work. On the users side, it means pressure to upgrade. To > justify that cost, there needs to be enough user-visible benefit (like > new features) in a given release. > > For the last 10 years for OpenStack, that balance has been around six > months. Six months let us accumulate enough new development that it was > worth upgrading to / integrating the new version, while giving enough > time to actually do the work. It also aligned well with Foundation > events cadence, allowing to synchronize in-person developer meetings > date with start of cycles. > > For sure I'm not talking on behalf of every project (or I might be, but I just don't know the dynamics across well enough). Anyways I think this assessment is missing one critical point, which is a release being the break off point for bikeshedding. I see a lot of genuine urgency to finally make a decision that has been going back and forth from the early cycle and the difference of opinion being finally solved when we're hitting the feature freeze / RC time. This does not only apply to feature work but lots of bugs as well that do not have active community members pushing them to get fixed (and backported) early. That push before RC is tagged is and seems to have been steadily intense few weeks before release, it's taxing for sure but it also ensures that we do get things done or make a real active decision to push it down the line for at least another half a year. I think the biggest actual drawback of longer release cycle is to lose this checkpoint (lets be honest here, no-one cares about milestones). Not that longer release would only make those hard decisions of "Are we including the work in this release or not" be even rarer occasion but the push before RC would intensify a lot when we have double the time of accumulating the review dept before we actually have to have that discussion. I think more of valuable contributions would be lost by losing traction (and people just deciding to carry the patches as one more downstream only thing as the community can't get around it). What changed > ------------ > > The major recent change affecting this trade-off is that the pace of new > development in OpenStack slowed down. The rhythm of changes was divided > by 3 between 2015 and 2021, reflecting that OpenStack is now a mature > and stable solution, where accessing the latest features is no longer a > major driver. That reduces some of the pressure for releasing more > often. At the same time, we have more users every day, with larger and > larger deployments, and keeping those clusters constantly up to date is > an operational challenge. That increases the pressure to release less > often. In essence, OpenStack is becoming much more like a LTS > distribution than a web browser -- something users like moving slow. > > Over the past years, project teams also increasingly decoupled > individual components from the "coordinated release". More and more > components opted for an independent or intermediary-released model, > where they can put out releases in the middle of a cycle, making new > features available to their users. This increasingly opens up the > possibility of a longer "coordinated release" which would still allow > development teams to follow "release early, release often" best > practices. All that recent evolution means it is (again) time to > reconsider if the 6-month cadence is what serves our community best, and > in particular if a longer release cadence would not suit us better. > > The release management team position on the debate > -------------------------------------------------- > > While releasing less often would definitely reduce the load on the > release management team, most of the team work being automated, we do > not think it should be a major factor in motivating the decision. We > should not adjust the cadence too often though, as there is a one-time > cost in switching our processes. In terms of impact, we expect that a > switch to a longer cycle will encourage more project teams to adopt a > "with-intermediary" release model (rather than the traditional "with-rc" > single release per cycle), which may lead to abandoning the latter, > hence simplifying our processes. Longer cycles might also discourage > people to commit to PTL or release liaison work. We'd probably need to > manage expectations there, and encourage more frequent switches (or > create alternate models). > While most of the PTL positions have not been resolved in elections for a while, this is a great note to keep in mind. Fundamentally we should keep that process as the main means to select new PTLs and pushing more and more towards handovers without even room for debate might bite us one day. > > If the decision is made to switch to a longer cycle, the release > management team recommends to switch to one year directly. That would > avoid changing it again anytime soon, and synchronizing on a calendar > year is much simpler to follow and communicate. We also recommend > announcing the change well in advance. We currently have an opportunity > of making the switch when we reach the end of the release naming > alphabet, which would also greatly simplify the communications around > the change. > > Finally, it is worth mentioning the impact on the stable branch work. > Releasing less often would likely impact the number of stable branches > that we keep on maintaining, so that we do not go too much in the past > (and hit unmaintained distributions or long-gone dependencies). We > currently maintain releases for 18 months before they switch to extended > maintenance, which results in between 3 and 4 releases being maintained > at the same time. We'd recommend switching to maintaining one-year > releases for 24 months, which would result in between 2 and 3 releases > being maintained at the same time. Such a change would lead to longer > maintenance for our users while reducing backporting work for our > developers. > > -- > Thierry Carrez (ttx) > On behalf of the OpenStack Release Management team > > In general my 2 cents for the proposals and couple of ideas maybe to consider easing the pain of current model: I really think the slowing phase (not due to lack of work items) is real and the biggest concern I have for a longer cycle is that it would be driving us to lose even more momentum and valuable contributions (see my comment at the end of "current trade-off"). It would also increase a lot of the pressure to backport features (no matter how much we claim this not being the case, we've seen it very clearly first hand downstream when we moved our product cycle to cover multiple upstream development cycles). Like Dan mentioned early on this thread the downstream releases from different sources won't align anyways and I think would contribute even more towards the drive of doing work downstream rather than upstream. In general I hate the idea of LTS even more than a longer cycle as then you effectively need to maintain 2 forks and put a lot of pressure on future contributors to align them just so that you can have yet another forking point to start diverting again. Perhaps a couple of things we could consider for easing the pain of upgrades over multiple cycles, the load of release itself and still balance the momentum of the "break point": 1) Compress the last few weeks of release with maybe aligning feature freeze with RC and bringing the final library releases closer as well. This might require us to have RC of clients to ease the integration pressure. This could result to few more RCs being tagged as things are being discovered later but them tags are fairly cheap. 2) Have any breaking changes (removing of deprecated anything) and disruptive db migrations happening only on cycle released in the second half of the year, while the first half would focus on bug fixes, non-intrusive feature work, etc. 3) move to max single community goal per year that targets to be finished on that second release (as they seem to be more and more disruptive in nature). As a bonus point I'd like to see a call to action to cut the test loads we're generating. I think our gating across the projects has some serious redundancy on them (I still remember the panic when our average check and gate runs reached 1hr mark). It's been great to see the efforts of covering more with our testing and that is very important too, but I still think we're eating a lot of infra resources that could be freed up (specially easing the last weeks of any release point) without losing the quality of our testing. This would give us the opportunity to have a real coordinated release as a checkpoint to get things done, but allow distributions and consumers to worry about the major upgrade pain of only one release per year. It would give us still 2 times a year to hype about the new release and all the positives coming with it and keep the planning of work, resources & commitments in more manageable chunks. Most importantly, apart from not allowing the breaking changes in the first release of the year we should treat both of them as 1st class citizens of releases, not "Development and LTS" or anything like that, just concentration of pain to the later one. - Erno "jokke" Kuvaja -------------- next part -------------- An HTML attachment was scrubbed... URL: From lokendrarathour at gmail.com Wed Dec 8 13:54:13 2021 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Wed, 8 Dec 2021 19:24:13 +0530 Subject: [Triple0] [Undercloud] Overcloud deployment getting failed at Introspection In-Reply-To: References: Message-ID: Hi Team, Please help update for any support here. best regards, Lokendra On Tue, Dec 7, 2021 at 5:56 PM Lokendra Rathour wrote: > Hi Team, > we are trying to deploy the triple0 Stein release on Centos 7. Undercloud > is getting deployed successfully but the introspection of the overcloud > node is getting failed with the reason as below: > > *on the DNS podman logs:* > ()[root at undercloud dhcp-hostsdir]# tail -f > /var/log/ironic-inspector/dnsmasq.log > Dec 7 11:21:08 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) > f8:f2:1e:9e:82:51 ignored > Dec 7 11:21:08 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) > f8:f2:1e:9e:82:50 ignored > Dec 7 11:21:12 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) > f8:f2:1e:9e:82:51 ignored > > *Checking the TCPDUMP at the same time gives:* > 11:38:38.308660 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, > Request from f8:f2:1e:9e:82:51 (oui Unknown), length 300 > 11:38:38.435173 IP6 fe80::f53a:6448:4a29:e051 > ff02::16: HBH ICMP6, > multicast listener report v2, 1 group record(s), length 28 > 11:38:38.435184 IP6 fe80::f53a:6448:4a29:e051 > ff02::16: HBH ICMP6, > multicast listener report v2, 1 group record(s), length 28 > 11:38:41.155784 IP6 fe80::322d:3253:e3b2:316e > ff02::2: ICMP6, router > solicitation, length 8 > 11:38:41.394304 IP6 fe80::f53a:6448:4a29:e051 > ff02::2: ICMP6, router > solicitation, length 8 > 11:38:42.563182 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, > Request from f8:f2:1e:9e:82:50 (oui Unknown), length 300 > 11:38:42.611973 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, > Request from f8:f2:1e:9e:82:51 (oui Unknown), length 300 > > *We observed that DHCP Server is not offering the IP to the baremetal. * > *DHCP discovery comes at br-ctlplane but no ip is offered.* > > Any input would be really helpful. > > Best Regards, > Lokendra > > > -- ~ Lokendra www.inertiaspeaks.com www.inertiagroups.com skype: lokendrarathour -------------- next part -------------- An HTML attachment was scrubbed... URL: From vikarnatathe at gmail.com Wed Dec 8 14:04:33 2021 From: vikarnatathe at gmail.com (Vikarna Tathe) Date: Wed, 8 Dec 2021 19:34:33 +0530 Subject: [Triple0] [Undercloud] Overcloud deployment getting failed at Introspection In-Reply-To: References: Message-ID: Hi Lokendra, Please check your yaml file which you prepared for introspection. The mac address should be correct. Vikarna On Wed, 8 Dec, 2021, 19:30 Lokendra Rathour, wrote: > Hi Team, > > Please help update for any support here. > > best regards, > Lokendra > > > On Tue, Dec 7, 2021 at 5:56 PM Lokendra Rathour > wrote: > >> Hi Team, >> we are trying to deploy the triple0 Stein release on Centos 7. Undercloud >> is getting deployed successfully but the introspection of the overcloud >> node is getting failed with the reason as below: >> >> *on the DNS podman logs:* >> ()[root at undercloud dhcp-hostsdir]# tail -f >> /var/log/ironic-inspector/dnsmasq.log >> Dec 7 11:21:08 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) >> f8:f2:1e:9e:82:51 ignored >> Dec 7 11:21:08 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) >> f8:f2:1e:9e:82:50 ignored >> Dec 7 11:21:12 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) >> f8:f2:1e:9e:82:51 ignored >> >> *Checking the TCPDUMP at the same time gives:* >> 11:38:38.308660 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, >> Request from f8:f2:1e:9e:82:51 (oui Unknown), length 300 >> 11:38:38.435173 IP6 fe80::f53a:6448:4a29:e051 > ff02::16: HBH ICMP6, >> multicast listener report v2, 1 group record(s), length 28 >> 11:38:38.435184 IP6 fe80::f53a:6448:4a29:e051 > ff02::16: HBH ICMP6, >> multicast listener report v2, 1 group record(s), length 28 >> 11:38:41.155784 IP6 fe80::322d:3253:e3b2:316e > ff02::2: ICMP6, router >> solicitation, length 8 >> 11:38:41.394304 IP6 fe80::f53a:6448:4a29:e051 > ff02::2: ICMP6, router >> solicitation, length 8 >> 11:38:42.563182 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, >> Request from f8:f2:1e:9e:82:50 (oui Unknown), length 300 >> 11:38:42.611973 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, >> Request from f8:f2:1e:9e:82:51 (oui Unknown), length 300 >> >> *We observed that DHCP Server is not offering the IP to the baremetal. * >> *DHCP discovery comes at br-ctlplane but no ip is offered.* >> >> Any input would be really helpful. >> >> Best Regards, >> Lokendra >> >> >> > > -- > ~ Lokendra > www.inertiaspeaks.com > www.inertiagroups.com > skype: lokendrarathour > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From strigazi at gmail.com Wed Dec 8 14:35:46 2021 From: strigazi at gmail.com (Spyros Trigazis) Date: Wed, 8 Dec 2021 15:35:46 +0100 Subject: [magnum] dropping mesos code In-Reply-To: References: Message-ID: Hello Mohammed, I do not think anyone is interested in it anymore, we could just drop it. Even if resources are left behind they can be cleaned up with heat/other openstack APIs. Spyros On Sat, Dec 4, 2021 at 11:53 AM Mohammed Naser wrote: > Hi there, > > Does it make sense for us to drop the Mesos driver code from the code base? > > I don't know of anyone that actually uses it and I believe it's likely > not been tested for years. > > Mohammed > > -- > Mohammed Naser > VEXXHOST, Inc. > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From tobias.urdin at binero.com Wed Dec 8 14:50:01 2021 From: tobias.urdin at binero.com (Tobias Urdin) Date: Wed, 8 Dec 2021 14:50:01 +0000 Subject: [magnum] dropping mesos code In-Reply-To: References: Message-ID: Hello Mohammed, I agree, we should cleanup things that has been broken for years now. Best regards Tobias On 8 Dec 2021, at 15:35, Spyros Trigazis > wrote: Hello Mohammed, I do not think anyone is interested in it anymore, we could just drop it. Even if resources are left behind they can be cleaned up with heat/other openstack APIs. Spyros On Sat, Dec 4, 2021 at 11:53 AM Mohammed Naser > wrote: Hi there, Does it make sense for us to drop the Mesos driver code from the code base? I don't know of anyone that actually uses it and I believe it's likely not been tested for years. Mohammed -- Mohammed Naser VEXXHOST, Inc. -------------- next part -------------- An HTML attachment was scrubbed... URL: From vikarnatathe at gmail.com Wed Dec 8 15:22:48 2021 From: vikarnatathe at gmail.com (Vikarna Tathe) Date: Wed, 8 Dec 2021 20:52:48 +0530 Subject: [Triple0] [Undercloud] Overcloud deployment getting failed at Introspection In-Reply-To: References: Message-ID: Json file and not yaml file On Wed, 8 Dec, 2021, 19:34 Vikarna Tathe, wrote: > Hi Lokendra, > > Please check your yaml file which you prepared for introspection. The mac > address should be correct. > > Vikarna > > > On Wed, 8 Dec, 2021, 19:30 Lokendra Rathour, > wrote: > >> Hi Team, >> >> Please help update for any support here. >> >> best regards, >> Lokendra >> >> >> On Tue, Dec 7, 2021 at 5:56 PM Lokendra Rathour < >> lokendrarathour at gmail.com> wrote: >> >>> Hi Team, >>> we are trying to deploy the triple0 Stein release on Centos 7. >>> Undercloud is getting deployed successfully but the introspection of the >>> overcloud node is getting failed with the reason as below: >>> >>> *on the DNS podman logs:* >>> ()[root at undercloud dhcp-hostsdir]# tail -f >>> /var/log/ironic-inspector/dnsmasq.log >>> Dec 7 11:21:08 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) >>> f8:f2:1e:9e:82:51 ignored >>> Dec 7 11:21:08 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) >>> f8:f2:1e:9e:82:50 ignored >>> Dec 7 11:21:12 dnsmasq-dhcp[8]: DHCPDISCOVER(br-ctlplane) >>> f8:f2:1e:9e:82:51 ignored >>> >>> *Checking the TCPDUMP at the same time gives:* >>> 11:38:38.308660 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, >>> Request from f8:f2:1e:9e:82:51 (oui Unknown), length 300 >>> 11:38:38.435173 IP6 fe80::f53a:6448:4a29:e051 > ff02::16: HBH ICMP6, >>> multicast listener report v2, 1 group record(s), length 28 >>> 11:38:38.435184 IP6 fe80::f53a:6448:4a29:e051 > ff02::16: HBH ICMP6, >>> multicast listener report v2, 1 group record(s), length 28 >>> 11:38:41.155784 IP6 fe80::322d:3253:e3b2:316e > ff02::2: ICMP6, router >>> solicitation, length 8 >>> 11:38:41.394304 IP6 fe80::f53a:6448:4a29:e051 > ff02::2: ICMP6, router >>> solicitation, length 8 >>> 11:38:42.563182 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, >>> Request from f8:f2:1e:9e:82:50 (oui Unknown), length 300 >>> 11:38:42.611973 IP 0.0.0.0.bootpc > 255.255.255.255.bootps: BOOTP/DHCP, >>> Request from f8:f2:1e:9e:82:51 (oui Unknown), length 300 >>> >>> *We observed that DHCP Server is not offering the IP to the baremetal. * >>> *DHCP discovery comes at br-ctlplane but no ip is offered.* >>> >>> Any input would be really helpful. >>> >>> Best Regards, >>> Lokendra >>> >>> >>> >> >> -- >> ~ Lokendra >> www.inertiaspeaks.com >> www.inertiagroups.com >> skype: lokendrarathour >> >> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: From radoslaw.piliszek at gmail.com Wed Dec 8 15:39:53 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Wed, 8 Dec 2021 16:39:53 +0100 Subject: [kolla] Dropping support for vmtp in Yoga Message-ID: Dear Users of Kolla! The Kolla project, so far, provided vmtp [1] image and ansible role. However, the status of those has been unknown (the ansible role looks pretty stale, still referencing ubuntu 16.04!) and now the vmtp image is failing to build so we are forced to disable it. [2] The project itself [1] looks rather unmaintained and it's outside of the openstack namespace so our plan is to drop support for it entirely in this cycle (Yoga). Please reply to this mail if you are using vmtp with Kolla in some recent release and are willing to contribute to let us keep supporting it. [1] https://opendev.org/x/vmtp [2] https://review.opendev.org/c/openstack/kolla/+/820043 Kind regards, -yoctozepto From christian.rohmann at inovex.de Wed Dec 8 16:12:19 2021 From: christian.rohmann at inovex.de (Christian Rohmann) Date: Wed, 8 Dec 2021 17:12:19 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: Message-ID: <9a3cf269-6b89-fb32-6dc5-4fd708171cb2@inovex.de> On 03/12/2021 21:01, Cedric Lemarchand wrote: > We faced the same use case, and we end up doing something like this: > > - disable security on the network provider (or on the instance port > where pxe boot is needed) > - deploy or rescue the intance with an ipxe image > As for DHCP, you can very much use the OpenStack DHCP and just set the options like next-server / server-ip-address ?* https://docs.openstack.org/api-ref/network/v2/?expanded=update-port-detail#extra-dhcp-option-extra-dhcp-opt-extension ?* https://docs.openstack.org/neutron/latest/ovn/dhcp_opts.html and this IP then points to your VM with the TFTP and other PXE / iPXE tooling running. Regards Christian From sbauza at redhat.com Wed Dec 8 16:22:55 2021 From: sbauza at redhat.com (Sylvain Bauza) Date: Wed, 8 Dec 2021 17:22:55 +0100 Subject: [nova][placement] Cancelling Dec 21st and Dec 28th IRC meetings Message-ID: As we agreed on yesterday's meeting [1] we are about to cancel Dec 21st and Dec 28th meetings as we won't have a quorum for contributors during those weeks :-) -Sylvain [1] https://meetings.opendev.org/meetings/nova/2021/nova.2021-12-07-16.00.html -------------- next part -------------- An HTML attachment was scrubbed... URL: From openinfradn at gmail.com Wed Dec 8 16:34:35 2021 From: openinfradn at gmail.com (open infra) Date: Wed, 8 Dec 2021 22:04:35 +0530 Subject: cpu pinning In-Reply-To: References: Message-ID: Managed to set cpu_dedicated_setin nova. Thanks, Sean! On Thu, Dec 2, 2021 at 9:16 PM Sean Mooney wrote: > On Thu, 2021-12-02 at 08:58 +0530, open infra wrote: > > Hi, > > > > I have created a flavor with following properties and created an > instance. > > Instance failed with the error "No valid host was found. There are not > > enough hosts available." > > When I set the cpu policy as 'shared' I can create the instance. The > host > > machine has two numa nodes and a total of 128 vcpu. > > I can not figure out what's missing here. > i suspect the issue is not with the flavor but with yoru host configurtion. > > you likely need to defience cpu_dedicated_set and cpu_shared_set in the > nova.conf > > we do not support mixing pinned and floating cpus on the same host unless > you partion the cpu pool > using cpu_dedicated_set and cpu_shared_set. > > as of train cpu_dedicated_set replaced vcpu_pin_set as the supported way > to report the pool of cpus to be > used for pinned vms to placment. > > if you do "openstack resource provider inventory show " > it should detail the avaiabel pcpu and vcpu inventories. > when you use hw:cpu_policy='dedicated' it will claim PCPUs not VCPUs in > placment. > That is likely the issue you are encountering. > > by default we have a fallback query to make this work while you are > upgrading > > > https://docs.openstack.org/nova/latest/configuration/config.html#workarounds.disable_fallback_pcpu_query > > which we should be disabling by default soon. > > but i suspect that is likely why you are getting the no valid host. > > to debug this properly you should enable debug logging on the nova > schduler and then confirm if you got > host back form placment and then if the numa toplogy filter is rejectign > the host or not. > > without the schduler debug logs for the instance creation we cannt really > help more then this since we do not have the info required. > > > > controller-1:~$ openstack flavor show dn.large -c properties > > > > > +------------+--------------------------------------------------------------------------------------------------------+ > > > > > Field | Value > > | > > > > > +------------+--------------------------------------------------------------------------------------------------------+ > > > > > properties | hw:cpu_cores='2', hw:cpu_policy='dedicated', > > hw:cpu_sockets='1', hw:cpu_threads='2', hw:numa_nodes='1' | > > > > > +------------+--------------------------------------------------------------------------------------------------------+ > > > > controller-1:~$ openstack hypervisor stats show > > > > +----------------------+--------+ > > > > > Field | Value | > > > > +----------------------+--------+ > > > > > count | 1 | > > > > > current_workload | 0 | > > > > > disk_available_least | 187 | > > > > > free_disk_gb | 199 | > > > > > free_ram_mb | 308787 | > > > > > local_gb | 219 | > > > > > local_gb_used | 20 | > > > > > memory_mb | 515443 | > > > > > memory_mb_used | 206656 | > > > > > running_vms | 7 | > > > > > vcpus | 126 | > > > > > vcpus_used | 49 | > > > > +----------------------+--------+ > > > > > > > > Regards, > > > > Danishka > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From openstack at a.spamming.party Wed Dec 8 17:57:56 2021 From: openstack at a.spamming.party (Jean-Philippe Evrard) Date: Wed, 08 Dec 2021 18:57:56 +0100 Subject: [all][tc] Relmgt team position on release cadence In-Reply-To: <13d6c4a5-b508-c9ed-8392-b6aaaeb70774@openstack.org> References: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> <8301ef05-4011-43e9-beec-66f6aeeb00f9@www.fastmail.com> <20211129130957.3poysdeqsufkp3pg@yuggoth.org> <17d72ae5ee2.b80bfd38119150.5945079300853303270@ghanshyammann.com> <20211130220913.ydk5zhyvkdl7g7zc@yuggoth.org> <6536dfae-cb75-4263-856d-f4dbdfa2db01@www.fastmail.com> <13d6c4a5-b508-c9ed-8392-b6aaaeb70774@openstack.org> Message-ID: <080FC5C2-53D6-4EA5-B0E1-AF96295A44DF@evrard.me> Hello, On 8 Dec 2021, at 12:16, Thierry Carrez wrote: > I'm still wrapping my head around your proposal... It appears you want to drop the "stable branch" part of the release rather than the regular tagging and tracking of what has been tested together (the "integrated release"). > > It definitely sounds possible to me to: > - have all components use an intermediary-released model (release as needed, no common feature freeze, no RCs) > - have regular points in time where we communicate combinations of components that work together > - not create stable branches for those components, not backport any bugfix and just roll forward (single branch development) That?s what I am proposing indeed. I have added the exception that _projects_ can decide to go multi branch, where it makes sense. The TC decides what warrants a branch, to have a common behaviour across the whole openstack (see examples below). > Then you would still have a comparison baseline ("I run X"), and you would still have "openstack releases" that you can communicate and generate excitement around. And it would definitely reduce the backporting work happening upstream. Correct. > However I am not sure I see how that proposal solves the upgrade pressure, or make downstream distribution work any easier... which are the two major reasons people ask for a longer cycle in the current system. If anything, that would make both worse, no? 1) Small backports don?t have to happen until a big, disruptive, work happens. This relieves a bit of pressure on contributions, and make the contributions more meaningful tbh. (cf. the "oh it wasn't backported to this branch" syndrome ;)). 2) When a new branch is created, the _project_ could decide what to do with the branch. ?We will abandon this legacy code in x?. The TC might decide a series of rules of when to close branches (or if it?s even allowed) 3) Downstream distributions would still get maintained software, and in fact it would be easier to understand. e.g. for packagers: The master branch rolls forward, and has regular tags (intermediary-released), all of them are not disruptive for users/packaging. If a project decides to branch (because they brought a very disruptive change), then it could be considered for packagers as creating a ?new version? of a package, obsoleting an older package. Very simple for packagers AND users. 4) The result of a strong ?let?s never branch? model is that we can have easier upgrades: We can decide that the upgrades between two version of the same branch have to be seemless, while a switch from an older branch to a newer branch must require a ?migration? tool in place. Regards, Jean-Philippe Evrard (evrardjp) -------------- next part -------------- An HTML attachment was scrubbed... URL: From openstack at a.spamming.party Wed Dec 8 18:04:48 2021 From: openstack at a.spamming.party (Jean-Philippe Evrard) Date: Wed, 08 Dec 2021 19:04:48 +0100 Subject: [all][tc] Relmgt team position on release cadence In-Reply-To: References: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> <8301ef05-4011-43e9-beec-66f6aeeb00f9@www.fastmail.com> <20211129130957.3poysdeqsufkp3pg@yuggoth.org> <17d72ae5ee2.b80bfd38119150.5945079300853303270@ghanshyammann.com> <20211130220913.ydk5zhyvkdl7g7zc@yuggoth.org> <6536dfae-cb75-4263-856d-f4dbdfa2db01@www.fastmail.com> Message-ID: Hello, On Wed, Dec 8, 2021, at 13:44, Erno Kuvaja wrote: > Absolutely, as someone who needs to think both upstream and downstream aspects of the release and has cared a lot about stable maintenance of OpenStack for years, what you're proposing here is literally pushing all our stable maintenance work to downstream and not even trying to share the efforts with the rest of the community. Not at all. I am fine with sharing the effort with the community. We just need to be "smarter" about the branching, and do the efforts right. > I also feel like this would turn really quickly to the point where if you're not following master on your deployment or using our product directly, I'd be happy to avice you to upgrade to the latest release and see if your problem still exists or go and talk to your deployment tool guys if they have seen the issue with whatever hashes they happen to deploy. Let's be honest, when we are finding a bug, the first question we ask is "which version do you run", no? We don't ask for a branch, we ask for a SHA. Nothing changes there ;) > I could have seen this as a tempting model 8-9 years ago when the competition was who is trailing master the closest and everyone was focusing on getting the next absolutely necessary feature set in to make OpenStack usable. Not now when we have matured considerably and are looking to provide stable production environments rather than rolling new code into production on a weekly basis. I'm not sure I even managed to grasp your proposal fully, but it really feels like half a step forward and mile, mile and half backwards. I don't think it's ever a race to follow the latest master branch. It's never been for me at least. For me it's about: - being smart about the definition of branching - forcing certain requirements on when branch, to be consistent in the community, WHILE bringing useful features for users - giving freedom to projects (of course while keeping consistency with the help of the TC). Being stable is fine. In fact, it's the dream. Stable doesn't mean stale, however ;) >> >> I see people talking about changing releases for years now, I haven't seen a single change in our behaviour. (or maybe I missed something?). Is that stockholm syndrome? ;) >> > I see this as a matter that handful few want to bring up every few months (for years now, we've had this discussion probably close to dozen times) and I have a feeling that the majority of the community is just tired of copypasting the same arguments every round to avoid breaking what works and genuinely is waiting for the thread being buried for the next few months so they can get back to work. I feel the same. But instead of burying I would like to act. Regards, JP -------------- next part -------------- An HTML attachment was scrubbed... URL: From mgrinnell at datto.com Wed Dec 8 18:36:17 2021 From: mgrinnell at datto.com (Matthew Grinnell) Date: Wed, 8 Dec 2021 13:36:17 -0500 Subject: Swift Account Reaper Deletion Status Message-ID: Hello, I was curious about expected account-reaper behavior after deleting an account and if there is a way to be able to tell when all data for the account has been reaped. When I issue a delete on a swift account, I see subsequent HEAD requests on the account then return 410 Gone. Testing with debug logging I can see that after passing the delay_reaping time I see the reaper report my test object is deleted, HEAD requests against the account still return the same 410 Gone. Is there a way to tell when everything for an account has been purged, ie does the 410 return change at some point when that happens? Thanks, Matt -------------- next part -------------- An HTML attachment was scrubbed... URL: From smooney at redhat.com Wed Dec 8 18:51:20 2021 From: smooney at redhat.com (Sean Mooney) Date: Wed, 08 Dec 2021 18:51:20 +0000 Subject: [all][tc] Relmgt team position on release cadence In-Reply-To: References: <40f6f51b-6903-1afe-7318-0689b2af482d@openstack.org> <8301ef05-4011-43e9-beec-66f6aeeb00f9@www.fastmail.com> <20211129130957.3poysdeqsufkp3pg@yuggoth.org> <17d72ae5ee2.b80bfd38119150.5945079300853303270@ghanshyammann.com> <20211130220913.ydk5zhyvkdl7g7zc@yuggoth.org> <6536dfae-cb75-4263-856d-f4dbdfa2db01@www.fastmail.com> Message-ID: <4202932ad0aec67e82a305e0a3d3b23a5b7cc00d.camel@redhat.com> On Wed, 2021-12-08 at 19:04 +0100, Jean-Philippe Evrard wrote: > Hello, > > On Wed, Dec 8, 2021, at 13:44, Erno Kuvaja wrote: > > Absolutely, as someone who needs to think both upstream and downstream aspects of the release and has cared a lot about stable maintenance of OpenStack for years, what you're proposing here is literally pushing all our stable maintenance work to downstream and not even trying to share the efforts with the rest of the community. > > Not at all. I am fine with sharing the effort with the community. We just need to be "smarter" about the branching, and do the efforts right. i dont nessisarly think our corrent branching is in anyway not smart. > > > I also feel like this would turn really quickly to the point where if you're not following master on your deployment or using our product directly, I'd be happy to avice you to upgrade to the latest release and see if your problem still exists or go and talk to your deployment tool guys if they have seen the issue with whatever hashes they happen to deploy. > > Let's be honest, when we are finding a bug, the first question we ask is "which version do you run", no? We don't ask for a branch, we ask for a SHA. Nothing changes there ;) > > > I could have seen this as a tempting model 8-9 years ago when the competition was who is trailing master the closest and everyone was focusing on getting the next absolutely necessary feature set in to make OpenStack usable. Not now when we have matured considerably and are looking to provide stable production environments rather than rolling new code into production on a weekly basis. I'm not sure I even managed to grasp your proposal fully, but it really feels like half a step forward and mile, mile and half backwards. > > I don't think it's ever a race to follow the latest master branch. It's never been for me at least. > > For me it's about: > - being smart about the definition of branching > - forcing certain requirements on when branch, to be consistent in the community, WHILE bringing useful features for users > - giving freedom to projects (of course while keeping consistency with the help of the TC). for me creating a branch ahs alwasy been to state tha twe have forzen the feature set for a given releas andy you can integrate it into your product or produciton with the expectation that it will only recive bug fixes goign forward. unless we decide just to adopt a semver branching model i dont really see how what your propsoing will be benifical to consuming upstream release or prodcutising downstream to me this woudl be a regression and make my life harder. if we take a semver apprcoh and bump the major version only when backward incomaptable change are made and defience some suport boundary on when tha tis allowed then that might be an alternive branchign stragy that could work but nova at least still recived enough feature reuqest that we may still need to bump the major version every 6 months. with that said we did not make db schema change between train and wallaby so we have stablised some what compared to the early years where we had several in each release. > > Being stable is fine. In fact, it's the dream. Stable doesn't mean stale, however ;) > > > > > > > I see people talking about changing releases for years now, I haven't seen a single change in our behaviour. (or maybe I missed something?). Is that stockholm syndrome? ;) > > > > > I see this as a matter that handful few want to bring up every few months (for years now, we've had this discussion probably close to dozen times) and I have a feeling that the majority of the community is just tired of copypasting the same arguments every round to avoid breaking what works and genuinely is waiting for the thread being buried for the next few months so they can get back to work. > > I feel the same. But instead of burying I would like to act. > > Regards, > JP From yipikai7 at gmail.com Wed Dec 8 19:52:13 2021 From: yipikai7 at gmail.com (Cedric Lemarchand) Date: Wed, 8 Dec 2021 20:52:13 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: <9a3cf269-6b89-fb32-6dc5-4fd708171cb2@inovex.de> References: <9a3cf269-6b89-fb32-6dc5-4fd708171cb2@inovex.de> Message-ID: Thanks for the heads up. Pointer is about ovn, I didnt find anything regarding ovs. Did the Neutron API support these options both for ovn and ovs ? On Wed, Dec 8, 2021, 17:12 Christian Rohmann wrote: > On 03/12/2021 21:01, Cedric Lemarchand wrote: > > We faced the same use case, and we end up doing something like this: > > > > - disable security on the network provider (or on the instance port > > where pxe boot is needed) > > - deploy or rescue the intance with an ipxe image > > > > As for DHCP, you can very much use the OpenStack DHCP and just set the > options like next-server / server-ip-address > > * > > https://docs.openstack.org/api-ref/network/v2/?expanded=update-port-detail#extra-dhcp-option-extra-dhcp-opt-extension > * https://docs.openstack.org/neutron/latest/ovn/dhcp_opts.html > > > and this IP then points to your VM with the TFTP and other PXE / iPXE > tooling running. > > > > > Regards > > > Christian > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From gsteinmuller at vexxhost.com Wed Dec 8 20:42:51 2021 From: gsteinmuller at vexxhost.com (=?UTF-8?Q?Guilherme_Steinm=C3=BCller?=) Date: Wed, 8 Dec 2021 17:42:51 -0300 Subject: [magnum] dropping mesos code In-Reply-To: References: Message-ID: Hey there, I took the liberty to kick off a patch to start the mesos code removal [1]. It's a lot of lines of code and files removed lol But I believe it's going to be an improvement on the current code as mesos is not really being used out there. [1] https://review.opendev.org/c/openstack/magnum/+/821124 Regards Guilherme Steinmuller On Wed, Dec 8, 2021 at 11:54 AM Tobias Urdin wrote: > Hello Mohammed, > > I agree, we should cleanup things that has been broken for years now. > > Best regards > Tobias > > On 8 Dec 2021, at 15:35, Spyros Trigazis wrote: > > Hello Mohammed, > > I do not think anyone is interested in it anymore, we could just drop it. > Even if resources are left behind they can be cleaned up with heat/other > openstack APIs. > > Spyros > > On Sat, Dec 4, 2021 at 11:53 AM Mohammed Naser > wrote: > >> Hi there, >> >> Does it make sense for us to drop the Mesos driver code from the code >> base? >> >> I don't know of anyone that actually uses it and I believe it's likely >> not been tested for years. >> >> Mohammed >> >> -- >> Mohammed Naser >> VEXXHOST, Inc. >> >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Thu Dec 9 01:57:14 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Wed, 08 Dec 2021 19:57:14 -0600 Subject: [all][tc] Technical Committee next weekly meeting on Dec 9th at 1500 UTC In-Reply-To: <17d9221334f.11a9bcb48483928.5563521669059290320@ghanshyammann.com> References: <17d9221334f.11a9bcb48483928.5563521669059290320@ghanshyammann.com> Message-ID: <17d9ce8acee.ebacf8b4634894.6597607769014056305@ghanshyammann.com> Hello Everyone, Below is the agenda for Tomorrow's TC IRC meeting schedule at 1500 UTC. https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting == Agenda for tomorrow's TC meeting == * Roll call * Follow up on past action items * Gate health check ** Fixing Zuul config error in OpenStack *** https://etherpad.opendev.org/p/zuul-config-error-openstack * Pain points discussion meeting updates ** https://etherpad.opendev.org/p/pain-point-elimination * Yoga tracker ** https://etherpad.opendev.org/p/tc-yoga-tracker * Updates on community-wide goal ** http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026218.html ** Proposed community goal for FIPS compatibility and compliance *** https://review.opendev.org/c/openstack/governance/+/816587 * Adjutant need PTLs and maintainers ** http://lists.openstack.org/pipermail/openstack-discuss/2021-October/025555.html * Open Reviews ** https://review.opendev.org/q/projects:openstack/governance+is:open -gmann ---- On Mon, 06 Dec 2021 17:43:09 -0600 Ghanshyam Mann wrote ---- > Hello Everyone, > > Technical Committee's next weekly meeting is scheduled for Dec 9th at 1500 UTC. > > If you would like to add topics for discussion, please add them to the below wiki page by > Wednesday, Dec 8th, at 2100 UTC. > > https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting > > -gmann > > From zaitcev at redhat.com Thu Dec 9 02:00:31 2021 From: zaitcev at redhat.com (Pete Zaitcev) Date: Wed, 8 Dec 2021 20:00:31 -0600 Subject: Swift Account Reaper Deletion Status In-Reply-To: References: Message-ID: <20211208200031.0e51a552@niphredil.zaitcev.lan> On Wed, 8 Dec 2021 13:36:17 -0500 Matthew Grinnell wrote: > [...] I see the > reaper report my test object is deleted, HEAD requests against the account > still return the same 410 Gone. Is there a way to tell when everything for > an account has been purged, ie does the 410 return change at some point > when that happens? The account server continues to return 410 until replicator reclaims the account DB. swift/proxy/controllers/account.py: if resp.status_int == HTTP_NOT_FOUND: if resp.headers.get('X-Account-Status', '').lower() == 'deleted': resp.status = HTTP_GONE The basic problem here is that the Swift proxy cannot possibly know if your data is not hiding somewhere on a handoff that is coincidentaly offline. However, the reclaim period is large (one week) and operators are strictly told never let anything that's been down that long back into the cluster. So after a week replicator deletes the account DB and only then you have a "guarantee" that everything is gone. It is an operational guarantee though. Cannot get you a better one in a distributed system. If you really want to have things GONE, you have to encrypt them to begin with, then destroy keys at the decision time. The key store is centralized, so it's not a subject to the general distributed system problem as per above. -- Pete From amotoki at gmail.com Thu Dec 9 02:14:27 2021 From: amotoki at gmail.com (Akihiro Motoki) Date: Thu, 9 Dec 2021 11:14:27 +0900 Subject: [tc][sig][i18n] What's the status for Xena translation? Message-ID: Hi, # I include TC and SIG tags to get broader attractions to this topic. # Also Cc'ed to Ian (i18n SIG char) and Andreas. I would like to ask the status of Xena translation in translate.openstack.org (Zanata). In past releases, stable-XXX branches were created on Zanata, but a branch for Xena translation has not been created yet. Ajaeger usually prepared these branches as a volunteer (I am really appreciated) but AFAIK he no longer works on OpenStack actively. Who takes care of the translatio preparation around the release now? I hope the i18n SIG is not dead. Thanks, Akihiro Motoki (amotoki) From ricolin at ricolky.com Thu Dec 9 04:42:19 2021 From: ricolin at ricolky.com (Rico Lin) Date: Thu, 9 Dec 2021 12:42:19 +0800 Subject: [tc][all] Follow up pain point discussion In-Reply-To: References: Message-ID: Dear all First, thank all who join us in the discussion or help on the etherpad [1] Here are wrap-ups of our first meeting yesterday: In the meeting, we have quick work through pain points for a couple of projects: Cinder, Ironic, Neutron, QA, Horizon. And put some action or tags(like `FIX NEEDED` or `INFO NEEDED`). As for Nova, there is already have a good quality of replies and actions, gmann will help with discussing tags with rest of Nova team. Horizon team will go through the rest item in team meeting and finish tagging after. Also, TC picks the log level topic (check out log level discussion in [2]) as one of the topics in next TC meeting. So if you're interested, please check out [1] and kindly provide your feedback. We plan to go through with other projects in our next pain point meeting, which we expect to schedule around mid. of January. Meanwhile, if teams can help with providing replies and tags will be very helpful. As for possible cross-project pain points: The OSC discussion went very well. The discussion includes OpenStack Client and OpenStackSDK. OSC seems to get more works done than the last time it was proposed as a community goal (back in Train cycle?), so TC has set up a topic in the next TC meeting to discuss the possibility to set this as a community goal (for z or after-z cycles). RabbitMQ was discussed briefly but didn't have any specific suggestions or actions. So if you have any, please kindly provide your feedback in this ML or [3]. If you would like to suggest other cross-project pain points, kindly suggest in ML or [1] and help TCs to identify them. Finally, we collect pain points and hope to get things fixed/improved, so if you're interested to help with targeting any of the items in [1], you're most welcome to reach out to project teams or at least provide your feedback in [1]. [1] https://etherpad.opendev.org/p/pain-point-elimination [2] https://etherpad.opendev.org/p/pain-point-elimination#L261 [3] https://etherpad.opendev.org/p/pain-point-elimination#L21 *Rico Lin* OIF Individual Board of directors, OpenStack TC, Multi-arch SIG chair, Heat PTL, Senior Software Engineer at EasyStack On Tue, Dec 7, 2021 at 4:39 PM Rico Lin wrote: > Dear all > > For pain point discussion, > Let's schedule the video meeting for *Dec. 8th, 15:00 UTC time* > Location: https://meet.google.com/xsx-inbw-mos > > *Rico Lin* > OIF Individual Board of directors, OpenStack TC, Multi-arch SIG chair, > Heat PTL, > Senior Software Engineer at EasyStack > > > On Fri, Dec 3, 2021 at 5:37 PM Rico Lin wrote: > >> Dear all >> >> The pain point collection etherpad [1] seems to get good attention from >> teams. >> We plan to host a one-hour video meeting to keep the discussion going. >> Please fill in the doodle to select your preferred time if you can join >> us: >> https://doodle.com/poll/kur2mfywceerqcvh?utm_source=poll&utm_medium=link >> The discussion will be hosted in video format. >> >> Here are some items (but not limited to) I expect we will discuss and >> hopefully put some actions/efforts on. >> >> - General Pain point: RabbitMQ >> - General Pain point: OpenStackClient support >> - General project pain point discussion: How to encourage teams to >> work/reply on pain points. >> >> Meanwhile, we encourage all to check the etherpad and provide your >> feedback. >> Also if teams can help to answer questions/concerns on etherpad (like >> Nova team does) will be greatly appreciated. >> >> [1] https://etherpad.opendev.org/p/pain-point-elimination >> >> *Rico Lin* >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From hberaud at redhat.com Thu Dec 9 07:45:10 2021 From: hberaud at redhat.com (Herve Beraud) Date: Thu, 9 Dec 2021 08:45:10 +0100 Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: <5dd6d28f-9955-7ca5-0ab8-0c5ce11ceb7e@redhat.com> References: <5dd6d28f-9955-7ca5-0ab8-0c5ce11ceb7e@redhat.com> Message-ID: Le mer. 8 d?c. 2021 ? 11:48, Bogdan Dobrelya a ?crit : > Please see inline > > >> I read this with great interest because we are seeing this issue. > Questions: > >> > >> 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. > Should we be upgrading our Train clusters to use 3.8.x? > >> 2. Document [2] recommends policy > '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible > playbooks, nor in any of the config files in the RMQ container. What would > this look like in Ansible, and what should the resulting container config > look like? > >> 3. It appears that we are not setting "amqp_durable_queues = True". > What does this setting look like in Ansible, and what file does it go into? > > > > Note that even having rabbit HA policies adjusted like that and its HA > > replication factor [0] decreased (e.g. to a 2), there still might be > > high churn caused by a large enough number of replicated durable RPC > > topic queues. And that might cripple the cloud down with the incurred > > I/O overhead because a durable queue requires all messages in it to be > > persisted to a disk (for all the messaging cluster replicas) before they > > are ack'ed by the broker. > > > > Given that said, Oslo messaging would likely require a more granular > > control for topic exchanges and the durable queues flag - to tell it to > > declare as durable only the most critical paths of a service. A single > > config setting and a single control exchange per a service might be not > > enough. > > Also note that therefore, amqp_durable_queue=True requires dedicated > control exchanges configured for each service. Those that use > 'openstack' as a default cannot turn the feature ON. Changing it to a > service specific might also cause upgrade impact, as described in the > topic [3]. > > The same is true for `amqp_auto_delete=True`. That requires dedicated control exchanges else it won't work if each service defines its own policy on a shared control exchange (e.g `openstack`) and if policies differ from each other. > [3] https://review.opendev.org/q/topic:scope-config-opts > > > > > There are also race conditions with durable queues enabled, like [1]. A > > solution could be where each service declare its own dedicated control > > exchange with its own configuration. > > > > Finally, openstack components should add perhaps a *.next CI job to test > > it with durable queues, like [2] > > > > [0] https://www.rabbitmq.com/ha.html#replication-factor > > > > [1] > > > https://zuul.opendev.org/t/openstack/build/aa514dd788f34cc1be3800e6d7dba0e8/log/controller/logs/screen-n-cpu.txt > > > > [2] https://review.opendev.org/c/openstack/nova/+/820523 > > > >> > >> Does anyone have a sample set of RMQ config files that they can share? > >> > >> It looks like my Outlook has ruined the link; reposting: > >> [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > > > > > -- > > Best regards, > > Bogdan Dobrelya, > > Irc #bogdando > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando > > > -- Herv? Beraud Senior Software Engineer at Red Hat irc: hberaud https://github.com/4383/ https://twitter.com/4383hberaud -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Thu Dec 9 10:00:05 2021 From: zigo at debian.org (Thomas Goirand) Date: Thu, 9 Dec 2021 11:00:05 +0100 Subject: [tc][all] Follow up pain point discussion In-Reply-To: References: Message-ID: Hi, I missed this thread and the discussion, though I've added to the etherpad: Glance FIX NEEDED - Glance doesn't work well over uwsgi, it's the only service with this state (appart from Swfit). We would need Glance to fully gate using uwsgi in the CI. FIX NEEDED - Glance upload on Ceph is painfully slow. So much that many operators simply don't use the Ceph backend. Swift FIX NEEDED - Swift continues to use Eventlet. Switching to uwsgi increases performances dramatically. Switching to uwsgi is possible for swift-proxy, swift-container and swift-account, but it completely fails for swift-object. Swift needs to fully switch to uwsgi in the CI, and repair the swift-proxy to swift-object protocol to follow *real* http specifictions (at some point, swift-object drifted away from respecting http protocol, which makes it impossible to run it under uwsgi). I hope this helps, Cheers, Thomas Goirand (zigo) On 12/9/21 5:42 AM, Rico Lin wrote: > Dear all > > First, thank?all?who join us in the discussion or help on?the?etherpad [1] > > Here are wrap-ups of our first meeting yesterday: > > In the meeting, we have quick work through pain points for a couple of > projects: Cinder, Ironic, Neutron, QA, Horizon. > And put some action or tags(like `FIX NEEDED` or `INFO NEEDED`). > As for Nova, there?is already have?a good quality of replies and > actions, gmann will help with discussing tags with rest of Nova team. > Horizon?team will go through?the rest item?in?team meeting and finish > tagging after. > Also, TC picks the log level topic?(check out log level discussion in > [2])?as one of the topics in next TC meeting. > So if you're interested, please check out [1] and kindly provide > your?feedback. > > We plan to go through with other projects in our next pain point > meeting, which we expect to schedule around mid. of January. > Meanwhile, if teams can help with providing replies and tags will be > very helpful. > > As for possible cross-project?pain points: > The OSC discussion went very well. The discussion includes?OpenStack > Client and OpenStackSDK. > OSC seems to get more works done than the last time it was proposed as a > community goal (back in Train cycle?), > so TC has set up a topic in the next?TC meeting to discuss the > possibility to set this as a community goal (for z or after-z cycles). > > RabbitMQ was discussed briefly but didn't have any specific suggestions > or actions. So if you have any, please kindly provide your feedback in > this ML or [3]. > > If you would like to suggest other cross-project pain points, kindly > suggest in ML or [1] and help TCs to identify them. > > > Finally, we collect pain points and hope to get things fixed/improved, > so if you're interested to help with targeting any of the items in [1], > you're most welcome to reach out to project teams or at least provide > your feedback in [1].? > > [1]?https://etherpad.opendev.org/p/pain-point-elimination > > [2]?https://etherpad.opendev.org/p/pain-point-elimination#L261 > > [3]?https://etherpad.opendev.org/p/pain-point-elimination#L21 > > > *Rico Lin* > OIF Individual Board of directors, OpenStack TC, Multi-arch SIG chair, > Heat PTL,? > Senior Software Engineer at EasyStack > > > On Tue, Dec 7, 2021 at 4:39 PM Rico Lin > wrote: > > Dear all > > For?pain point discussion, > Let's schedule the video meeting for?*Dec. 8th, 15:00 UTC time* > Location: https://meet.google.com/xsx-inbw-mos > > > *Rico Lin* > OIF Individual Board of directors, OpenStack TC, Multi-arch SIG > chair, Heat PTL,? > Senior Software Engineer at EasyStack > > > On Fri, Dec 3, 2021 at 5:37 PM Rico Lin > wrote: > > Dear all > > The pain point collection etherpad [1] seems to get good > attention from teams. > We plan to host a one-hour video meeting to keep the discussion > going. > Please fill in the doodle to?select your preferred?time if you > can join us: > https://doodle.com/poll/kur2mfywceerqcvh?utm_source=poll&utm_medium=link > > The discussion will be hosted?in video?format. > > Here are some items (but not limited to) I expect we will > discuss and hopefully put some actions/efforts on. > > * General Pain point: RabbitMQ > * General?Pain point:?OpenStackClient support > * General project pain point discussion: How to encourage > teams to work/reply on pain points. > > Meanwhile, we encourage all to check the etherpad and provide > your feedback. > Also if teams can help to answer questions/concerns on etherpad > (like Nova team does) will be greatly appreciated. > > [1] https://etherpad.opendev.org/p/pain-point-elimination > > > *Rico Lin* > From S.Kieske at mittwald.de Thu Dec 9 10:56:41 2021 From: S.Kieske at mittwald.de (Sven Kieske) Date: Thu, 9 Dec 2021 10:56:41 +0000 Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: References: <7d613f7f8f71497db28b5e27ddcb8224@verisign.com> Message-ID: <7ae9d381119c188c5885f3e008b3bb603b8221be.camel@mittwald.de> On Mo, 2021-12-06 at 08:40 +0000, Mark Goddard wrote: > Should we (Kolla) consider adding more of these to our defaults? Or > documenting them in our RabbitMQ guide? definitely +1 on both, this stuff took us a while to figure out! -- Mit freundlichen Gr??en / Regards Sven Kieske Systementwickler / systems engineer ? ? Mittwald CM Service GmbH & Co. KG K?nigsberger Stra?e 4-6 32339 Espelkamp ? Tel.: 05772 / 293-900 Fax: 05772 / 293-333 ? https://www.mittwald.de ? Gesch?ftsf?hrer: Robert Meyer, Florian J?rgens ? St.Nr.: 331/5721/1033, USt-IdNr.: DE814773217, HRA 6640, AG Bad Oeynhausen Komplement?rin: Robert Meyer Verwaltungs GmbH, HRB 13260, AG Bad Oeynhausen Informationen zur Datenverarbeitung im Rahmen unserer Gesch?ftst?tigkeit? gem?? Art. 13-14 DSGVO sind unter www.mittwald.de/ds abrufbar. -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 833 bytes Desc: This is a digitally signed message part URL: From openstack at a.spamming.party Thu Dec 9 15:57:27 2021 From: openstack at a.spamming.party (Jean-Philippe Evrard) Date: Thu, 09 Dec 2021 16:57:27 +0100 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> Message-ID: <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> Hello folks, The current governance requires the projects to follow certain guidelines [1], some of which are already covered by skyline. One of such guidelines is the PTI [2], which is detailed in more details for python in [3]. I am interested by the case of Skyline. Skyline is a truely new project for OpenStack, bringing value to the ecosystem. It is forward looking. Yes, it is different than existing projects. Yes, it doesn't pass the pti for python [3]. If you check the PTI for golang [4], many of those commands could be easily wired on Skyline project using a Makefile (of course it wouldn't pass the dependency management part of that document, for which we require glide [5]... unmaintained since 2019, ahem... [6]). So the question is: Is OpenStack ready to move forward and adapt to new kinds of packaging/software? If yes, let's modify the PTIs (or python PTI) to cover the case for skyline, and allow that team to move forward. It's not like it's doing something very different than some new python projects. Yes, OpenStack is mature... I (simply) hope it's also a place for casual contributions, and where developers can have fun with tech. And yes, it won't please everyone from day one (especially distros), but I hope we can work this out together, rather than discard the project because it doesn't fit the mold. Regards, Jean-Philippe Evrard (evrardjp) [1]: https://governance.openstack.org/tc/reference/new-projects-requirements.html [2]: https://governance.openstack.org/tc/reference/project-testing-interface.html [3]: https://governance.openstack.org/tc/reference/pti/python.html [4]: https://governance.openstack.org/tc/reference/pti/golang.html [5]: https://github.com/Masterminds/glide [6]: https://travis-ci.org/github/Masterminds/glide From mkopec at redhat.com Thu Dec 9 16:23:30 2021 From: mkopec at redhat.com (Martin Kopec) Date: Thu, 9 Dec 2021 17:23:30 +0100 Subject: [all][qa][triplo] Help need to maintain OpenStack health dashboard In-Reply-To: References: <17d76bee71c.cfa2c4c1178950.6715520613968754830@ghanshyammann.com> Message-ID: We're gonna have a call to discuss collaboration between qa and tripleo and the next steps regarding openstack-health. The call details: next Thursday December 16th 1630 UTC Video call link: https://meet.google.com/cmr-yzaq-twp Feel free to join, On Wed, 1 Dec 2021 at 17:52, Ronelle Landy wrote: > > > On Wed, Dec 1, 2021 at 11:11 AM Ghanshyam Mann > wrote: > >> Hello Everyone, >> >> In the QA meeting[1], we discussed the help needed to maintain the >> OpenStack health dashboard which is >> broken currently and the QA team does not have the JS developer and >> bandwidth to fix it. While discussing >> it meeting, we found that the TripleO team might be working on a similar >> dashboard for their CI results. If that >> is the case, can we collaborate on maintaining the existing dashboard? >> > > We had a discussion with Martin Kopec about collaboration with QA in > maintaining a TripleO focused dashboard. > We have two dashboards in progress at the moment - one focused on jobs > running in https://review.rdoproject.org/zuul/status - > http://ci-health-rdo.tripleo.org/ and one where we are starting to track > failures in select jobs running on https://zuul.openstack.org/ - > http://ci-health.tripleo.org/ > > If you would like to collaborate on this work, please ping us on #oooq on > (OFTC) join our community call on Tuesdays at 1:30pm UTC and we can discuss > further. > > Thanks! > >> >> OpenStack health dashboard: >> https://opendev.org/openstack/openstack-health >> Repo: http://status.openstack.org/openstack-health/#/ >> >> [1] >> https://meetings.opendev.org/meetings/qa/2021/qa.2021-11-16-15.00.log.html#l-71 >> >> -gmann >> >> -- Martin Kopec Senior Software Quality Engineer Red Hat EMEA -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Thu Dec 9 17:06:17 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Thu, 09 Dec 2021 11:06:17 -0600 Subject: [tc][all] Follow up pain point discussion In-Reply-To: References: Message-ID: <17da028f2c6.f53fb4e822720.4072888890953590866@ghanshyammann.com> ---- On Thu, 09 Dec 2021 04:00:05 -0600 Thomas Goirand wrote ---- > > Hi, > > I missed this thread and the discussion, though I've added to the etherpad: > > Glance > > FIX NEEDED - Glance doesn't work well over uwsgi, it's the only > service with this state (appart from Swfit). We would need Glance to > fully gate using uwsgi in the CI. We run glance in uwsgi mode by default since wallaby [1] (GLANCE_STANDALONE=False means uwsgi) and from this patch onwards[2], even import workflow is run in uwsgi by default in CI. [1] https://github.com/openstack/devstack/blob/6c849e371384e468679d3d030fe494a36587c505/lib/glance#L85 [2] https://review.opendev.org/c/openstack/devstack/+/742332 -gmann > > FIX NEEDED - Glance upload on Ceph is painfully slow. So much that > many operators simply don't use the Ceph backend. > > Swift > > FIX NEEDED - Swift continues to use Eventlet. Switching to uwsgi > increases performances dramatically. Switching to uwsgi is possible for > swift-proxy, swift-container and swift-account, but it completely fails > for swift-object. Swift needs to fully switch to uwsgi in the CI, and > repair the swift-proxy to swift-object protocol to follow *real* http > specifictions (at some point, swift-object drifted away from respecting > http protocol, which makes it impossible to run it under uwsgi). > > I hope this helps, > Cheers, > > Thomas Goirand (zigo) > > On 12/9/21 5:42 AM, Rico Lin wrote: > > Dear all > > > > First, thank all who join us in the discussion or help on the etherpad [1] > > > > Here are wrap-ups of our first meeting yesterday: > > > > In the meeting, we have quick work through pain points for a couple of > > projects: Cinder, Ironic, Neutron, QA, Horizon. > > And put some action or tags(like `FIX NEEDED` or `INFO NEEDED`). > > As for Nova, there is already have a good quality of replies and > > actions, gmann will help with discussing tags with rest of Nova team. > > Horizon team will go through the rest item in team meeting and finish > > tagging after. > > Also, TC picks the log level topic (check out log level discussion in > > [2]) as one of the topics in next TC meeting. > > So if you're interested, please check out [1] and kindly provide > > your feedback. > > > > We plan to go through with other projects in our next pain point > > meeting, which we expect to schedule around mid. of January. > > Meanwhile, if teams can help with providing replies and tags will be > > very helpful. > > > > As for possible cross-project pain points: > > The OSC discussion went very well. The discussion includes OpenStack > > Client and OpenStackSDK. > > OSC seems to get more works done than the last time it was proposed as a > > community goal (back in Train cycle?), > > so TC has set up a topic in the next TC meeting to discuss the > > possibility to set this as a community goal (for z or after-z cycles). > > > > RabbitMQ was discussed briefly but didn't have any specific suggestions > > or actions. So if you have any, please kindly provide your feedback in > > this ML or [3]. > > > > If you would like to suggest other cross-project pain points, kindly > > suggest in ML or [1] and help TCs to identify them. > > > > > > Finally, we collect pain points and hope to get things fixed/improved, > > so if you're interested to help with targeting any of the items in [1], > > you're most welcome to reach out to project teams or at least provide > > your feedback in [1]. > > > > [1] https://etherpad.opendev.org/p/pain-point-elimination > > > > [2] https://etherpad.opendev.org/p/pain-point-elimination#L261 > > > > [3] https://etherpad.opendev.org/p/pain-point-elimination#L21 > > > > > > *Rico Lin* > > OIF Individual Board of directors, OpenStack TC, Multi-arch SIG chair, > > Heat PTL, > > Senior Software Engineer at EasyStack > > > > > > On Tue, Dec 7, 2021 at 4:39 PM Rico Lin > > wrote: > > > > Dear all > > > > For pain point discussion, > > Let's schedule the video meeting for *Dec. 8th, 15:00 UTC time* > > Location: https://meet.google.com/xsx-inbw-mos > > > > > > *Rico Lin* > > OIF Individual Board of directors, OpenStack TC, Multi-arch SIG > > chair, Heat PTL, > > Senior Software Engineer at EasyStack > > > > > > On Fri, Dec 3, 2021 at 5:37 PM Rico Lin > > wrote: > > > > Dear all > > > > The pain point collection etherpad [1] seems to get good > > attention from teams. > > We plan to host a one-hour video meeting to keep the discussion > > going. > > Please fill in the doodle to select your preferred time if you > > can join us: > > https://doodle.com/poll/kur2mfywceerqcvh?utm_source=poll&utm_medium=link > > > > The discussion will be hosted in video format. > > > > Here are some items (but not limited to) I expect we will > > discuss and hopefully put some actions/efforts on. > > > > * General Pain point: RabbitMQ > > * General Pain point: OpenStackClient support > > * General project pain point discussion: How to encourage > > teams to work/reply on pain points. > > > > Meanwhile, we encourage all to check the etherpad and provide > > your feedback. > > Also if teams can help to answer questions/concerns on etherpad > > (like Nova team does) will be greatly appreciated. > > > > [1] https://etherpad.opendev.org/p/pain-point-elimination > > > > > > *Rico Lin* > > > > > From smooney at redhat.com Thu Dec 9 17:21:55 2021 From: smooney at redhat.com (Sean Mooney) Date: Thu, 09 Dec 2021 17:21:55 +0000 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> Message-ID: <9e6f9cfca9137c1531241947fdc585a0decc6259.camel@redhat.com> On Thu, 2021-12-09 at 16:57 +0100, Jean-Philippe Evrard wrote: > Hello folks, > > The current governance requires the projects to follow certain guidelines [1], some of which are already covered by skyline. One of such guidelines is the PTI [2], which is detailed in more details for python in [3]. > > I am interested by the case of Skyline. > Skyline is a truely new project for OpenStack, bringing value to the ecosystem. It is forward looking. > Yes, it is different than existing projects. > Yes, it doesn't pass the pti for python [3]. > > If you check the PTI for golang [4], many of those commands could be easily wired on Skyline project using a Makefile (of course it wouldn't pass the dependency management part of that document, for which we require glide [5]... unmaintained since 2019, ahem... [6]). > > So the question is: Is OpenStack ready to move forward and adapt to new kinds of packaging/software? If yes, let's modify the PTIs (or python PTI) to cover the case for skyline, and allow that team to move forward. > It's not like it's doing something very different than some new python projects. using a make file to me woudl be alarge regression in packaging and it does not align with the upstream recomendation for package form the python packaging athority. i support using projects.toml and the new standardised build system and i woudl be open to changeing the pti to allow that but? i dont think tha tallow make files would be a good thing. the use of poetry is also an interesting choice that has some advandtges but it will break much of our tooling with regards to upper-constraints.txt. i woudl personally find it rather suprising as a downstream packageer or end user to consume skyline as currently wtieen as either a standar python proejct or an openstack one if i had to use make to build or install it. i am happy to see that tox is used but since its callign dirctly into make i really dont think it being used the way it is intended to be. > > Yes, OpenStack is mature... I (simply) hope it's also a place for casual contributions, and where developers can have fun with tech. And yes, it won't please everyone from day one (especially distros), but I hope we can work this out together, rather than discard the project because it doesn't fit the mold. > is there a reaosbon bveyond its fun to do it this way.? to me this add barries to contibutors using the proejct and to packagees consuming it. > Regards, > Jean-Philippe Evrard (evrardjp) > > [1]: https://governance.openstack.org/tc/reference/new-projects-requirements.html > [2]: https://governance.openstack.org/tc/reference/project-testing-interface.html > [3]: https://governance.openstack.org/tc/reference/pti/python.html > [4]: https://governance.openstack.org/tc/reference/pti/golang.html > [5]: https://github.com/Masterminds/glide > [6]: https://travis-ci.org/github/Masterminds/glide > From akanevsk at redhat.com Thu Dec 9 17:43:26 2021 From: akanevsk at redhat.com (Arkady Kanevsky) Date: Thu, 9 Dec 2021 11:43:26 -0600 Subject: [Interop] No WG meeting on Dec 23 and 30. Message-ID: Happy holidays. -- Arkady Kanevsky, Ph.D. Phone: 972 707-6456 Corporate Phone: 919 729-5744 ext. 8176456 -------------- next part -------------- An HTML attachment was scrubbed... URL: From fungi at yuggoth.org Thu Dec 9 18:03:10 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Thu, 9 Dec 2021 18:03:10 +0000 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> Message-ID: <20211209180309.qnqfd3dflvj4uqul@yuggoth.org> On 2021-12-09 16:57:27 +0100 (+0100), Jean-Philippe Evrard wrote: [...] > So the question is: Is OpenStack ready to move forward and adapt > to new kinds of packaging/software? If yes, let's modify the PTIs > (or python PTI) to cover the case for skyline, and allow that team > to move forward. It's not like it's doing something very different > than some new python projects. > > Yes, OpenStack is mature... I (simply) hope it's also a place for > casual contributions, and where developers can have fun with tech. > And yes, it won't please everyone from day one (especially > distros), but I hope we can work this out together, rather than > discard the project because it doesn't fit the mold. [...] The reason I raised those points on the review was in order to highlight inconsistencies with OpenStack as a whole. If OpenStack is defined by anything, it's consistent development workflows and automation, which allow small groups of people to take advantage of economies of scale in order to minimize maintenance over time across the whole of the set of projects which comprise it. Inconsistency has an inherent cost which the entire community bears, from tracking integration failures in development to keeping concise recipes for packaging in distributions to operators being able to configure all services the same way and expect the same logging formats. Most of the benefits a project gets from being "part of OpenStack" are derived directly from the established standards and norms which enable OpenStack to be treated as a cohesive whole rather than a mere "bucket of parts" with tooling choices made independently by their respective development teams. Deviation from these expectations should be deliberate and come with a clear rationale which benefits OpenStack as a whole rather than one project alone. Let's take the current example point by point: * Skyline implements its own routines for things like configuration and logging rather than using Oslo libraries. Should all OpenStack projects "move forward" to this model, choose their own configuration and logging formats, implement their own solutions for documenting these, and so on? Does inconsistency here benefit operators in some way? * Skyline defines the interdependencies between its components with a mix of "monorepo" (multiple packages from a single repository) and Git submodule references (pinning specific commit ids as virtual subtrees of another repository), rather than establishing a loose coupling with explicit version numbers and package relationships. This choice prevents them from taking advantage of features in Zuul like cross-project change dependencies. Is this a pattern the other OpenStack projects should follow? Does it benefit distribution package maintainers to need to track fundamentally different repository layouts and build processes for each of them? * Skyline performs its release versioning by committing the version strings to files in its repositories rather than assigning them at build time from Git tag metadata. This leads to ambiguity as to what exact state of the repository represents that version, as well as opening up the possibility of forgetting to merge the version update before tagging (a very common problem which drove OpenStack to rely on its current model). Wholly different tools are also needed in order to track versions for these projects as a result. Should the rest of OpenStack's projects follow suit there? Has something changed to make the idea of using Git to signal releases a bad one? * Skyline uses makefiles to script its build and test workflows rather than tox, adding a tox wrapper around the make targets in order to nominally comply with the OpenStack PTI. As a result, it's unable to rely on most of tox's features for environment isolation, package management, path overrides, et cetera. OpenStack projects have solved a lot of cross-project development workflow challenges through applying consistent changes across the tox.ini files of repositories, should those solutions be abandoned and recreated for make instead? * Skyline does not put its external Python package dependencies in central lists nor use tools which would allow it to rely on OpenStack's global constraints implementation, making it very hard for them to guarantee that their software can run with the same versions of shared dependencies as the rest of OpenStack. Is this unnecessary? Have we given up on the idea that two pieces of OpenStack software can be coinstalled into one environment, or that Linux distributions won't be stuck having to package many different versions of the same software because OpenStack projects don't agree on which versions work for them any longer? To be perfectly clear, I don't oppose accepting Skyline as a project in OpenStack, even in its current state, but I do feel like it's not going to end up released as part of OpenStack unless it can address the challenges I outlined above. Most of these are fairly superficial changes to implement, however, and can be done by simply copying the solutions already used in other OpenStack projects or retroactively applying our cookie-cutters. "Moving OpenStack forward" to be more like Skyline, on the other hand, is bound to be far more work than the Skyline team can handle on its own and take far longer, though I'm not opposed to the idea of discussing which of these differences may make sense for OpenStack to adopt in the future. My deeper concern is that these differences are a clear indication that the software was written outside the OpenStack community, with no original consideration for consistency with OpenStack itself, and so there are probably more significant architectural modifications needed (beyond integrating oslo.config and oslo.log) which I'm not well-placed to spot. Those can addressed in time too, and while Skyline is an official team within OpenStack, but they will still need to be taken care of at some point early in the life of the project either by becoming like the rest of OpenStack or helping the rest of OpenStack become more like Skyline. -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From katonalala at gmail.com Thu Dec 9 19:05:48 2021 From: katonalala at gmail.com (Lajos Katona) Date: Thu, 9 Dec 2021 20:05:48 +0100 Subject: [neutron] Drivers meeting - Friday 10.12.2021 - cancelled Message-ID: Hi Neutron Drivers! Due to the lack of agenda, let's cancel tomorrow's drivers meeting. We have only one RFE from last week: https://bugs.launchpad.net/neutron/+bug/1953170 ([RFE] Unify quota engine API) but we discussed that last week (see [0]). See You on the meeting next week. [0]: https://meetings.opendev.org/meetings/neutron_drivers/2021/neutron_drivers.2021-12-03-14.01.log.html#l-76 Lajos Katona (lajoskatona) -------------- next part -------------- An HTML attachment was scrubbed... URL: From ryan.bannon at gmail.com Thu Dec 9 19:25:07 2021 From: ryan.bannon at gmail.com (Ryan Bannon) Date: Thu, 9 Dec 2021 14:25:07 -0500 Subject: Force specific Region for resource creation Message-ID: Hello all, Question: is there any way to associate a Region to a Project? Would this require custom code/development in OpenStack? (By my understanding, a Project lives outside the concept of a Region, and resources associated with a Project can target any Region.) Tx, Ryan -------------- next part -------------- An HTML attachment was scrubbed... URL: From jaegerandi at gmail.com Thu Dec 9 08:43:23 2021 From: jaegerandi at gmail.com (Andreas Jaeger) Date: Thu, 9 Dec 2021 09:43:23 +0100 Subject: [tc][sig][i18n] What's the status for Xena translation? In-Reply-To: References: Message-ID: [I'm not sure whether this mail goes through since I'm not subscribed to the mailing lists] If anybody needs some guideance on what needs to be done, I'm happy to bring anybody up to speed and guide through the process. Ian has done this in the past as well, so he can do this himself or help others as well. Andreas On Thu, Dec 9, 2021 at 3:14 AM Akihiro Motoki wrote: > > Hi, > > # I include TC and SIG tags to get broader attractions to this topic. > # Also Cc'ed to Ian (i18n SIG char) and Andreas. > > I would like to ask the status of Xena translation in > translate.openstack.org (Zanata). > In past releases, stable-XXX branches were created on Zanata, > but a branch for Xena translation has not been created yet. > > Ajaeger usually prepared these branches as a volunteer (I am really > appreciated) > but AFAIK he no longer works on OpenStack actively. > > Who takes care of the translatio preparation around the release now? > I hope the i18n SIG is not dead. > > Thanks, > Akihiro Motoki (amotoki) -- Andreas Jaeger jaegerandi at gmail.com From oleksandr.mykhalskyi at NetCracker.com Thu Dec 9 17:17:03 2021 From: oleksandr.mykhalskyi at NetCracker.com (Oleksandr Mykhalskyi) Date: Thu, 9 Dec 2021 17:17:03 +0000 Subject: [mistral] How to import openstack actions from mistral-extra Message-ID: <8f1179c97de24f1aa0488e8a0ed86331@NetCracker.com> Hello, In release Pike, I was able to import openstack actions to mistral DB using "mistral-db-manage --config-file /etc/mistral/mistral.conf populate" As I understood, from release Ussuri, all openstack actions were moved to separate mistral-extra library. But it`s unclear, how to import openstack actions to DB now. Command "mistral-db-manage populate" creates only few std* actions. I didn`t find any info in mistral documentation https://docs.openstack.org/mistral/latest/user/index.html about these changes. In https://github.com/openstack/mistral-extra also there is no useful readme. Can you please clarify, how to work with mistral-extra library and openstack actions in latest openstack releases? Thank you! Oleksandr Mykhalskyi, System Engineer, IT Servers and Services, LAUT Netcracker Technology +38 (044) 238-8727 ext 6404 | office ________________________________ The information transmitted herein is intended only for the person or entity to which it is addressed and may contain confidential, proprietary and/or privileged material. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon, this information by persons or entities other than the intended recipient is prohibited. If you received this in error, please contact the sender and delete the material from any computer. -------------- next part -------------- An HTML attachment was scrubbed... URL: From ozzzo at yahoo.com Thu Dec 9 21:40:48 2021 From: ozzzo at yahoo.com (Albert Braden) Date: Thu, 9 Dec 2021 21:40:48 +0000 (UTC) Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: <14084e87df22458caa7668eea8b496b6@verisign.com> References: <5dd6d28f-9955-7ca5-0ab8-0c5ce11ceb7e@redhat.com> <14084e87df22458caa7668eea8b496b6@verisign.com> Message-ID: <1147779219.1274196.1639086048233@mail.yahoo.com> Replying from my home email because I've been asked to not email the list from my work email anymore, until I get permission from upper management. I'm not sure I follow. I was planning to add 2 lines to etc/kolla/config/global.conf: [oslo_messaging_rabbit] amqp_durable_queues = False Is that not sufficient? What is involved in configuring dedicated control exchanges for each service? What would that look like in the config? From: Herve Beraud Sent: Thursday, December 9, 2021 2:45 AM To: Bogdan Dobrelya Cc: openstack-discuss at lists.openstack.org Subject: [EXTERNAL] Re: [ops] [kolla] RabbitMQ High Availability ? Caution:?This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe.? ? ? Le?mer. 8 d?c. 2021 ??11:48, Bogdan Dobrelya a ?crit?: Please see inline >> I read this with great interest because we are seeing this issue. Questions: >> >> 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? >> 2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? >> 3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? > > Note that even having rabbit HA policies adjusted like that and its HA > replication factor [0] decreased (e.g. to a 2), there still might be > high churn caused by a large enough number of replicated durable RPC > topic queues. And that might cripple the cloud down with the incurred > I/O overhead because a durable queue requires all messages in it to be > persisted to a disk (for all the messaging cluster replicas) before they > are ack'ed by the broker. > > Given that said, Oslo messaging would likely require a more granular > control for topic exchanges and the durable queues flag - to tell it to > declare as durable only the most critical paths of a service. A single > config setting and a single control exchange per a service might be not > enough. Also note that therefore, amqp_durable_queue=True requires dedicated control exchanges configured for each service. Those that use 'openstack' as a default cannot turn the feature ON. Changing it to a service specific might also cause upgrade impact, as described in the topic [3]. ? The same is true for `amqp_auto_delete=True`. That requires dedicated control exchanges else it won't work if each service defines its own policy on a shared control exchange (e.g `openstack`) and if policies differ from each other. ? [3] https://review.opendev.org/q/topic:scope-config-opts > > There are also race conditions with durable queues enabled, like [1]. A > solution could be where each service declare its own dedicated control > exchange with its own configuration. > > Finally, openstack components should add perhaps a *.next CI job to test > it with durable queues, like [2] > > [0] https://www.rabbitmq.com/ha.html#replication-factor > > [1] > https://zuul.opendev.org/t/openstack/build/aa514dd788f34cc1be3800e6d7dba0e8/log/controller/logs/screen-n-cpu.txt > > [2] https://review.opendev.org/c/openstack/nova/+/820523 > >> >> Does anyone have a sample set of RMQ config files that they can share? >> >> It looks like my Outlook has ruined the link; reposting: >> [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando -- Best regards, Bogdan Dobrelya, Irc #bogdando -- Herv? Beraud Senior Software Engineer at Red Hat irc: hberaud https://github.com/4383/ https://twitter.com/4383hberaud -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Fri Dec 10 00:18:47 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Thu, 09 Dec 2021 18:18:47 -0600 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> Message-ID: <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> ---- On Thu, 09 Dec 2021 09:57:27 -0600 Jean-Philippe Evrard wrote ---- > Hello folks, > > The current governance requires the projects to follow certain guidelines [1], some of which are already covered by skyline. One of such guidelines is the PTI [2], which is detailed in more details for python in [3]. > > I am interested by the case of Skyline. > Skyline is a truely new project for OpenStack, bringing value to the ecosystem. It is forward looking. > Yes, it is different than existing projects. > Yes, it doesn't pass the pti for python [3]. > > If you check the PTI for golang [4], many of those commands could be easily wired on Skyline project using a Makefile (of course it wouldn't pass the dependency management part of that document, for which we require glide [5]... unmaintained since 2019, ahem... [6]). > > So the question is: Is OpenStack ready to move forward and adapt to new kinds of packaging/software? If yes, let's modify the PTIs (or python PTI) to cover the case for skyline, and allow that team to move forward. > It's not like it's doing something very different than some new python projects. > > Yes, OpenStack is mature... I (simply) hope it's also a place for casual contributions, and where developers can have fun with tech. And yes, it won't please everyone from day one (especially distros), but I hope we can work this out together, rather than discard the project because it doesn't fit the mold. Yes, those are the inconsistency we currently have in the Skyline project, but that is not what they will stick to. In Yoga PTG[1], the skyline team iterates through all these points and plans to improve those and be consistent with the other OpenStack projects. They said they will work on packaging, PTI, using Oslo code etc. Boxiang Zhu mentioned the same in the governance patch also[2]. If they deny moving Skyline towards the OpenStack way of packing, testing, using Oslo, etc., we should consider whether to reject the OpenStack official project or modify our new project application criteria. But this is not the case, and they agree to improve all those things. Now we have more clear direction from them in the governance patch also, so I feel this is good to go as an official project, and while they will be in OpenStack, we all can help them to improve the things in OpenStack way. Also, TC discussed in PTG the new concept of 'Tech pre-review' where we will monitor such new projects' health, contribution, and direction going in the direction of OpenStack way. This 'Tech pre-review' is not up yet, so let's see how we define it. [1] http://lists.openstack.org/pipermail/openstack-discuss/2021-October/025541.html [2] https://review.opendev.org/c/openstack/governance/+/814037 [3] https://etherpad.opendev.org/p/tc-yoga-ptg#L263 -gmann > > Regards, > Jean-Philippe Evrard (evrardjp) > > [1]: https://governance.openstack.org/tc/reference/new-projects-requirements.html > [2]: https://governance.openstack.org/tc/reference/project-testing-interface.html > [3]: https://governance.openstack.org/tc/reference/pti/python.html > [4]: https://governance.openstack.org/tc/reference/pti/golang.html > [5]: https://github.com/Masterminds/glide > [6]: https://travis-ci.org/github/Masterminds/glide > > From hberaud at redhat.com Fri Dec 10 07:55:36 2021 From: hberaud at redhat.com (Herve Beraud) Date: Fri, 10 Dec 2021 08:55:36 +0100 Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: <1147779219.1274196.1639086048233@mail.yahoo.com> References: <5dd6d28f-9955-7ca5-0ab8-0c5ce11ceb7e@redhat.com> <14084e87df22458caa7668eea8b496b6@verisign.com> <1147779219.1274196.1639086048233@mail.yahoo.com> Message-ID: If you plan to let `amqp_durable_queues = False` (i.e if you plan to keep this config equal to false), then you don't need to add these config lines as this is already the default value [1]. [1] https://opendev.org/openstack/oslo.messaging/src/branch/master/oslo_messaging/_drivers/amqp.py#L34 Le jeu. 9 d?c. 2021 ? 22:40, Albert Braden a ?crit : > Replying from my home email because I've been asked to not email the list > from my work email anymore, until I get permission from upper management. > > I'm not sure I follow. I was planning to add 2 lines to > etc/kolla/config/global.conf: > > [oslo_messaging_rabbit] > amqp_durable_queues = False > > Is that not sufficient? What is involved in configuring dedicated control > exchanges for each service? What would that look like in the config? > > > From: Herve Beraud > Sent: Thursday, December 9, 2021 2:45 AM > To: Bogdan Dobrelya > Cc: openstack-discuss at lists.openstack.org > Subject: [EXTERNAL] Re: [ops] [kolla] RabbitMQ High Availability > > > > Caution: This email originated from outside the organization. Do not click > links or open attachments unless you recognize the sender and know the > content is safe. > > > > > > Le mer. 8 d?c. 2021 ? 11:48, Bogdan Dobrelya a > ?crit : > > Please see inline > > >> I read this with great interest because we are seeing this issue. > Questions: > >> > >> 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. > Should we be upgrading our Train clusters to use 3.8.x? > >> 2. Document [2] recommends policy > '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible > playbooks, nor in any of the config files in the RMQ container. What would > this look like in Ansible, and what should the resulting container config > look like? > >> 3. It appears that we are not setting "amqp_durable_queues = True". > What does this setting look like in Ansible, and what file does it go into? > > > > Note that even having rabbit HA policies adjusted like that and its HA > > replication factor [0] decreased (e.g. to a 2), there still might be > > high churn caused by a large enough number of replicated durable RPC > > topic queues. And that might cripple the cloud down with the incurred > > I/O overhead because a durable queue requires all messages in it to be > > persisted to a disk (for all the messaging cluster replicas) before they > > are ack'ed by the broker. > > > > Given that said, Oslo messaging would likely require a more granular > > control for topic exchanges and the durable queues flag - to tell it to > > declare as durable only the most critical paths of a service. A single > > config setting and a single control exchange per a service might be not > > enough. > > Also note that therefore, amqp_durable_queue=True requires dedicated > control exchanges configured for each service. Those that use > 'openstack' as a default cannot turn the feature ON. Changing it to a > service specific might also cause upgrade impact, as described in the > topic [3]. > > > > The same is true for `amqp_auto_delete=True`. That requires dedicated > control exchanges else it won't work if each service defines its own policy > on a shared control exchange (e.g `openstack`) and if policies differ from > each other. > > > > [3] https://review.opendev.org/q/topic:scope-config-opts > > > > > There are also race conditions with durable queues enabled, like [1]. A > > solution could be where each service declare its own dedicated control > > exchange with its own configuration. > > > > Finally, openstack components should add perhaps a *.next CI job to test > > it with durable queues, like [2] > > > > [0] https://www.rabbitmq.com/ha.html#replication-factor > > > > [1] > > > https://zuul.opendev.org/t/openstack/build/aa514dd788f34cc1be3800e6d7dba0e8/log/controller/logs/screen-n-cpu.txt > > > > [2] https://review.opendev.org/c/openstack/nova/+/820523 > > > >> > >> Does anyone have a sample set of RMQ config files that they can share? > >> > >> It looks like my Outlook has ruined the link; reposting: > >> [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > > > > > -- > > Best regards, > > Bogdan Dobrelya, > > Irc #bogdando > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando > > > > > -- > > Herv? Beraud > > Senior Software Engineer at Red Hat > > irc: hberaud > > https://github.com/4383/ > > https://twitter.com/4383hberaud > -- Herv? Beraud Senior Software Engineer at Red Hat irc: hberaud https://github.com/4383/ https://twitter.com/4383hberaud -------------- next part -------------- An HTML attachment was scrubbed... URL: From openstack at a.spamming.party Fri Dec 10 08:16:59 2021 From: openstack at a.spamming.party (Jean-Philippe Evrard) Date: Fri, 10 Dec 2021 09:16:59 +0100 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> Message-ID: Hello, On Fri, Dec 10, 2021, at 01:18, Ghanshyam Mann wrote: > Yes, those are the inconsistency we currently have in the Skyline > project, but that is not what they will stick to. In Yoga PTG[1], the > skyline team iterates through all these points and plans to > improve those and be consistent with the other OpenStack projects. They > said they will > work on packaging, PTI, using Oslo code etc. > > Boxiang Zhu mentioned the same in the governance patch also[2]. The questions I have still stand: Do we need to be more lax in the governance to not _require_ to follow the PTI (and/or have an intention to follow through)? Can the TC make an exception? When is the deadline for projects not following our testing to become compliant? What is the risk if the newly accepted project is not doing it? Moving it to a non-official project? Is that worth the effort? > If they deny moving Skyline towards the OpenStack way of packing, > testing, using Oslo, etc., we should consider whether to reject the > OpenStack official project or modify our new > project application criteria. But this is not the case, and they agree > to improve all those things. I agree. Let's not talk hypothetics. > Now we have more clear direction from them in the governance patch > also, so I feel this is good > to go as an official project, and while they will be in OpenStack, we > all can help them to > improve the things in OpenStack way. Perfect, great news. > > Also, TC discussed in PTG the new concept of 'Tech pre-review' where we > will monitor > such new projects' health, contribution, and direction going in the > direction of OpenStack way. This > 'Tech pre-review' is not up yet, so let's see how we define it. I am sorry I have missed that from the notes I read. > > [1] > http://lists.openstack.org/pipermail/openstack-discuss/2021-October/025541.html > [2] https://review.opendev.org/c/openstack/governance/+/814037 > [3] https://etherpad.opendev.org/p/tc-yoga-ptg#L263 > > -gmann > > > > > Regards, > > Jean-Philippe Evrard (evrardjp) > > > > [1]: > https://governance.openstack.org/tc/reference/new-projects-requirements.html > > [2]: > https://governance.openstack.org/tc/reference/project-testing-interface.html > > [3]: https://governance.openstack.org/tc/reference/pti/python.html > > [4]: https://governance.openstack.org/tc/reference/pti/golang.html > > [5]: https://github.com/Masterminds/glide > > [6]: https://travis-ci.org/github/Masterminds/glide > > > > From openstack at a.spamming.party Fri Dec 10 09:43:10 2021 From: openstack at a.spamming.party (Jean-Philippe Evrard) Date: Fri, 10 Dec 2021 10:43:10 +0100 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <20211209180309.qnqfd3dflvj4uqul@yuggoth.org> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <20211209180309.qnqfd3dflvj4uqul@yuggoth.org> Message-ID: Helllo, On Thu, Dec 9, 2021, at 19:03, Jeremy Stanley wrote: > The reason I raised those points on the review was in order to > highlight inconsistencies with OpenStack as a whole. If OpenStack is > defined by anything, it's consistent development workflows and > automation, which allow small groups of people to take advantage of > economies of scale in order to minimize maintenance over time across > the whole of the set of projects which comprise it. Inconsistency > has an inherent cost which the entire community bears, from tracking > integration failures in development to keeping concise recipes for > packaging in distributions to operators being able to configure all > services the same way and expect the same logging formats. I agree that consistency is important, and helps us in many ways. I don't think it's the "defining" factor of OpenStack, however. Inconsistency in openstack is in many places, from the users' to the developer's perspective. It doesn't mean I am _looking for_ inconsistency. I am merely embracing the differences (and I am eager to remove them, if we can). > Most of the benefits a project gets from being "part of OpenStack" > are derived directly from the established standards and norms which > enable OpenStack to be treated as a cohesive whole rather than a > mere "bucket of parts" with tooling choices made independently by > their respective development teams. Deviation from these > expectations should be deliberate and come with a clear rationale > which benefits OpenStack as a whole rather than one project alone. Correct again! :) For the case of skyline, one could made a case to start building packages with a new way, using poetry, rather than setuptools. It could mean we need to ask the skyline team to help on improving the infrastructure and testing. Instead of arriving to a standard testing, we embrace the difference and provide a possible new way of working. We could have two kinds of python projects. Instead of accepting that difference when a single project asking for it, maybe we can just tally who wanted this, and the TC decides at the end "when" it is worth doing. My whole point is that we need to document better and agree which parts benefits OpenStack as a whole. We should also have indicators for checking how the "ecosystem" is behaving. If we are considering the amount of new teams in OpenStack, it's not very encouraging. (It might not be a metric, I don't know, I rather leave Thierry judge there). In other words: Is it really better to change the skyline team to adapt our mold, or should we have multiple molds? We agreed to go the way that skyline should adapt. I am simply raising the fact that some other people/projects won't do the effort to adapt to existing mold, because they won't necessarily see the value. (I won't prevent my teams to work on non-openstack projects ;)) > Let's take the current example point by point: > > * Skyline implements its own routines for things like > configuration and logging rather than using Oslo libraries. > Should all OpenStack projects "move forward" to this model, > choose their own configuration and logging formats, implement > their own solutions for documenting these, and so on? Does > inconsistency here benefit operators in some way? Is Oslo a requirement to be an OpenStack project now? AFAIK it wasn't (see swift). It doesn't mean you can't be consistent for logs/configuration from the user perspective by not using oslo. Side note: Bringing oslo projects for configuration and logging isn't enough to bring consistency from the user perspective. > * Skyline defines the interdependencies between its components > with a mix of "monorepo" (multiple packages from a single > repository) and Git submodule references (pinning specific > commit ids as virtual subtrees of another repository), rather > than establishing a loose coupling with explicit version numbers > and package relationships. This choice prevents them from taking > advantage of features in Zuul like cross-project change > dependencies. Is this a pattern the other OpenStack projects > should follow? Which is why I am highlighting that _as of today_, the governance requires certain expectations, which needs clarifications (or exceptions) if we want to integrate Skyline. I know I am a bit pedantic in the following question: Is (nowadays) the expectation for projects to be testable by Zuul (+release-able by zuul, ... list to be defined) or is the expectation for projects to be following the PTI guidelines (tested by zuul)? In other words: Do we need to enforce the tooling or the results (a frequently released and well tested software)? I would like the TC to think about this. > Does it benefit distribution package maintainers > to need to track fundamentally different repository layouts and > build processes for each of them? Isn't that exactly what the distros are doing right now with the packages they are maintaining? Not every software is like openstack. Yes, almost all openstack projects are similar to each other, so they can be packaged the same way. Yet, some official openstack projects aren't, as of today. So, some distros decide to not package them? Is that a problem for us? I didn't think so. Projects/Teams need to understand that if they don't follow "x", they will not have "y". Which sounds fair to me. Keep in mind that we are _already doing exceptions_. Not all projects are similar to nova, glance, neutron, keystone, cinder. (Sorry for the many I missed). > * Skyline performs its release versioning by committing the > version strings to files in its repositories rather than > assigning them at build time from Git tag metadata. This leads > to ambiguity as to what exact state of the repository represents > that version, as well as opening up the possibility of > forgetting to merge the version update before tagging (a very > common problem which drove OpenStack to rely on its current > model). Wholly different tools are also needed in order to > track versions for these projects as a result. Should the rest > of OpenStack's projects follow suit there? Has something changed > to make the idea of using Git to signal releases a bad one? You're raising again a valid point, for which the skyline team will have to find a solution. For me, it boils down to defining that projects should follow a "project release interface", which explains the expected behaviour in regards of releasing for projects. We don't have such. I am not claiming we should, I am merely asking us to think about what we are doing. Projects could indeed move to use pbr/setuptools. However, I am not sure many developers understand why. Should we simply explain what is expected in terms of building the version from git? People might even come with interesting solutions. Promoting creativity will create tech debt, but might also reduce it over time. What if we don't need pbr anymore? Will we even notice? > * Skyline uses makefiles to script its build and test workflows > rather than tox, adding a tox wrapper around the make targets > in order to nominally comply with the OpenStack PTI. As a > result, it's unable to rely on most of tox's features for > environment isolation, package management, path overrides, et > cetera. OpenStack projects have solved a lot of cross-project > development workflow challenges through applying consistent > changes across the tox.ini files of repositories, should those > solutions be abandoned and recreated for make instead? I didn't say so. I am only partially agreeing with you there, for once :) For me, tox is a test runner that provides many nice features thanks to its code (and integration with setuptools) I would prefer if python projects using tox for testing interface indeed. Should that be the standard? I don't know. Does it make sense to rely on tox features for isolations/package management/other? I don't know. You might be aware that it's possible to integrate tox with poetry. Does it provide all we _need_? I don't know: We aren't explicit in the requirements of what we need, we just talk about tools. If I were to rephrase: Is it easier to fit the existing mold or clarify what we need, why, and how we can deal with the tech debt? Probably the former. Should we do it? I don't know. The TC should however evaluate this seriously. > * Skyline does not put its external Python package dependencies in > central lists nor use tools which would allow it to rely on > OpenStack's global constraints implementation, making it very > hard for them to guarantee that their software can run with the > same versions of shared dependencies as the rest of OpenStack. > Is this unnecessary? Have we given up on the idea that two > pieces of OpenStack software can be coinstalled into one > environment, or that Linux distributions won't be stuck having > to package many different versions of the same software because > OpenStack projects don't agree on which versions work for them > any longer? Now, we are hitting the nail, IMO :) This is a bit of the issue's core, isn't it? I think that's what we need to document. While I (personally) don't believe it's important nowadays (cf. containers), I am sure distros will disagree with me. It's important to know who we are pleasing, as we can't please everyone ;) Let's be clear about the audience. If we intend to please the distros, did we ask about RedHat/Canonical committments on skyline? Are they even remotely interested? If they want to use it, wouldn't it make sense they do the work to make sure skyline is packageable for them, by improving upstream? If they don't, why do we care? > To be perfectly clear, I don't oppose accepting Skyline as a project > in OpenStack, even in its current state, but I do feel like it's not > going to end up released as part of OpenStack unless it can address > the challenges I outlined above. Most of these are fairly > superficial changes to implement, however, and can be done by simply > copying the solutions already used in other OpenStack projects or > retroactively applying our cookie-cutters. I agree with you. I am merely asking for clarity. However, I want to think as a whole. > "Moving OpenStack > forward" to be more like Skyline, on the other hand, is bound to be > far more work than the Skyline team can handle on its own and take > far longer, though I'm not opposed to the idea of discussing which > of these differences may make sense for OpenStack to adopt in the > future. Depends on the definition of "Moving OpenStack forward". Do you mean all modifying all projects to be like skyline + integrate the testing? That was never my point. Do you mean "ensure a single project can be behaving like a majority of other big openstack projects are"? That's indeed more work for a team, compared to if they had their work done outside OpenStack governance. Should we use this energy for thinking(/doing something) about OpenStack future? IMO, absolutely. And for those arriving here: Thanks for reading my rant until the end ;) Jean-Philippe Evrard (evrardjp) From zigo at debian.org Fri Dec 10 09:50:44 2021 From: zigo at debian.org (Thomas Goirand) Date: Fri, 10 Dec 2021 10:50:44 +0100 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> Message-ID: <0eb643ff-5168-3aa8-6f24-ee21f802d109@debian.org> On 12/10/21 9:16 AM, Jean-Philippe Evrard wrote: > Hello, > > On Fri, Dec 10, 2021, at 01:18, Ghanshyam Mann wrote: >> Yes, those are the inconsistency we currently have in the Skyline >> project, but that is not what they will stick to. In Yoga PTG[1], the >> skyline team iterates through all these points and plans to >> improve those and be consistent with the other OpenStack projects. They >> said they will >> work on packaging, PTI, using Oslo code etc. >> >> Boxiang Zhu mentioned the same in the governance patch also[2]. > > The questions I have still stand: Do we need to be more lax in the governance to not _require_ to follow the PTI (and/or have an intention to follow through)? Can the TC make an exception? IMO, we shouldn't allow an exception. Otherwise, it's going to become super hard to contribute to OpenStack if every project is different. It is a pandora box that should not be opened. That being said, I don't think poetry is a problem for Debian anymore: it's been packaged already. :) Cheers, Thomas Goirand (zigo) From smooney at redhat.com Fri Dec 10 11:27:16 2021 From: smooney at redhat.com (Sean Mooney) Date: Fri, 10 Dec 2021 11:27:16 +0000 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <0eb643ff-5168-3aa8-6f24-ee21f802d109@debian.org> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> <0eb643ff-5168-3aa8-6f24-ee21f802d109@debian.org> Message-ID: On Fri, 2021-12-10 at 10:50 +0100, Thomas Goirand wrote: > On 12/10/21 9:16 AM, Jean-Philippe Evrard wrote: > > Hello, > > > > On Fri, Dec 10, 2021, at 01:18, Ghanshyam Mann wrote: > > > Yes, those are the inconsistency we currently have in the Skyline > > > project, but that is not what they will stick to. In Yoga PTG[1], the > > > skyline team iterates through all these points and plans to > > > improve those and be consistent with the other OpenStack projects. They > > > said they will > > > work on packaging, PTI, using Oslo code etc. > > > > > > Boxiang Zhu mentioned the same in the governance patch also[2]. > > > > The questions I have still stand: Do we need to be more lax in the governance to not _require_ to follow the PTI (and/or have an intention to follow through)? Can the TC make an exception? > > IMO, we shouldn't allow an exception. Otherwise, it's going to become > super hard to contribute to OpenStack if every project is different. It > is a pandora box that should not be opened. i would agree with this we should not make an excption however we could work on evolving the PTI over time. > > That being said, I don't think poetry is a problem for Debian anymore: > it's been packaged already. :) poetry i see as less of a porblem then using make honestly. as someone who has avoided using make for most of the last decade moderatly succesfully that is perhaps the most objectionable part of its build system. one concern i would have with using poetry is just makeing sure that upper-constratints.txt is used correctly based on https://github.com/python-poetry/poetry/pull/4005 and https://github.com/python-poetry/poetry/issues/3225 i dont think that poetry will ever actuly support upper-constarits so we would need to have some tooling to generate the equivalent for there format. i still think coinstallablity is an important requirement for being an offical python porject and moving to poetry or other packaging may not support that. i recently have been workign on a new python project which i was considering proposing to opendev/openstack at some point in the future i used https://pyscaffold.org/ to generate the skeloton of the project and do have a?pyproject.toml https://github.com/SeanMooney/arbiterd/blob/master/pyproject.toml but have intentioally choose to continue to use setuptools so that it would be simple to support the pti in the future i think the main delta i have right now is i dont have my deps in requirement.txt but i can simply generate that in the future https://github.com/SeanMooney/arbiterd/blob/master/setup.cfg#L49-L77 that is just because pyscaffold put them in setup.cfg and i havent bothered to move them. i likely would have to add pbr for ChangeLog and AUTHORS generation and some minor other tweaks but you can have a pretty moddern an painless setup and be pti compliant at the same time. > > Cheers, > > Thomas Goirand (zigo) > From zigo at debian.org Fri Dec 10 11:30:34 2021 From: zigo at debian.org (Thomas Goirand) Date: Fri, 10 Dec 2021 12:30:34 +0100 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <20211209180309.qnqfd3dflvj4uqul@yuggoth.org> Message-ID: Hi Jean-Philippe. Please read what I'm writing keeping in mind that I'm *very* interested in Skyline, and would love to have it fully packaged in Debian. I probably will do that when I consider the project ready for it, which unfortunately, in the current case, isn't the case (ie: it would need too much effort from my side for it). So I'm writing in the hope to get things improved... :) On 12/10/21 10:43 AM, Jean-Philippe Evrard wrote: > For the case of skyline, one could made a case to start building packages with a new way, using poetry, rather than setuptools. It could mean we need to ask the skyline team to help on improving the infrastructure and testing. >From my package maintainer point of view, the main problem I see is more then way things are layout (ie: the libs folder of skyline-apiserver that contains many projects). I don't really mind anymore the poetry thing. This shouldn't be so hard to address, right? > Instead of arriving to a standard testing, we embrace the difference and provide a possible new way of working. We could have two kinds of python projects. Instead of accepting that difference when a single project asking for it, maybe we can just tally who wanted this, and the TC decides at the end "when" it is worth doing. I don't think you understand the problem here. The issue is wiring Skyline to the OpenDev CI. With the way things are done, it's simply very hard to do, unless you're ready to spend *a lot* of time with the infra team, and if they have enough time to let you do that (probably they are already very busy). > In other words: Is it really better to change the skyline team to adapt our mold, or should we have multiple molds? > We agreed to go the way that skyline should adapt. I am simply raising the fact that some other people/projects won't do the effort to adapt to existing mold, because they won't necessarily see the value. > (I won't prevent my teams to work on non-openstack projects ;)) Same story: if the infra can handle the "new way" that you're proposing, why not. But we're not there yet at all. >> Let's take the current example point by point: >> >> * Skyline implements its own routines for things like >> configuration and logging rather than using Oslo libraries. >> Should all OpenStack projects "move forward" to this model, >> choose their own configuration and logging formats, implement >> their own solutions for documenting these, and so on? Does >> inconsistency here benefit operators in some way? > > Is Oslo a requirement to be an OpenStack project now? > AFAIK it wasn't (see swift). You wont make your case by pointing at the defects of other projects. You could also have pointed at Horizon's local_setting.py. IMO it's a much better example beacuse it's a very good example: it's a *REAL PAIN* for operators (for example: for puppet, that has no way of handling that configuration file using something else than a ERB template). > Side note: Bringing oslo projects for configuration and logging isn't enough to bring consistency from the user perspective. Probably, but it's a starting point. >> * Skyline defines the interdependencies between its components >> with a mix of "monorepo" (multiple packages from a single >> repository) and Git submodule references (pinning specific >> commit ids as virtual subtrees of another repository), rather >> than establishing a loose coupling with explicit version numbers >> and package relationships. This choice prevents them from taking >> advantage of features in Zuul like cross-project change >> dependencies. Is this a pattern the other OpenStack projects >> should follow? > > Which is why I am highlighting that _as of today_, the governance requires certain expectations, which needs clarifications (or exceptions) if we want to integrate Skyline. > > I know I am a bit pedantic in the following question: > Is (nowadays) the expectation for projects to be testable by Zuul (+release-able by zuul, ... list to be defined) or is the expectation for projects to be following the PTI guidelines (tested by zuul)? I'm not from the TC, but ... hopefully BOTH ! :) >> Does it benefit distribution package maintainers >> to need to track fundamentally different repository layouts and >> build processes for each of them? > > Isn't that exactly what the distros are doing right now with the packages they are maintaining? Not every software is like openstack. Yes, almost all openstack projects are similar to each other, so they can be packaged the same way. > > Yet, some official openstack projects aren't, as of today. So, some distros decide to not package them? Could you please be more specific? Maybe with some examples? I'm having a hard time trying to figure out what you have in mind here... :) > Keep in mind that we are _already doing exceptions_. Not all projects are similar to nova, glance, neutron, keystone, cinder. (Sorry for the many I missed). Which one are you thinking about? :) >> * Skyline performs its release versioning by committing the >> version strings to files in its repositories rather than >> assigning them at build time from Git tag metadata. This leads >> to ambiguity as to what exact state of the repository represents >> that version, as well as opening up the possibility of >> forgetting to merge the version update before tagging (a very >> common problem which drove OpenStack to rely on its current >> model). Wholly different tools are also needed in order to >> track versions for these projects as a result. Should the rest >> of OpenStack's projects follow suit there? Has something changed >> to make the idea of using Git to signal releases a bad one? > > You're raising again a valid point, for which the skyline team will have to find a solution. >From a distro perspective, it is easy to build multiple binary packages out of a single source tree. However, it's not possible to produce different versions of the binaries, if we have a single source package. So IMO, this should be addressed as a top priority. > For me, it boils down to defining that projects should follow a "project release interface", which explains the expected behaviour in regards of releasing for projects. We don't have such. I am not claiming we should, I am merely asking us to think about what we are doing. Yes, you should think about it! :) Would it be a huge effort to separte things into multiple projects instead of one big repo, and then start tagging properly? > Projects could indeed move to use pbr/setuptools. However, I am not sure many developers understand why. > Should we simply explain what is expected in terms of building the version from git? People might even come with interesting solutions. Promoting creativity will create tech debt, but might also reduce it over time. > What if we don't need pbr anymore? Will we even notice? >From my perspective, pbr/setuptools isn't the main problem. Single repo + not tagging is. > For me, tox is a test runner that provides many nice features thanks to its code (and integration with setuptools) > I would prefer if python projects using tox for testing interface indeed. > Should that be the standard? I don't know. > Does it make sense to rely on tox features for isolations/package management/other? I don't know. > You might be aware that it's possible to integrate tox with poetry. Does it provide all we _need_? I don't know: We aren't explicit in the requirements of what we need, we just talk about tools. > > If I were to rephrase: Is it easier to fit the existing mold or clarify what we need, why, and how we can deal with the tech debt? Probably the former. Should we do it? I don't know. The TC should however evaluate this seriously. What Jeremy wrote: it's important to be able to use the global requirements repository, so we can check your project works with the others. I don't think the TC will have any satisfying answer if you can't provide a way to check your projects works with the rest of the OpenStack world. >> * Skyline does not put its external Python package dependencies in >> central lists nor use tools which would allow it to rely on >> OpenStack's global constraints implementation, making it very >> hard for them to guarantee that their software can run with the >> same versions of shared dependencies as the rest of OpenStack. >> Is this unnecessary? Have we given up on the idea that two >> pieces of OpenStack software can be coinstalled into one >> environment, or that Linux distributions won't be stuck having >> to package many different versions of the same software because >> OpenStack projects don't agree on which versions work for them >> any longer? > > Now, we are hitting the nail, IMO :) > > This is a bit of the issue's core, isn't it? I think that's what we need to document. > While I (personally) don't believe it's important nowadays (cf. containers), I am sure distros will disagree with me. Indeed, I strongly disagree! :) > It's important to know who we are pleasing, as we can't please everyone ;) > Let's be clear about the audience. > > If we intend to please the distros, did we ask about RedHat/Canonical committments on skyline? Are they even remotely interested? I am *very* interested. :) If you didn't know, I've been packaging OpenStack in Debian since the very beginning (ie: in 2011 for me). > If they want to use it, wouldn't it make sense they do the work to make sure skyline is packageable for them, by improving upstream? Do you understand that *all* the distros are understaffed? Do you really expect package maintainers to address *ALL* of the issues they see in *ALL* of the OpenStack projects? That's not reasonable, and it wont ever happen, simply because we can't. > If they don't, why do we care? Why wouldn't you care about integration? Being package-able also means your project can be integrated in *any* environment, not only distro. If you don't care, why do you then care to have Skyline as an official project then? It doesn't make sense to me. Cheers, Thomas Goirand (zigo) From opentastic at gmail.com Fri Dec 10 13:20:52 2021 From: opentastic at gmail.com (Edward Hope-Morley) Date: Fri, 10 Dec 2021 13:20:52 +0000 Subject: [charms] nominate Hemanth Nakkina for charms-core Message-ID: <3c167485-421c-4861-bbad-6f588b923ee9@gmail.com> I would like propose Hemanth Nakkina for charms-core. Hemanth has really stepped up his efforts to provide invaluable contributions both in code and reviews and I beleive he is in a good position to be considered for core where he will be able to provide additional value to the project. Hemanth has demonstrated rigour for quality through his contributions and is always keen to help others with his deep knowledge of the charms and Openstack. Here is a list of all his contributions: patches: https://review.opendev.org/q/owner:hemanth.nakkina%2540canonical.com reviews: https://review.opendev.org/q/reviewedby:hemanth.nakkina%2540canonical.com I hope you will join me in voting for Hemanth. - Ed From alex.kavanagh at canonical.com Fri Dec 10 13:58:04 2021 From: alex.kavanagh at canonical.com (Alex Kavanagh) Date: Fri, 10 Dec 2021 13:58:04 +0000 Subject: [charms] nominate Hemanth Nakkina for charms-core In-Reply-To: <3c167485-421c-4861-bbad-6f588b923ee9@gmail.com> References: <3c167485-421c-4861-bbad-6f588b923ee9@gmail.com> Message-ID: On Fri, Dec 10, 2021 at 1:24 PM Edward Hope-Morley wrote: > I would like propose Hemanth Nakkina for charms-core. Hemanth has really > stepped up his efforts to provide invaluable contributions both in code > and reviews and I beleive he is in a good position to be considered for > core where he will be able to provide additional value to the project. > Hemanth has demonstrated rigour for quality through his contributions > and is always keen to help others with his deep knowledge of the charms > and Openstack. Here is a list of all his contributions: > > patches: > https://review.opendev.org/q/owner:hemanth.nakkina%2540canonical.com > reviews: > https://review.opendev.org/q/reviewedby:hemanth.nakkina%2540canonical.com > > I hope you will join me in voting for Hemanth. > +1 from me; Hemanth has submitted good patches and thoughtful reviews in my experience. He would be a good member of charms-core. Cheers Alex. > - Ed > > > -- Alex Kavanagh - Software Engineer OpenStack Engineering - Data Centre Development - Canonical Ltd -------------- next part -------------- An HTML attachment was scrubbed... URL: From fungi at yuggoth.org Fri Dec 10 14:11:36 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Fri, 10 Dec 2021 14:11:36 +0000 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: References: <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> <0eb643ff-5168-3aa8-6f24-ee21f802d109@debian.org> Message-ID: <20211210141136.uqnxdvnc4otxplyz@yuggoth.org> On 2021-12-10 11:27:16 +0000 (+0000), Sean Mooney wrote: [...] > i likely would have to add pbr for ChangeLog and AUTHORS > generation and some minor other tweaks but you can have a pretty > moddern an painless setup and be pti compliant at the same time. As an aside, most (if not all?) of the work needed to make PBR a functioning PEP 517 build backend is in 5.7.0, with improvements in master (not yet released) which eliminate use of some deprecated setuptools interfaces. A setup.py is still needed along with your pyproject.toml now, but there's no more calling into the setup.py install interface behind the scenes, and all the build requirements get correctly installed by pip now instead. -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From pierre at stackhpc.com Fri Dec 10 14:25:55 2021 From: pierre at stackhpc.com (Pierre Riteau) Date: Fri, 10 Dec 2021 15:25:55 +0100 Subject: [CLOUDKITTY][VICTORIA] - policy prohibit report:get_summary to admin In-Reply-To: References: Message-ID: Hello Ga?l, This fix will not be included in RDO packages until we issue a new victoria release, which only then can be packaged as RPM. I have proposed release patches for both victoria and wallaby. They need to be approved by Rafael. Using Kolla source images is another option, since in recent versions (including Victoria) they use the tip of stable branches. Best wishes, Pierre On Tue, 23 Nov 2021 at 16:31, Ga?l THEROND wrote: > > ah ah! Was exactly that indeed! > > So, CentOS cloudkitty common package is not using the latest patch fixing the issue -> http://mirror.centos.org/centos/8/cloud/x86_64/openstack-victoria/Packages/o/openstack-cloudkitty-common-13.0.0-1.el8.noarch.rpm > Thanks a lot for the hint! > > Will patch it downstream waiting for COS patch. > > Le mar. 23 nov. 2021 ? 15:15, Ga?l THEROND a ?crit : >> >> aaaaah nice catch! I'll check that out as I use CentOS packages; it may actually just be that! >> >> Thanks a lot! >> >> Le mar. 23 nov. 2021 ? 15:09, Rafael Weing?rtner a ?crit : >>> >>> I guess that the rule "context_is_admin" might have some weird definition in your version. Can you check it? >>> >>> On Tue, Nov 23, 2021 at 12:06 PM Rafael Weing?rtner wrote: >>>> >>>> Can you check this one? >>>> https://review.opendev.org/c/openstack/cloudkitty/+/785132 >>>> >>>> On Tue, Nov 23, 2021 at 12:01 PM Ga?l THEROND wrote: >>>>> >>>>> Hi everyone! >>>>> >>>>> Today I faced a weird situation with one of our cloud platforms using victoria release. >>>>> >>>>> When trying to get a summary of projects rates would it be through Horizon or CLI using the admin user of the platform we've got the following error message: >>>>> >>>>> https://paste.opendev.org/show/bIgG6owrN9B2F3O7iqYG/ >>>>> >>>>> From my understanding of the default policies of cloudkitty, this error seems to be a bit odd as the admin user profile actually match the default rules. >>>>> >>>>> At least as exposed in: >>>>> >>>>> https://opendev.org/openstack/cloudkitty/src/branch/stable/victoria/cloudkitty/common/policies/base.py >>>>> and >>>>> https://opendev.org/openstack/cloudkitty/src/branch/stable/victoria/cloudkitty/common/policies/v1/report.py >>>>> >>>>> Unless I misunderstood something (please correct me if I'm wrong), it's supposed to at least be ok with the matching. >>>> >>>> >>>> >>>> -- >>>> Rafael Weing?rtner >>> >>> >>> >>> -- >>> Rafael Weing?rtner From fungi at yuggoth.org Fri Dec 10 14:30:45 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Fri, 10 Dec 2021 14:30:45 +0000 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: References: <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <20211209180309.qnqfd3dflvj4uqul@yuggoth.org> Message-ID: <20211210143045.afkmz4g5ezeaz5ac@yuggoth.org> On 2021-12-10 10:43:10 +0100 (+0100), Jean-Philippe Evrard wrote: > On Thu, Dec 9, 2021, at 19:03, Jeremy Stanley wrote: [...] > > * Skyline performs its release versioning by committing the > > version strings to files in its repositories rather than > > assigning them at build time from Git tag metadata. This leads > > to ambiguity as to what exact state of the repository represents > > that version, as well as opening up the possibility of > > forgetting to merge the version update before tagging (a very > > common problem which drove OpenStack to rely on its current > > model). Wholly different tools are also needed in order to > > track versions for these projects as a result. Should the rest > > of OpenStack's projects follow suit there? Has something changed > > to make the idea of using Git to signal releases a bad one? > > You're raising again a valid point, for which the skyline team > will have to find a solution. > > For me, it boils down to defining that projects should follow a > "project release interface", which explains the expected behaviour > in regards of releasing for projects. We don't have such. I am not > claiming we should, I am merely asking us to think about what we > are doing. > > Projects could indeed move to use pbr/setuptools. However, I am > not sure many developers understand why. Should we simply explain > what is expected in terms of building the version from git? People > might even come with interesting solutions. Promoting creativity > will create tech debt, but might also reduce it over time. What if > we don't need pbr anymore? Will we even notice? Isn't that exactly what I did? You'll notice here and in my review I didn't say "use PBR," I outlined the properties desired instead. It's the case that using PBR is likely the easiest way to do that since we already have a lot of examples, but like you I don't think that the implementation is important as long as we can find a suitable solution to the problems we face. That said, using a consistent implementation across as many projects as possible increases our familiarity and ability to address new issues with the fewest set of solutions. > > * Skyline uses makefiles to script its build and test workflows > > rather than tox, adding a tox wrapper around the make targets > > in order to nominally comply with the OpenStack PTI. As a > > result, it's unable to rely on most of tox's features for > > environment isolation, package management, path overrides, et > > cetera. OpenStack projects have solved a lot of cross-project > > development workflow challenges through applying consistent > > changes across the tox.ini files of repositories, should those > > solutions be abandoned and recreated for make instead? > > I didn't say so. I am only partially agreeing with you there, for > once :) > > For me, tox is a test runner that provides many nice features > thanks to its code (and integration with setuptools) I would > prefer if python projects using tox for testing interface indeed. > Should that be the standard? I don't know. Does it make sense to > rely on tox features for isolations/package management/other? I > don't know. You might be aware that it's possible to integrate tox > with poetry. Does it provide all we _need_? I don't know: We > aren't explicit in the requirements of what we need, we just talk > about tools. > > If I were to rephrase: Is it easier to fit the existing mold or > clarify what we need, why, and how we can deal with the tech debt? > Probably the former. Should we do it? I don't know. The TC should > however evaluate this seriously. Let's be clear, I have nothing against make. In fact, I somewhat prefer it, since it's useful for projects written in all sorts of different programming languages beyond Python. At one point we even came very close to recommending OpenStack switch from tox to make during one of the more unfortunate backward-incompatible tox releases some years ago. Clark had a proof of concept recreating the fundamentals of tox itself in Makefiles, and it wasn't half bad. But having one project use make while the rest use tox isn't really great for anyone, it turns the make-based project into a second-class citizen within the whole. Its maintainers have to do a lot more work to support the same workflows with a different tool, and the community as a whole incurs some additional burden in working around that difference (additional CI jobs, separate functionality, different developer documentation). If we think make is a superior solution we should use it across all of OpenStack and not just in a few random projects. > > * Skyline does not put its external Python package dependencies in > > central lists nor use tools which would allow it to rely on > > OpenStack's global constraints implementation, making it very > > hard for them to guarantee that their software can run with the > > same versions of shared dependencies as the rest of OpenStack. > > Is this unnecessary? Have we given up on the idea that two > > pieces of OpenStack software can be coinstalled into one > > environment, or that Linux distributions won't be stuck having > > to package many different versions of the same software because > > OpenStack projects don't agree on which versions work for them > > any longer? > > Now, we are hitting the nail, IMO :) > > This is a bit of the issue's core, isn't it? I think that's what > we need to document. While I (personally) don't believe it's > important nowadays (cf. containers), I am sure distros will > disagree with me. It's important to know who we are pleasing, as > we can't please everyone ;) Let's be clear about the audience. [...] Just my opinion, but I think buying into container hype to the point where we make OpenStack only deployable through containerization would be a massive step backwards. -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From openstack at a.spamming.party Fri Dec 10 16:29:42 2021 From: openstack at a.spamming.party (Jean-Philippe Evrard) Date: Fri, 10 Dec 2021 17:29:42 +0100 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <20211210143045.afkmz4g5ezeaz5ac@yuggoth.org> References: <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <20211209180309.qnqfd3dflvj4uqul@yuggoth.org> <20211210143045.afkmz4g5ezeaz5ac@yuggoth.org> Message-ID: Hello, On Fri, Dec 10, 2021, at 15:30, Jeremy Stanley wrote: > Isn't that exactly what I did? You'll notice here and in my review I > didn't say "use PBR," I outlined the properties desired instead. This is exactly what you did, and what I like in your answer. Everything you said, as wise as it is, isn't documented in the governance [1]. You have knowledge, and that knowledge on why certain things are done should be recorded for the future. The more we share, the more people will be able to challenge it or understand it. > That said, using a consistent implementation across as many projects > as possible increases our familiarity and ability to address new > issues with the fewest set of solutions. Totally agree. I have so many points I want to raise. I realise asking what I have in mind will cause more trouble on this mailing list than the value it will bring. I just want the TC to think broader than the "we used to do it this way, and this is the only way". I am sure the topic raised some questions, and will give enough food for thoughts. Hence, I hope it will result in the improvement of OpenStack :) > Just my opinion, but I think buying into container hype to the point > where we make OpenStack only deployable through containerization > would be a massive step backwards. I didn't say it was a goal or if I consider this important. I just wanted to mention that they are other ways to consume software than distribution packages. Side note: I have spent to much time on this, so I will just step out now. I hope the TC got enough details about what to make out of this in the future. If you need an extra $0.02, you know where to find me :) Regards, Jean-Philippe Evrard (evrardjp) [1]: https://governance.openstack.org/tc/reference/new-projects-requirements.html From ozzzo at yahoo.com Fri Dec 10 16:50:02 2021 From: ozzzo at yahoo.com (Albert Braden) Date: Fri, 10 Dec 2021 16:50:02 +0000 (UTC) Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: References: <5dd6d28f-9955-7ca5-0ab8-0c5ce11ceb7e@redhat.com> <14084e87df22458caa7668eea8b496b6@verisign.com> <1147779219.1274196.1639086048233@mail.yahoo.com> Message-ID: <986294621.1553814.1639155002132@mail.yahoo.com> Sorry, that was a transcription error. I thought "True" and my fingers typed "False." The correct lines are: [oslo_messaging_rabbit] amqp_durable_queues = True On Friday, December 10, 2021, 02:55:55 AM EST, Herve Beraud wrote: If you plan to let `amqp_durable_queues = False` (i.e if you plan to keep this config equal to false), then you don't need to add these config lines as this is already the default value [1]. [1] https://opendev.org/openstack/oslo.messaging/src/branch/master/oslo_messaging/_drivers/amqp.py#L34 Le?jeu. 9 d?c. 2021 ??22:40, Albert Braden a ?crit?: Replying from my home email because I've been asked to not email the list from my work email anymore, until I get permission from upper management. I'm not sure I follow. I was planning to add 2 lines to etc/kolla/config/global.conf: [oslo_messaging_rabbit] amqp_durable_queues = False Is that not sufficient? What is involved in configuring dedicated control exchanges for each service? What would that look like in the config? From: Herve Beraud Sent: Thursday, December 9, 2021 2:45 AM To: Bogdan Dobrelya Cc: openstack-discuss at lists.openstack.org Subject: [EXTERNAL] Re: [ops] [kolla] RabbitMQ High Availability ? Caution:?This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe.? ? ? Le?mer. 8 d?c. 2021 ??11:48, Bogdan Dobrelya a ?crit?: Please see inline >> I read this with great interest because we are seeing this issue. Questions: >> >> 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? >> 2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? >> 3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? > > Note that even having rabbit HA policies adjusted like that and its HA > replication factor [0] decreased (e.g. to a 2), there still might be > high churn caused by a large enough number of replicated durable RPC > topic queues. And that might cripple the cloud down with the incurred > I/O overhead because a durable queue requires all messages in it to be > persisted to a disk (for all the messaging cluster replicas) before they > are ack'ed by the broker. > > Given that said, Oslo messaging would likely require a more granular > control for topic exchanges and the durable queues flag - to tell it to > declare as durable only the most critical paths of a service. A single > config setting and a single control exchange per a service might be not > enough. Also note that therefore, amqp_durable_queue=True requires dedicated control exchanges configured for each service. Those that use 'openstack' as a default cannot turn the feature ON. Changing it to a service specific might also cause upgrade impact, as described in the topic [3]. ? The same is true for `amqp_auto_delete=True`. That requires dedicated control exchanges else it won't work if each service defines its own policy on a shared control exchange (e.g `openstack`) and if policies differ from each other. ? [3] https://review.opendev.org/q/topic:scope-config-opts > > There are also race conditions with durable queues enabled, like [1]. A > solution could be where each service declare its own dedicated control > exchange with its own configuration. > > Finally, openstack components should add perhaps a *.next CI job to test > it with durable queues, like [2] > > [0] https://www.rabbitmq.com/ha.html#replication-factor > > [1] > https://zuul.opendev.org/t/openstack/build/aa514dd788f34cc1be3800e6d7dba0e8/log/controller/logs/screen-n-cpu.txt > > [2] https://review.opendev.org/c/openstack/nova/+/820523 > >> >> Does anyone have a sample set of RMQ config files that they can share? >> >> It looks like my Outlook has ruined the link; reposting: >> [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando -- Best regards, Bogdan Dobrelya, Irc #bogdando -- Herv? Beraud Senior Software Engineer at Red Hat irc: hberaud https://github.com/4383/ https://twitter.com/4383hberaud -- Herv? BeraudSenior Software Engineer at Red Hatirc: hberaudhttps://github.com/4383/https://twitter.com/4383hberaud -------------- next part -------------- An HTML attachment was scrubbed... URL: From billy.olsen at canonical.com Fri Dec 10 17:07:00 2021 From: billy.olsen at canonical.com (Billy Olsen) Date: Fri, 10 Dec 2021 10:07:00 -0700 Subject: [charms] nominate Hemanth Nakkina for charms-core In-Reply-To: References: <3c167485-421c-4861-bbad-6f588b923ee9@gmail.com> Message-ID: On Fri, Dec 10, 2021 at 7:08 AM Alex Kavanagh wrote: > > > On Fri, Dec 10, 2021 at 1:24 PM Edward Hope-Morley > wrote: > >> I would like propose Hemanth Nakkina for charms-core. Hemanth has really >> stepped up his efforts to provide invaluable contributions both in code >> and reviews and I beleive he is in a good position to be considered for >> core where he will be able to provide additional value to the project. >> Hemanth has demonstrated rigour for quality through his contributions >> and is always keen to help others with his deep knowledge of the charms >> and Openstack. Here is a list of all his contributions: >> >> patches: >> https://review.opendev.org/q/owner:hemanth.nakkina%2540canonical.com >> reviews: >> https://review.opendev.org/q/reviewedby:hemanth.nakkina%2540canonical.com >> >> I hope you will join me in voting for Hemanth. >> > > +1 from me; Hemanth has submitted good patches and thoughtful reviews in > my experience. He would be a good member of charms-core. > > Cheers > Alex. > > I concur with Alex here. +1 from me as well. - Billy > >> - Ed >> >> >> > > -- > Alex Kavanagh - Software Engineer > OpenStack Engineering - Data Centre Development - Canonical Ltd > -------------- next part -------------- An HTML attachment was scrubbed... URL: From smooney at redhat.com Fri Dec 10 17:26:19 2021 From: smooney at redhat.com (Sean Mooney) Date: Fri, 10 Dec 2021 17:26:19 +0000 Subject: [nova][cinder][devstack][qa] ceph installs python3-logutils which conflicts with neutron Message-ID: hi o/ gibi noticed that the?nova-ceph-multistore job started failing this morning with ERROR: Cannot uninstall 'logutils'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. https://bugs.launchpad.net/nova/+bug/1954427 we quickly identifed that python3-logutils was beeing pulled in by the devstack-plugin-ceph as a sideeffect fo installing ceph. functions-common:apt_get:1207 : sudo DEBIAN_FRONTEND=noninteractive http_proxy= https_proxy= no_proxy= apt-get --option Dpkg::Options::=--force-confold --assume-yes install ceph libnss3- tools python3-rados python3-rbd Reading package lists... Building dependency tree... Reading state information... The following packages were automatically installed and are no longer required: libboost-iostreams1.71.0 libboost-thread1.71.0 Use 'sudo apt autoremove' to remove them. The following additional packages will be installed: ceph-base ceph-common ceph-mgr ceph-mgr-modules-core ceph-mon ceph-osd libbabeltrace1 libcephfs2 libdw1 libgoogle-perftools4 libjaeger libleveldb1d liblttng-ust-ctl4 liblttng-ust0 libnuma1 liboath0 librabbitmq4 librados2 libradosstriper1 librbd1 librdkafka1 librgw2 libsnappy1v5 libtcmalloc-minimal4 liburcu6 logrotate python-pastedeploy-tpl python3-bcrypt python3-bs4 python3-ceph-argparse python3-ceph-common python3-cephfs python3-cffi-backend python3-cherrypy3 python3-cryptography python3-dateutil python3-jwt python3-logutils python3-mako python3-markupsafe python3-openssl python3-paste python3-pastedeploy python3-pecan python3-prettytable python3-rgw python3-simplegeneric python3-singledispatch python3-soupsieve python3-tempita python3-waitress python3-webob python3-webtest python3-werkzeug to adress this i have proposed a patch based on how we workaround unremovable python packages in devstack fixup_tools script too the devstack-plugin-ceph repo to adress this. https://review.opendev.org/c/openstack/devstack-plugin-ceph/+/821396/ untill that merges expect to see all ceph based jobs to fail when installing neutron. regards sean From thierry at openstack.org Fri Dec 10 17:32:32 2021 From: thierry at openstack.org (Thierry Carrez) Date: Fri, 10 Dec 2021 18:32:32 +0100 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> Message-ID: <514915c2-79c1-10ce-547f-94576bb01b9b@openstack.org> Ghanshyam Mann wrote: > [...] > Yes, those are the inconsistency we currently have in the Skyline project, but that is not what they will stick to. In Yoga PTG[1], the skyline team iterates through all these points and plans to > improve those and be consistent with the other OpenStack projects. They said they will> work on packaging, PTI, using Oslo code etc. Ultimately there is a chicken and egg situation there -- they need to work on making it more like OpenStack, but it's difficult to justify the time to work on this until it actually is accepted as an OpenStack project. We should communicate expectations clearly, make sure that promises are made, and then there is a leap of faith. But IMHO it is a very small leap as the TC can still control if it should be included in the next release, and the TC can also remove the project team if it never gets where it needs to be. > Boxiang Zhu mentioned the same in the governance patch also[2]. > > If they deny moving Skyline towards the OpenStack way of packing, testing, using Oslo, etc., we should consider whether to reject the OpenStack official project or modify our new > project application criteria. But this is not the case, and they agree to improve all those things. > > Now we have more clear direction from them in the governance patch also, so I feel this is good > to go as an official project, and while they will be in OpenStack, we all can help them to > improve the things in OpenStack way. I agree. Expectations have been clearly communicated. It's time for us to give them a good chance to meet those. -- Thierry Carrez (ttx) From gmann at ghanshyammann.com Fri Dec 10 18:48:25 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Fri, 10 Dec 2021 12:48:25 -0600 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> Message-ID: <17da5acd08a.1243b54b2101642.1273683447221903066@ghanshyammann.com> ---- On Fri, 10 Dec 2021 02:16:59 -0600 Jean-Philippe Evrard wrote ---- > Hello, > > On Fri, Dec 10, 2021, at 01:18, Ghanshyam Mann wrote: > > Yes, those are the inconsistency we currently have in the Skyline > > project, but that is not what they will stick to. In Yoga PTG[1], the > > skyline team iterates through all these points and plans to > > improve those and be consistent with the other OpenStack projects. They > > said they will > > work on packaging, PTI, using Oslo code etc. > > > > Boxiang Zhu mentioned the same in the governance patch also[2]. > > The questions I have still stand: Do we need to be more lax in the governance to not _require_ to follow the PTI (and/or have an intention to follow through)? Can the TC make an exception? When is the deadline for projects not following our testing to become compliant? What is the risk if the newly accepted project is not doing it? Moving it to a non-official project? Is that worth the effort? Those are all good points. We discussed around similar one in PTG and even considering if we should introduce different level of the official project like CNCF (Sandbox, graduate) or our old model or integrated vs incubated so that we can make it clear what projects are in-progress of becoming the OpenStack way and which are already. But these add more complications than solve the problem. We agreed to try a new concept with 'Tech Preview' to monitor the project health as well as the OpenStack consistency. Let's wait for that to be up and there we can review/improve the process. Or PTI change if needed. We will also define what is the deadline for projects to be in 'Tech Preview process. Let's wait for the Skyline team to improve the things which hey already mentioned to do and I think that can be done in a better way if they are in OpenStack community. If they do not then it can be dropped from OpenStack governance is nothing different from any existing projects do that. -gmann > > > If they deny moving Skyline towards the OpenStack way of packing, > > testing, using Oslo, etc., we should consider whether to reject the > > OpenStack official project or modify our new > > project application criteria. But this is not the case, and they agree > > to improve all those things. > > I agree. Let's not talk hypothetics. > > > Now we have more clear direction from them in the governance patch > > also, so I feel this is good > > to go as an official project, and while they will be in OpenStack, we > > all can help them to > > improve the things in OpenStack way. > > Perfect, great news. > > > > > Also, TC discussed in PTG the new concept of 'Tech pre-review' where we > > will monitor > > such new projects' health, contribution, and direction going in the > > direction of OpenStack way. This > > 'Tech pre-review' is not up yet, so let's see how we define it. > > I am sorry I have missed that from the notes I read. > > > > > [1] > > http://lists.openstack.org/pipermail/openstack-discuss/2021-October/025541.html > > [2] https://review.opendev.org/c/openstack/governance/+/814037 > > [3] https://etherpad.opendev.org/p/tc-yoga-ptg#L263 > > > > -gmann > > > > > > > > Regards, > > > Jean-Philippe Evrard (evrardjp) > > > > > > [1]: > > https://governance.openstack.org/tc/reference/new-projects-requirements.html > > > [2]: > > https://governance.openstack.org/tc/reference/project-testing-interface.html > > > [3]: https://governance.openstack.org/tc/reference/pti/python.html > > > [4]: https://governance.openstack.org/tc/reference/pti/golang.html > > > [5]: https://github.com/Masterminds/glide > > > [6]: https://travis-ci.org/github/Masterminds/glide > > > > > > > > From gmann at ghanshyammann.com Fri Dec 10 18:51:51 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Fri, 10 Dec 2021 12:51:51 -0600 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: <514915c2-79c1-10ce-547f-94576bb01b9b@openstack.org> References: <17c84b557a6.12930f0e21106266.7388077538937209855@ghanshyammann.com> <446265bd-eb57-a5a6-2f5d-937c6cdad372@debian.org> <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <17da1b4e9cc.df721cfb35645.2419671251769404839@ghanshyammann.com> <514915c2-79c1-10ce-547f-94576bb01b9b@openstack.org> Message-ID: <17da5aff3a0.c58e5125101745.7894421896103761771@ghanshyammann.com> ---- On Fri, 10 Dec 2021 11:32:32 -0600 Thierry Carrez wrote ---- > Ghanshyam Mann wrote: > > [...] > > Yes, those are the inconsistency we currently have in the Skyline project, but that is not what they will stick to. In Yoga PTG[1], the skyline team iterates through all these points and plans to > improve those and be consistent with the other OpenStack projects. > They said they will> work on packaging, PTI, using Oslo code etc. > Ultimately there is a chicken and egg situation there -- they need to > work on making it more like OpenStack, but it's difficult to justify the > time to work on this until it actually is accepted as an OpenStack project. > > We should communicate expectations clearly, make sure that promises are > made, and then there is a leap of faith. But IMHO it is a very small > leap as the TC can still control if it should be included in the next > release, and the TC can also remove the project team if it never gets > where it needs to be. > > > Boxiang Zhu mentioned the same in the governance patch also[2]. > > > > If they deny moving Skyline towards the OpenStack way of packing, testing, using Oslo, etc., we should consider whether to reject the OpenStack official project or modify our new > > project application criteria. But this is not the case, and they agree to improve all those things. > > > > Now we have more clear direction from them in the governance patch also, so I feel this is good > > to go as an official project, and while they will be in OpenStack, we all can help them to > > improve the things in OpenStack way. > > I agree. Expectations have been clearly communicated. It's time for us > to give them a good chance to meet those. Indeed. Also, I think we are making assumptions too early (I mentioned in this thread), if Skyline team deny doing OpenStack way then it is better to discuss all these points otherwise, we should just wait for them to improve the things and even better is that we (who need this project or think it is an interested one) should be part of their team and improve things. -gmann > > -- > Thierry Carrez (ttx) > > From pierre at stackhpc.com Fri Dec 10 21:02:57 2021 From: pierre at stackhpc.com (Pierre Riteau) Date: Fri, 10 Dec 2021 22:02:57 +0100 Subject: [tc][all] Follow up pain point discussion In-Reply-To: References: Message-ID: On Mon, 6 Dec 2021 at 11:59, Stephen Finucane wrote: > novaclient ? full feature parity since 5.5.0 > > We finally introduced a proper '--block-device' option for 'server create' in 5.5.0, which closed our remaining RFE. There have been bugfixes since but it's feature complete, if not bug free. What about `nova host-evacuate-live` and a few other commands identified in doc/source/cli/data/nova.csv? aggregate-cache-images,,Request images be cached. (Supported by API versions '2.81' - '2.latest') [hint: use '-- os-compute-api-version' flag to show help message for proper version] host-evacuate,,Evacuate all instances from failed host. host-evacuate-live,,Live migrate all instances off the specified host to other available hosts. host-meta,,Set or Delete metadata on all instances of a host. host-servers-migrate,,Cold migrate all instances off the specified host to other available hosts. hypervisor-servers,,List servers belonging to specific hypervisors. hypervisor-uptime,,Display the uptime of the specified hypervisor. instance-action,,Show an action. instance-action-list,,List actions on a server. instance-usage-audit-log,,List/Get server usage audits. migration-list,,Print a list of migrations. version-list,,List all API versions. volume-update,,Update volume attachment. From gmann at ghanshyammann.com Fri Dec 10 21:50:58 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Fri, 10 Dec 2021 15:50:58 -0600 Subject: [all][tc] What's happening in Technical Committee: summary 10th Dec, 21: Reading: 10 min Message-ID: <17da653f057.e296ef341967.7707586008920002731@ghanshyammann.com> Hello Everyone, Here is this week's summary of the Technical Committee activities. 1. TC Meetings: ============ * Most of the meeting discussions are summarized below (Completed or in-progress activities section). Meeting full logs are available @ https://meetings.opendev.org/meetings/tc/2021/tc.2021-12-09-15.00.log.html * Next week's meeting is on IRC on 16th Dec, Thursday 15:00 UTC, feel free the topic on the agenda[1] by 15th Dec. 2. What we completed this week: ========================= * 'Consistent and Secure Default RBAC' is selected as community-wide goal [2] * Add NVidia vGPU plugin charm to OpenStack charms[3] 3. Activities In progress: ================== TC Tracker for Yoga cycle ------------------------------ * This etherpad includes the Yoga cycle targets/working items for TC[4]. Open Reviews ----------------- * 5 open reviews for ongoing activities[5]. OpenStack Pain points discussion ---------------------------------------- We had a separate video call on Wed 8th Dec and discussed four projects' pain points[6] and also added the tag to categorize them. Details are sent by Rico in a separate email[7] and we will continue the discussion in Jan. Stay tuned on that ML thread to know the next meeting details. Skyline as an official project --------------------------------- Discussion going on in ML thread[8] regarding the things/improvement to do in Skyline project for various things to make it consistent with other OpenStack projects. Skyline team has confirmed to work on all those points on the governance patch[9] which is under review by TC. TC position on release cadence ------------------------------------- Discussion is in-progress in ML thread[10]. No other updates from TC on this and we will set up a separate call to continue the discussion sometime in Jan (after the Christmas holidays). Yoga testing runtime ------------------------- Governance patch is up for review on keeping python3.6 support in Yoga[11] Fixing Zuul config error -------------------------- If your project is listed in zuul config error, please start planning to fix those[12] Community-wide goal updates ------------------------------------ You might have seen my email[13] about goal status and planning the work, below is the current status: * Selected (active) goals[14]: ** Migrate from oslo.rootwrap to oslo.privsep ** Consistent and Secure Default RBAC * Goal proposal under review: ** FIPS compatibility and compliance [15]. Adjutant need maintainers and PTLs ------------------------------------------- We are still waiting to hear from Braden, Albert on permission to work on this project[16]. Project updates ------------------- * Retire js-openstack-lib [17] 4. How to contact the TC: ==================== If you would like to discuss or give feedback to TC, you can reach out to us in multiple ways: 1. Email: you can send the email with tag [tc] on openstack-discuss ML[18]. 2. Weekly meeting: The Technical Committee conduct a weekly meeting every Thursday 15 UTC [19] 3. Ping us using 'tc-members' nickname on #openstack-tc IRC channel. [1] https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting [2] https://governance.openstack.org/tc/goals/selected/consistent-and-secure-rbac.html [3] https://review.opendev.org/c/openstack/governance/+/819856 [4] https://etherpad.opendev.org/p/tc-yoga-tracker [5] https://review.opendev.org/q/projects:openstack/governance+status:open [6] https://etherpad.opendev.org/p/pain-point-elimination [7] http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026245.html [8] http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026206.html [9] https://review.opendev.org/c/openstack/governance/+/814037 [10] http://lists.openstack.org/pipermail/openstack-discuss/2021-November/025684.html [11] https://review.opendev.org/c/openstack/governance/+/820195 [12] https://etherpad.opendev.org/p/zuul-config-error-openstack [13] http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026218.html [14] https://governance.openstack.org/tc/goals/selected/index.html [15] https://review.opendev.org/c/openstack/governance/+/816587 [16] http://lists.openstack.org/pipermail/openstack-discuss/2021-November/025786.html [17] https://review.opendev.org/c/openstack/governance/+/798540 [18] http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-discuss [19] http://eavesdrop.openstack.org/#Technical_Committee_Meeting -gmann From zigo at debian.org Fri Dec 10 22:08:01 2021 From: zigo at debian.org (Thomas Goirand) Date: Fri, 10 Dec 2021 23:08:01 +0100 Subject: [all][tc] Skyline as a new official project [was: What's happening in Technical Committee: summary 15th Oct, 21: Reading: 5 min] In-Reply-To: References: <93C08133-972B-44BE-9F2A-661A1B86651F@99cloud.net> <20211018121818.rerqlp7ek7z3rnya@yuggoth.org> <19a6d9d2-ed39-39cd-1c85-8fee97f93b8b@debian.org> <4b4a3c1f-67eb-9e77-ec54-b5d2a698c4a2@debian.org> <20211022224028.blm3djon2rrgbguk@yuggoth.org> <17d922a8b65.dcad9c98484045.6727468203019450661@ghanshyammann.com> <46f5f7d7-d0ba-4870-9c5a-1de3586bdaff@www.fastmail.com> <20211209180309.qnqfd3dflvj4uqul@yuggoth.org> <20211210143045.afkmz4g5ezeaz5ac@yuggoth.org> Message-ID: <9ea9126b-8ff4-f4a2-377e-c130c6b92c56@debian.org> On 12/10/21 5:29 PM, Jean-Philippe Evrard wrote: > Totally agree. I have so many points I want to raise. I realise asking what I have in mind will cause more trouble on this mailing list than the value it will bring. I just want the TC to think broader than the "we used to do it this way, and this is the only way". Willing to change things within OpenStack is a completely different topic than "doing things my own way". I very welcome change if you have a cross-project attitude but I don't welcome "we don't do like the others, because we can" type of thinking. Packaging Telemetry, I've hated many times the design decision that was just a little bit different. It wasn't better or worse, just different. And it caused a lot of troubles just because I had to learn "their way" of doing things. I suppose the experience was the same for contributors. For example, aodh has it's build-dependency (well, "test requirements" if you prefer the Python wording than the distro ones) in setup.cfg instead of test-requirements.txt. Both works, but when I upgrade any component, I diff the test-requirements.txt between version, to see if there's new Build-Depends: to add to my package. But with Aodh, I have to remember to do it with setup.cfg. Its *VERY* annoying. Yes, I'm aware that this way, Aodh can have different test-requirements for mysql and postgresql for example. But this is still very annoying. I would have very much prefer if the Telemetry guys pushed so that *ALL* projects would move to push test-requirements.txt within setup.cfg, and *THEN* I would have accepted the change. Do you see where I'm going? :) I have other examples in mind, but I believe what's above is enough. > I am sure the topic raised some questions, and will give enough food for thoughts. Hence, I hope it will result in the improvement of OpenStack :) If you don't do things on your own different than the others, and push every projects to move one direction at the same time then it's going to be an improvement. If you are the only one doing your own way, it's going to be a regression, and a pain for everyone. >> Just my opinion, but I think buying into container hype to the point >> where we make OpenStack only deployable through containerization >> would be a massive step backwards. > > I didn't say it was a goal or if I consider this important. > I just wanted to mention that they are other ways to consume software than distribution packages. Probably, but it's not up to you to decide for the others how they want to consume your deliverables. > Side note: I have spent to much time on this, so I will just step out now. > I hope the TC got enough details about what to make out of this in the future. > If you need an extra $0.02, you know where to find me :) > > Regards, > Jean-Philippe Evrard (evrardjp) I very much hope that Skyline will be in a super nice shape in 6 months from now, so I can include in in Debian Bookworm! :) By the way, we discussed Python packaging a lot, but never JS. There's a huge discussing that sooner or later will have to get started too. NPM is a very dangerous thing that often leads to "npm install world" (ie: a JS orgy that is barely maintainable). I have to admit that I didn't look closely into how that's done in Skyline. But let's just take something else as the worst example: Grafana. I've heard someone from Red Hat attempted to package it, and discovered it included (directly or indirectly) 1600 JS packages, and then gave-up. How is Skyline in this regard? Cheers, Thomas Goirand (zigo) From mkopec at redhat.com Sat Dec 11 12:38:27 2021 From: mkopec at redhat.com (Martin Kopec) Date: Sat, 11 Dec 2021 13:38:27 +0100 Subject: [Interop] No WG meeting on Dec 23 and 30. In-Reply-To: References: Message-ID: Happy holidays to you all! On Thu, 9 Dec 2021 at 18:47, Arkady Kanevsky wrote: > Happy holidays. > > -- > Arkady Kanevsky, Ph.D. > Phone: 972 707-6456 > Corporate Phone: 919 729-5744 ext. 8176456 > -- Martin Kopec Senior Software Quality Engineer Red Hat EMEA -------------- next part -------------- An HTML attachment was scrubbed... URL: From frank at tomatoservers.com Sat Dec 11 12:46:02 2021 From: frank at tomatoservers.com (=?UTF-8?B?RnJhbmsgSHdh?=) Date: Sat, 11 Dec 2021 15:46:02 +0300 Subject: =?UTF-8?B?UmVbMl06IFtJbnRlcm9wXSBObyBXRyBtZWV0aW5nIG9uIERlYyAyMyBhbmQg?= =?UTF-8?B?MzAu?= In-Reply-To: References: Message-ID: <1639226762.82325909@f195.i.mail.ru> Happy holidays to you! -- Frank Hwa >Saturday, 11 December 2021, 15:40 +03:00 from Martin Kopec : > >Happy holidays to you all! > >On Thu, 9 Dec 2021 at 18:47, Arkady Kanevsky < akanevsk at redhat.com > wrote: >>Happy holidays. >> >>-- >>Arkady Kanevsky, Ph.D. >>Phone: 972 707-6456 >>Corporate Phone: 919 729-5744 ext. 8176456 > > >-- >Martin Kopec >Senior Software Quality Engineer >Red Hat EMEA > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Sat Dec 11 20:21:38 2021 From: zigo at debian.org (Thomas Goirand) Date: Sat, 11 Dec 2021 21:21:38 +0100 Subject: Is heat-cfntools still an active project? Message-ID: <2dbe4219-7e93-d588-05ba-5490ddbb6784@debian.org> Hi, The last release of heat-cfntools was in 2016, and in Debian, we're considering removing removing boto (as boto is unmaintained upstream, replaced by boto3). If nobody does anything, we may remove both boto and heat-cfntools from Debian. Thoughts anyone? Cheers, Thomas Goirand (zigo) From yipikai7 at gmail.com Sun Dec 12 17:38:46 2021 From: yipikai7 at gmail.com (Cedric Lemarchand) Date: Sun, 12 Dec 2021 18:38:46 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: <9a3cf269-6b89-fb32-6dc5-4fd708171cb2@inovex.de> Message-ID: So I validated what from my point of view seems to be the most "supported" way to boot an Openstack instance in PXE : - add the flavor the extra specs "hw:boot_menu = 1" (thanks Sean) - spawn the instance - configure the extra_dhcp_opts of the associated port (thanks Christian). Its not supported with the CLI, but doable with an sdk, below a Pythonic example [1] - reboot the instance, hit esc at the bios prompt then third choice (iPXE) This allows the use of the Openstack DHCP and to keep port security enabled, which is nice. Cheers [1] --- conn.update_port(port_id, extra_dhcp_opts=[ {"opt_name": "server-ip-address","opt_value": "10.0.0.15"}, {"opt_name": "bootfile-name", "opt_value": "pxelinux.0"}, {"opt_name": "tftp-server", "opt_value": "10.0.0.15"} ] ) --- On Wed, Dec 8, 2021 at 8:52 PM Cedric Lemarchand wrote: > > Thanks for the heads up. > > Pointer is about ovn, I didnt find anything regarding ovs. > > Did the Neutron API support these options both for ovn and ovs ? > > On Wed, Dec 8, 2021, 17:12 Christian Rohmann wrote: >> >> On 03/12/2021 21:01, Cedric Lemarchand wrote: >> > We faced the same use case, and we end up doing something like this: >> > >> > - disable security on the network provider (or on the instance port >> > where pxe boot is needed) >> > - deploy or rescue the intance with an ipxe image >> > >> >> As for DHCP, you can very much use the OpenStack DHCP and just set the >> options like next-server / server-ip-address >> >> * >> https://docs.openstack.org/api-ref/network/v2/?expanded=update-port-detail#extra-dhcp-option-extra-dhcp-opt-extension >> * https://docs.openstack.org/neutron/latest/ovn/dhcp_opts.html >> >> >> and this IP then points to your VM with the TFTP and other PXE / iPXE >> tooling running. >> >> >> >> >> Regards >> >> >> Christian >> From smooney at redhat.com Sun Dec 12 21:08:05 2021 From: smooney at redhat.com (Sean Mooney) Date: Sun, 12 Dec 2021 21:08:05 +0000 Subject: [tc][all] Follow up pain point discussion In-Reply-To: References: Message-ID: On Fri, 2021-12-10 at 22:02 +0100, Pierre Riteau wrote: > On Mon, 6 Dec 2021 at 11:59, Stephen Finucane wrote: > > novaclient ? full feature parity since 5.5.0 > > > > We finally introduced a proper '--block-device' option for 'server create' in 5.5.0, which closed our remaining RFE. There have been bugfixes since but it's feature complete, if not bug free. > > What about `nova host-evacuate-live` and a few other commands > identified in doc/source/cli/data/nova.csv? this has deliberatlly not being ported to OSC and we never intend to add it. we advise people to impelemtnt this them selves as it purly a client side impleantion and discurage using the exsiting host-eveacuate-live as its error handeling if some live mirations fail is undefiend. you can do a better job yourself. > > aggregate-cache-images,,Request images be cached. (Supported by API > versions '2.81' - '2.latest') [hint: use '-- os-compute-api-version' > flag to show help message for proper version] > host-evacuate,,Evacuate all instances from failed host. > host-evacuate-live,,Live migrate all instances off the specified host > to other available hosts. all the host eveacutate apis were not ported for the same reasons as host-evacuate-live. this is something the nova team nolonger wants to maintain and we deliberatly chose to not port those to osc. > host-meta,,Set or Delete metadata on all instances of a host. again this is not something we want to maintain. its not part of the nova api and can be implented yourself in a few lines of bash. > host-servers-migrate,,Cold migrate all instances off the specified as with the other host-evacuate command this is condiers depercated and was intentionally not ported > host to other available hosts. > hypervisor-servers,,List servers belonging to specific hypervisors. openstack server list --all-projects --host by the way all the host-evacuate,host-evacuate-live and host-servers-migrate commands do is call the list endpoint as show aboove to get vms on a give host and loop over them calling evacuate, live migrate or cold migrate there is no error handeling and for hsot-server-migrate i dont think it actully confirms the migration. it may but the dosc dont say one way our another. it certenly wont do the right thing in most caes if there is an error so you should not use it if you really care about the vms. > hypervisor-uptime,,Display the uptime of the specified hypervisor. the uptime of a hyperivor is part of a deprectated api that was mainly used by xen that is technially avaiable as part of the hyperviors api now but os-host has been deprecated for years and it will eventraully be removed. > instance-action,,Show an action. > instance-action-list,,List actions on a server. the instance action are aviable via openstack server event {list,show} > instance-usage-audit-log,,List/Get server usage audits > usage is aviabel via "openstack usage {list,show}" > . > migration-list,,Print a list of migrations. migration info is aviable at "openstack server migration {abort,confirm,force complete,list,rever,show}" and you can also confrim or revert a migrate via "openstack server migrate {conrim,revert}" > version-list,,List all API versions. > volume-update,,Update volume attachment. volume updates shoudl be done via cinder all nova proxy apis have been deprecated for 3+ year proably even longer so no clinet shoudl still be using those the equivalent woudl openstack volume migrate i belvie. the volume-update api is really swap volume i belive > From amy at demarco.com Sun Dec 12 23:53:47 2021 From: amy at demarco.com (Amy Marrich) Date: Sun, 12 Dec 2021 17:53:47 -0600 Subject: [Diversity] D&I WG Meeting Time and IRC vs Video Poll Message-ID: Hi everyone, The Diversity and Inclusion WG is holding a poll to see if there would be interest in attending if it were on another day and time as well as if IRC vs Video would be preferred. We currently meet the first Monday of the month at 17:00 UTC. And hope to have a planning meeting as our first meeting of the year. Our first meeting of the new year is slated for January 3rd at 17:00 UTC and for IRC. We will take the poll results into consideration and may cancel the January meeting to allow folks to plan on attending in which case February's meeting would be the first of the year. Please vote for your preferred days and times AND also for IRC and/or Video. https://doodle.com/poll/mnzxkauzwf2isssm?utm_source=poll&utm_medium=link We are keeping to morning hours to encourage participation from EMEA but not overly early for PST folks to attend. We will keep the poll up through the end of the year. Please feel free to reach out with any questions or concerns and remember the Diversity and Inclusion WG is for ALL projects under the Open Infra Foundation! Thanks, Amy (spotz) -------------- next part -------------- An HTML attachment was scrubbed... URL: From xiaolihope1008 at 163.com Mon Dec 13 01:54:29 2021 From: xiaolihope1008 at 163.com (Di XiaoLi) Date: Mon, 13 Dec 2021 09:54:29 +0800 (GMT+08:00) Subject: [cyborg][nova][qa]Question about Accelerator unbinding Message-ID: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> hi, Cyborg and nova team: I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. Here are my questions: 1. What is the function of the arq unbind command in cyborg ? 2. How to unbind the accelerator which bounded to vm? Does nova or cyborg support this function now? Here are my steps: step1: openstack accelerator arq unbind ed205084-f58d-4fad-99b4-327a1398858f +---------------------+--------------------------------------+ | Field | Value | +---------------------+--------------------------------------+ | uuid | ed205084-f58d-4fad-99b4-327a1398858f | | state | Unbound | | device_profile_name | ssd | | hostname | None | | device_rp_uuid | None | | instance_uuid | None | | attach_handle_type | | | attach_handle_info | {} | +---------------------+--------------------------------------+ step2: login vm and check the device, but it still here. step3: stop vm and start vm, met the following error: "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" -------------- next part -------------- An HTML attachment was scrubbed... URL: From lokendrarathour at gmail.com Mon Dec 13 05:27:29 2021 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Mon, 13 Dec 2021 10:57:29 +0530 Subject: [Triple0][Undercloud ] Deployment failed because of Bonds Message-ID: Hi Team, we have created a bond with name bond-mgmt for the provisioning n/w and passed that name as local interface: Undercloud.conf: *..* *..* *# Network interface on the Undercloud that will be handling the PXE* *# boots and DHCP for Overcloud instances. (string value)* *local_interface = bond-mgmt* *..* *..* Because of this undercloud deployment is getting failed with the error as below: [root at localhost ~]# ovs-vsctl show b3db7071-58f7-4201-8fd5-9577dec0020e Bridge br-ctlplane fail_mode: standalone Port bond-mgmt Interface bond-mgmt error: "could not open network device bond-mgmt (No such device)" Port br-ctlplane Interface br-ctlplane type: internal ovs_version: "2.12.0" [root at localhost ~]# Query: - is bonding supported in Undercloud installation for provisioning n/w? - if yes then what is the correct way to do so. we are trying to deploy Train on Centos 8. -- ~ Lokendra skype: lokendrarathour -------------- next part -------------- An HTML attachment was scrubbed... URL: From ruslanas at lpic.lt Mon Dec 13 07:22:58 2021 From: ruslanas at lpic.lt (=?UTF-8?Q?Ruslanas_G=C5=BEibovskis?=) Date: Mon, 13 Dec 2021 09:22:58 +0200 Subject: [Triple0][Undercloud ] Deployment failed because of Bonds In-Reply-To: References: Message-ID: Try providing physical interface instead, just to see if it works. On Mon, 13 Dec 2021, 07:35 Lokendra Rathour, wrote: > Hi Team, > we have created a bond with name bond-mgmt for the provisioning n/w and > passed that name as local interface: > Undercloud.conf: > > *..* > *..* > > *# Network interface on the Undercloud that will be handling the PXE* > *# boots and DHCP for Overcloud instances. (string value)* > *local_interface = bond-mgmt* > *..* > *..* > > > > Because of this undercloud deployment is getting failed with the error as > below: > > [root at localhost ~]# ovs-vsctl show > b3db7071-58f7-4201-8fd5-9577dec0020e > Bridge br-ctlplane > fail_mode: standalone > Port bond-mgmt > Interface bond-mgmt > error: "could not open network device bond-mgmt (No such > device)" > Port br-ctlplane > Interface br-ctlplane > type: internal > ovs_version: "2.12.0" > [root at localhost ~]# > > Query: > > - is bonding supported in Undercloud installation for provisioning n/w? > - if yes then what is the correct way to do so. we are trying to > deploy Train on Centos 8. > > > > -- > ~ Lokendra > skype: lokendrarathour > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From lokendrarathour at gmail.com Mon Dec 13 07:30:54 2021 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Mon, 13 Dec 2021 13:00:54 +0530 Subject: [Triple0][Undercloud ] Deployment failed because of Bonds In-Reply-To: References: Message-ID: Hi Ruslanas, Thanks for the response. Yes, it does work with a physical interface. We are looking forward to configuring if something is needed to use the bond interface. any input on this would be helpfull. On Mon, Dec 13, 2021 at 12:53 PM Ruslanas G?ibovskis wrote: > Try providing physical interface instead, just to see if it works. > > On Mon, 13 Dec 2021, 07:35 Lokendra Rathour, > wrote: > >> Hi Team, >> we have created a bond with name bond-mgmt for the provisioning n/w and >> passed that name as local interface: >> Undercloud.conf: >> >> *..* >> *..* >> >> *# Network interface on the Undercloud that will be handling the PXE* >> *# boots and DHCP for Overcloud instances. (string value)* >> *local_interface = bond-mgmt* >> *..* >> *..* >> >> >> >> Because of this undercloud deployment is getting failed with the error as >> below: >> >> [root at localhost ~]# ovs-vsctl show >> b3db7071-58f7-4201-8fd5-9577dec0020e >> Bridge br-ctlplane >> fail_mode: standalone >> Port bond-mgmt >> Interface bond-mgmt >> error: "could not open network device bond-mgmt (No such >> device)" >> Port br-ctlplane >> Interface br-ctlplane >> type: internal >> ovs_version: "2.12.0" >> [root at localhost ~]# >> >> Query: >> >> - is bonding supported in Undercloud installation for provisioning >> n/w? >> - if yes then what is the correct way to do so. we are trying to >> deploy Train on Centos 8. >> >> >> >> -- >> ~ Lokendra >> skype: lokendrarathour >> >> >> -- ~ Lokendra skype: lokendrarathour -------------- next part -------------- An HTML attachment was scrubbed... URL: From songwenping at inspur.com Mon Dec 13 08:27:16 2021 From: songwenping at inspur.com (=?utf-8?B?QWxleCBTb25nICjlrovmloflubMp?=) Date: Mon, 13 Dec 2021 08:27:16 +0000 Subject: =?utf-8?B?562U5aSNOiBbbGlzdHMub3BlbnN0YWNrLm9yZ+S7o+WPkV1bY3lib3JnXVtu?= =?utf-8?Q?ova][qa]Question_about_Accelerator_unbinding?= In-Reply-To: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> References: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> Message-ID: Emm, the arq unbind cmd is used of deleting instance. Recently we donot support hot-plug/unplug device for instance. We only support binding devices when create instance and unbinding devices when delete instance. Best Regards. ???: Di XiaoLi [mailto:xiaolihope1008 at 163.com] ????: 2021?12?13? 9:54 ???: openstack-discuss ??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding hi, Cyborg and nova team: I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. Here are my questions: 1. What is the function of the arq unbind command in cyborg ? 2. How to unbind the accelerator which bounded to vm? Does nova or cyborg support this function now? Here are my steps: step1: openstack accelerator arq unbind ed205084-f58d-4fad-99b4-327a1398858f +---------------------+--------------------------------------+ | Field | Value | +---------------------+--------------------------------------+ | uuid | ed205084-f58d-4fad-99b4-327a1398858f | | state | Unbound | | device_profile_name | ssd | | hostname | None | | device_rp_uuid | None | | instance_uuid | None | | attach_handle_type | | | attach_handle_info | {} | +---------------------+--------------------------------------+ step2: login vm and check the device, but it still here. step3: stop vm and start vm, met the following error: "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3774 bytes Desc: not available URL: From xiaolihope1008 at 163.com Mon Dec 13 08:31:16 2021 From: xiaolihope1008 at 163.com (Di XiaoLi) Date: Mon, 13 Dec 2021 16:31:16 +0800 (GMT+08:00) Subject: =?utf-8?Q?=E5=9B=9E=E5=A4=8D=EF=BC=9A=E7=AD=94=E5=A4=8D:_[lists.opensta?= =?utf-8?Q?ck.org=E4=BB=A3=E5=8F=91][cyborg][nov?= =?utf-8?Q?a][qa]Question_about_Accelerator_unbinding?= In-Reply-To: References: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> Message-ID: <295c0cd5.394f.17db2eadc46.Coremail.xiaolihope1008@163.com> Okay, I got it. Thanks very much for your answer. ?2021?12?13? 16:29?Alex Song (???) ??? Emm, the arq unbind cmd is used of deleting instance. Recently we donot support hot-plug/unplug device for instance. We only support binding devices when create instance and unbinding devices when delete instance. Best Regards. ???: Di XiaoLi [mailto:xiaolihope1008 at 163.com] ????: 2021?12?13? 9:54 ???: openstack-discuss ??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding hi, Cyborg and nova team: I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. Here are my questions: 1. What is the function of the arq unbind command in cyborg ? 2. How to unbind the accelerator which bounded to vm? Does nova or cyborg support this function now? Here are my steps: step1: openstack accelerator arq unbind ed205084-f58d-4fad-99b4-327a1398858f +---------------------+--------------------------------------+ | Field | Value | +---------------------+--------------------------------------+ | uuid | ed205084-f58d-4fad-99b4-327a1398858f | | state | Unbound | | device_profile_name | ssd | | hostname | None | | device_rp_uuid | None | | instance_uuid | None | | attach_handle_type | | | attach_handle_info | {} | +---------------------+--------------------------------------+ step2: login vm and check the device, but it still here. step3: stop vm and start vm, met the following error: "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" -------------- next part -------------- An HTML attachment was scrubbed... URL: From songwenping at inspur.com Mon Dec 13 08:33:22 2021 From: songwenping at inspur.com (=?utf-8?B?QWxleCBTb25nICjlrovmloflubMp?=) Date: Mon, 13 Dec 2021 08:33:22 +0000 Subject: =?utf-8?B?562U5aSNOiBbbGlzdHMub3BlbnN0YWNrLm9yZ+S7o+WPkV1bY3lib3JnXVtu?= =?utf-8?Q?ova][qa]Question_about_Accelerator_unbinding?= References: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> Message-ID: Anyway, we should shield this cmd to avoid misoperation of arq and lead the instance start error for now. ???: Alex Song (???) ????: 2021?12?13? 16:27 ???: 'xiaolihope1008 at 163.com' ; 'openstack-discuss at lists.openstack.org' ??: ??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding Emm, the arq unbind cmd is used of deleting instance. Recently we donot support hot-plug/unplug device for instance. We only support binding devices when create instance and unbinding devices when delete instance. Best Regards. ???: Di XiaoLi [mailto:xiaolihope1008 at 163.com] ????: 2021?12?13? 9:54 ???: openstack-discuss > ??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding hi, Cyborg and nova team: I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. Here are my questions: 1. What is the function of the arq unbind command in cyborg ? 2. How to unbind the accelerator which bounded to vm? Does nova or cyborg support this function now? Here are my steps: step1: openstack accelerator arq unbind ed205084-f58d-4fad-99b4-327a1398858f +---------------------+--------------------------------------+ | Field | Value | +---------------------+--------------------------------------+ | uuid | ed205084-f58d-4fad-99b4-327a1398858f | | state | Unbound | | device_profile_name | ssd | | hostname | None | | device_rp_uuid | None | | instance_uuid | None | | attach_handle_type | | | attach_handle_info | {} | +---------------------+--------------------------------------+ step2: login vm and check the device, but it still here. step3: stop vm and start vm, met the following error: "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3774 bytes Desc: not available URL: From songwenping at inspur.com Mon Dec 13 08:36:33 2021 From: songwenping at inspur.com (=?utf-8?B?QWxleCBTb25nICjlrovmloflubMp?=) Date: Mon, 13 Dec 2021 08:36:33 +0000 Subject: =?utf-8?B?562U5aSNOiBbbGlzdHMub3BlbnN0YWNrLm9yZ+S7o+WPkV1bY3lib3JnXVtu?= =?utf-8?Q?ova][qa]Question_about_Accelerator_unbinding?= References: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> Message-ID: <73d8fedf28bb4be9bea4b07ce50c3540@inspur.com> Anyway, we should shield this cmd to avoid misoperation of arq and lead the instance start error for now. ???: Alex Song (???) ????: 2021?12?13? 16:27 ???: 'xiaolihope1008 at 163.com' ; 'openstack-discuss at lists.openstack.org' ??: ??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding Emm, the arq unbind cmd is used of deleting instance. Recently we donot support hot-plug/unplug device for instance. We only support binding devices when create instance and unbinding devices when delete instance. Best Regards. ???: Di XiaoLi [mailto:xiaolihope1008 at 163.com] ????: 2021?12?13? 9:54 ???: openstack-discuss > ??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding hi, Cyborg and nova team: I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. Here are my questions: 1. What is the function of the arq unbind command in cyborg ? 2. How to unbind the accelerator which bounded to vm? Does nova or cyborg support this function now? Here are my steps: step1: openstack accelerator arq unbind ed205084-f58d-4fad-99b4-327a1398858f +---------------------+--------------------------------------+ | Field | Value | +---------------------+--------------------------------------+ | uuid | ed205084-f58d-4fad-99b4-327a1398858f | | state | Unbound | | device_profile_name | ssd | | hostname | None | | device_rp_uuid | None | | instance_uuid | None | | attach_handle_type | | | attach_handle_info | {} | +---------------------+--------------------------------------+ step2: login vm and check the device, but it still here. step3: stop vm and start vm, met the following error: "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3774 bytes Desc: not available URL: From xiaolihope1008 at 163.com Mon Dec 13 08:46:04 2021 From: xiaolihope1008 at 163.com (Di XiaoLi) Date: Mon, 13 Dec 2021 16:46:04 +0800 (GMT+08:00) Subject: =?utf-8?Q?=E5=9B=9E=E5=A4=8D=EF=BC=9A=E7=AD=94=E5=A4=8D:_[lists.opensta?= =?utf-8?Q?ck.org=E4=BB=A3=E5=8F=91][cyborg][nov?= =?utf-8?Q?a][qa]Question_about_Accelerator_unbinding?= In-Reply-To: References: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> Message-ID: <50f5d17d.3b79.17db2f86af7.Coremail.xiaolihope1008@163.com> Yes, Alex, this command does confuse me a bit. Hope we can support bind/unbind accelerators in the future. I also found some errors in kolla-ansible when deploying cyborg. I want to know how active is the cyborg project? Shall we fix the deploy bug in kolla-ansible and backport the bug fixes to released branches? ?2021?12?13? 16:35?Alex Song (???) ??? Anyway, we should shield this cmd to avoid misoperation of arq and lead the instance start error for now. ???: Alex Song (???) ????: 2021?12?13? 16:27 ???: 'xiaolihope1008 at 163.com' ; 'openstack-discuss at lists.openstack.org' ??:??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding Emm, the arq unbind cmd is used of deleting instance. Recently we donot support hot-plug/unplug device for instance. We only support binding devices when create instance and unbinding devices when delete instance. Best Regards. ???: Di XiaoLi [mailto:xiaolihope1008 at 163.com] ????: 2021?12?13? 9:54 ???: openstack-discuss ??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding hi, Cyborg and nova team: I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. Here are my questions: 1. What is the function of the arq unbind command in cyborg ? 2. How to unbind the accelerator which bounded to vm? Does nova or cyborg support this function now? Here are my steps: step1: openstack accelerator arq unbind ed205084-f58d-4fad-99b4-327a1398858f +---------------------+--------------------------------------+ | Field | Value | +---------------------+--------------------------------------+ | uuid | ed205084-f58d-4fad-99b4-327a1398858f | | state | Unbound | | device_profile_name | ssd | | hostname | None | | device_rp_uuid | None | | instance_uuid | None | | attach_handle_type | | | attach_handle_info | {} | +---------------------+--------------------------------------+ step2: login vm and check the device, but it still here. step3: stop vm and start vm, met the following error: "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" -------------- next part -------------- An HTML attachment was scrubbed... URL: From arne.wiebalck at cern.ch Mon Dec 13 08:55:07 2021 From: arne.wiebalck at cern.ch (Arne Wiebalck) Date: Mon, 13 Dec 2021 09:55:07 +0100 Subject: [baremetal-sig][ironic] Tue Dec 14, 2021, 2pm UTC: Metal3 Overview Message-ID: <9b43905a-e407-6574-d720-7b0bf0ce73b8@cern.ch> Dear all, The Bare Metal SIG will meet tomorrow Tue Dec 14, 2021, at 2pm UTC on zoom. The meeting will feature a "topic-of-the-day" presentation by Dmitry Tantsur (dtantsur) with an "Overview of Metal3" (and I hear there might be a demo even!) As usual, all details on https://etherpad.opendev.org/p/bare-metal-sig Everyone is welcome: come along, learn from the expert and bring your questions! Cheers, Arne From songwenping at inspur.com Mon Dec 13 09:21:11 2021 From: songwenping at inspur.com (=?utf-8?B?QWxleCBTb25nICjlrovmloflubMp?=) Date: Mon, 13 Dec 2021 09:21:11 +0000 Subject: =?utf-8?B?562U5aSNOiDlm57lpI3vvJrnrZTlpI06IFtsaXN0cy5vcGVuc3RhY2sub3Jn?= =?utf-8?B?5Luj5Y+RXVtjeWJvcmddW25vdmFdW3FhXVF1ZXN0aW9uIGFib3V0IEFjY2Vs?= =?utf-8?Q?erator_unbinding?= In-Reply-To: <50f5d17d.3b79.17db2f86af7.Coremail.xiaolihope1008@163.com> References: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> <50f5d17d.3b79.17db2f86af7.Coremail.xiaolihope1008@163.com> Message-ID: <56bee1fcd5e04abfab812ca5ab6ca1d1@inspur.com> About the bind/unbind feature, we have discussed on the nova PTG[1], they are in our task list. Welcome to join us if you have interest. We also use kolla-ansible to deploy cyborg, but found no problem right now, if you meet error, pls report bugs on the launchpad[2] and fix them if you have time. [1] https://etherpad.opendev.org/p/nova-xena-ptg#L375 [2] https://bugs.launchpad.net/openstack-cyborg ???: Di XiaoLi [mailto:xiaolihope1008 at 163.com] ????: 2021?12?13? 16:46 ???: Alex Song (???) ??: openstack-discuss at lists.openstack.org ??: ?????: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding Yes, Alex, this command does confuse me a bit. Hope we can support bind/unbind accelerators in the future. I also found some errors in kolla-ansible when deploying cyborg. I want to know how active is the cyborg project? Shall we fix the deploy bug in kolla-ansible and backport the bug fixes to released branches? ?2021?12?13? 16:35? Alex Song (???) ??? Anyway, we should shield this cmd to avoid misoperation of arq and lead the instance start error for now. ???: Alex Song (???) ????: 2021?12?13? 16:27 ???: 'xiaolihope1008 at 163.com ' >; 'openstack-discuss at lists.openstack.org ' > ??: ??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding Emm, the arq unbind cmd is used of deleting instance. Recently we donot support hot-plug/unplug device for instance. We only support binding devices when create instance and unbinding devices when delete instance. Best Regards. ???: Di XiaoLi [mailto:xiaolihope1008 at 163.com] ????: 2021?12?13? 9:54 ???: openstack-discuss > ??: [lists.openstack.org??][cyborg][nova][qa]Question about Accelerator unbinding hi, Cyborg and nova team: I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. Here are my questions: 1. What is the function of the arq unbind command in cyborg ? 2. How to unbind the accelerator which bounded to vm? Does nova or cyborg support this function now? Here are my steps: step1: openstack accelerator arq unbind ed205084-f58d-4fad-99b4-327a1398858f +---------------------+--------------------------------------+ | Field | Value | +---------------------+--------------------------------------+ | uuid | ed205084-f58d-4fad-99b4-327a1398858f | | state | Unbound | | device_profile_name | ssd | | hostname | None | | device_rp_uuid | None | | instance_uuid | None | | attach_handle_type | | | attach_handle_info | {} | +---------------------+--------------------------------------+ step2: login vm and check the device, but it still here. step3: stop vm and start vm, met the following error: "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3774 bytes Desc: not available URL: From chkumar at redhat.com Mon Dec 13 09:52:21 2021 From: chkumar at redhat.com (Chandan Kumar) Date: Mon, 13 Dec 2021 15:22:21 +0530 Subject: [tripleo] Initial plans around stable/wallaby branch on CentOS Stream 9 Message-ID: Hello, The tripleo-ci team is working on drafting the plans for bootstrapping tripleo stable/wallaby branch on CentOS Stream 9. During adding the support of CentOS Stream 9 for tripleo master, we have added lots of patches for example: * tripleo-common: https://review.opendev.org/q/%28topic:%2522cs9%2522+OR+hashtag:%2522cs9%2522%29+status:merged+project:openstack/tripleo-common * tripleo-heat-templates: https://review.opendev.org/q/%28topic:%2522cs9%2522+OR+hashtag:%2522cs9%2522%29+status:merged+project:openstack/tripleo-heat-templates * https://review.opendev.org/q/%28topic:%2522cs9%2522+OR+hashtag:%2522cs9%2522%29+status:merged+project:openstack/tripleo-ansible Backporting these patches in chains for stable/wallaby, will create lots of merge conflicts (few of them are big changes) and might get messy. RDO wallaby DLRN builder bootstrap has also started. In order to avoid that, we are coming with following plan: - Pick the major change and backport that one and add all the remaining changes in that itself (not a clean backport): https://review.opendev.org/c/openstack/tripleo-common/+/821511 - Make sure these backported changes passed on CentOS-8 CI jobs - Once the RDO CS9 wallaby packages are available, Start bootstrapping the container and Image build job on CS9 via testproject. Let us know your thoughts around this. Thanks, Chandan Kumar From cjeanner at redhat.com Mon Dec 13 10:08:12 2021 From: cjeanner at redhat.com (=?UTF-8?Q?C=c3=a9dric_Jeanneret?=) Date: Mon, 13 Dec 2021 11:08:12 +0100 Subject: [tripleo] Initial plans around stable/wallaby branch on CentOS Stream 9 In-Reply-To: References: Message-ID: On 12/13/21 10:52, Chandan Kumar wrote: > Hello, > > The tripleo-ci team is working on drafting the plans for bootstrapping > tripleo stable/wallaby branch on CentOS Stream 9. > > During adding the support of CentOS Stream 9 for tripleo master, we have added > lots of patches for example: > * tripleo-common: > https://review.opendev.org/q/%28topic:%2522cs9%2522+OR+hashtag:%2522cs9%2522%29+status:merged+project:openstack/tripleo-common > * tripleo-heat-templates: > https://review.opendev.org/q/%28topic:%2522cs9%2522+OR+hashtag:%2522cs9%2522%29+status:merged+project:openstack/tripleo-heat-templates > * https://review.opendev.org/q/%28topic:%2522cs9%2522+OR+hashtag:%2522cs9%2522%29+status:merged+project:openstack/tripleo-ansible > > Backporting these patches in chains for stable/wallaby, will create > lots of merge conflicts (few of them are big changes) and might get > messy. > RDO wallaby DLRN builder bootstrap has also started. > > In order to avoid that, we are coming with following plan: > - Pick the major change and backport that one and add all the > remaining changes in that itself (not a clean backport): > https://review.opendev.org/c/openstack/tripleo-common/+/821511 > - Make sure these backported changes passed on CentOS-8 CI jobs > - Once the RDO CS9 wallaby packages are available, Start bootstrapping > the container and Image build job on CS9 via testproject. > > Let us know your thoughts around this. Hey Chandan, As discussed before: I'm all for this solution. There are multiple advantages, among them: less CI hassle, less conflicts, faster merges. Cheers, C. > > Thanks, > > Chandan Kumar > > -- C?dric Jeanneret (He/Him/His) Sr. Software Engineer - OpenStack Platform Deployment Framework TC Red Hat EMEA https://www.redhat.com/ -------------- next part -------------- A non-text attachment was scrubbed... Name: OpenPGP_signature Type: application/pgp-signature Size: 840 bytes Desc: OpenPGP digital signature URL: From rafaelweingartner at gmail.com Mon Dec 13 10:52:49 2021 From: rafaelweingartner at gmail.com (=?UTF-8?Q?Rafael_Weing=C3=A4rtner?=) Date: Mon, 13 Dec 2021 07:52:49 -0300 Subject: [CLOUDKITTY][VICTORIA] - policy prohibit report:get_summary to admin In-Reply-To: References: Message-ID: Hello Pierre, Releases approved. On Fri, Dec 10, 2021 at 11:26 AM Pierre Riteau wrote: > Hello Ga?l, > > This fix will not be included in RDO packages until we issue a new > victoria release, which only then can be packaged as RPM. > I have proposed release patches for both victoria and wallaby. They > need to be approved by Rafael. > > Using Kolla source images is another option, since in recent versions > (including Victoria) they use the tip of stable branches. > > Best wishes, > Pierre > > On Tue, 23 Nov 2021 at 16:31, Ga?l THEROND > wrote: > > > > ah ah! Was exactly that indeed! > > > > So, CentOS cloudkitty common package is not using the latest patch > fixing the issue -> > http://mirror.centos.org/centos/8/cloud/x86_64/openstack-victoria/Packages/o/openstack-cloudkitty-common-13.0.0-1.el8.noarch.rpm > > Thanks a lot for the hint! > > > > Will patch it downstream waiting for COS patch. > > > > Le mar. 23 nov. 2021 ? 15:15, Ga?l THEROND > a ?crit : > >> > >> aaaaah nice catch! I'll check that out as I use CentOS packages; it may > actually just be that! > >> > >> Thanks a lot! > >> > >> Le mar. 23 nov. 2021 ? 15:09, Rafael Weing?rtner < > rafaelweingartner at gmail.com> a ?crit : > >>> > >>> I guess that the rule "context_is_admin" might have some weird > definition in your version. Can you check it? > >>> > >>> On Tue, Nov 23, 2021 at 12:06 PM Rafael Weing?rtner < > rafaelweingartner at gmail.com> wrote: > >>>> > >>>> Can you check this one? > >>>> https://review.opendev.org/c/openstack/cloudkitty/+/785132 > >>>> > >>>> On Tue, Nov 23, 2021 at 12:01 PM Ga?l THEROND < > gael.therond at bitswalk.com> wrote: > >>>>> > >>>>> Hi everyone! > >>>>> > >>>>> Today I faced a weird situation with one of our cloud platforms > using victoria release. > >>>>> > >>>>> When trying to get a summary of projects rates would it be through > Horizon or CLI using the admin user of the platform we've got the following > error message: > >>>>> > >>>>> https://paste.opendev.org/show/bIgG6owrN9B2F3O7iqYG/ > >>>>> > >>>>> From my understanding of the default policies of cloudkitty, this > error seems to be a bit odd as the admin user profile actually match the > default rules. > >>>>> > >>>>> At least as exposed in: > >>>>> > >>>>> > https://opendev.org/openstack/cloudkitty/src/branch/stable/victoria/cloudkitty/common/policies/base.py > >>>>> and > >>>>> > https://opendev.org/openstack/cloudkitty/src/branch/stable/victoria/cloudkitty/common/policies/v1/report.py > >>>>> > >>>>> Unless I misunderstood something (please correct me if I'm wrong), > it's supposed to at least be ok with the matching. > >>>> > >>>> > >>>> > >>>> -- > >>>> Rafael Weing?rtner > >>> > >>> > >>> > >>> -- > >>> Rafael Weing?rtner > -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From rlandy at redhat.com Mon Dec 13 11:21:47 2021 From: rlandy at redhat.com (Ronelle Landy) Date: Mon, 13 Dec 2021 06:21:47 -0500 Subject: [tripleo] Initial plans around stable/wallaby branch on CentOS Stream 9 In-Reply-To: References: Message-ID: On Mon, Dec 13, 2021 at 5:12 AM C?dric Jeanneret wrote: > On 12/13/21 10:52, Chandan Kumar wrote: > > Hello, > > > > The tripleo-ci team is working on drafting the plans for bootstrapping > > tripleo stable/wallaby branch on CentOS Stream 9. > > > > During adding the support of CentOS Stream 9 for tripleo master, we have > added > > lots of patches for example: > > * tripleo-common: > > > https://review.opendev.org/q/%28topic:%2522cs9%2522+OR+hashtag:%2522cs9%2522%29+status:merged+project:openstack/tripleo-common > > * tripleo-heat-templates: > > > https://review.opendev.org/q/%28topic:%2522cs9%2522+OR+hashtag:%2522cs9%2522%29+status:merged+project:openstack/tripleo-heat-templates > > * > https://review.opendev.org/q/%28topic:%2522cs9%2522+OR+hashtag:%2522cs9%2522%29+status:merged+project:openstack/tripleo-ansible > > > > Backporting these patches in chains for stable/wallaby, will create > > lots of merge conflicts (few of them are big changes) and might get > > messy. > > RDO wallaby DLRN builder bootstrap has also started. > > > > In order to avoid that, we are coming with following plan: > > - Pick the major change and backport that one and add all the > > remaining changes in that itself (not a clean backport): > > https://review.opendev.org/c/openstack/tripleo-common/+/821511 > > - Make sure these backported changes passed on CentOS-8 CI jobs > > - Once the RDO CS9 wallaby packages are available, Start bootstrapping > > the container and Image build job on CS9 via testproject. > > > > Let us know your thoughts around this. > > Hey Chandan, > > As discussed before: I'm all for this solution. There are multiple > advantages, among them: less CI hassle, less conflicts, faster merges. > +1 to this approach. We looked at backporting patch by patch and it was much more complicated due to the same files changed in multiple patches. > > Cheers, > > C. > > > > > Thanks, > > > > Chandan Kumar > > > > > > -- > C?dric Jeanneret (He/Him/His) > Sr. Software Engineer - OpenStack Platform > Deployment Framework TC > Red Hat EMEA > https://www.redhat.com/ > -------------- next part -------------- An HTML attachment was scrubbed... URL: From lokendrarathour at gmail.com Mon Dec 13 12:28:33 2021 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Mon, 13 Dec 2021 17:58:33 +0530 Subject: [Triple0] Overcloud Deployment Network Queries Message-ID: Hi Team, As we are trying to deploy Triple0, regarding this we have the following queries: 1. On the Overcloud machine, I have four NICs available so is there a way to bind a particular NIC to a physnet. 1. Also, can we create multiple physnet, one being flat and the other being VLAN, binded with two different NICs respectively? 2. We have created a local registry during Undercloud installation on Undercloud, What is the way to configure local registry images for overcloud deployment? -- ~ Lokendra skype: lokendrarathour -------------- next part -------------- An HTML attachment was scrubbed... URL: From hberaud at redhat.com Mon Dec 13 12:42:10 2021 From: hberaud at redhat.com (Herve Beraud) Date: Mon, 13 Dec 2021 13:42:10 +0100 Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: <986294621.1553814.1639155002132@mail.yahoo.com> References: <5dd6d28f-9955-7ca5-0ab8-0c5ce11ceb7e@redhat.com> <14084e87df22458caa7668eea8b496b6@verisign.com> <1147779219.1274196.1639086048233@mail.yahoo.com> <986294621.1553814.1639155002132@mail.yahoo.com> Message-ID: So, your config snippet LGTM. Le ven. 10 d?c. 2021 ? 17:50, Albert Braden a ?crit : > Sorry, that was a transcription error. I thought "True" and my fingers > typed "False." The correct lines are: > > [oslo_messaging_rabbit] > amqp_durable_queues = True > > On Friday, December 10, 2021, 02:55:55 AM EST, Herve Beraud < > hberaud at redhat.com> wrote: > > > If you plan to let `amqp_durable_queues = False` (i.e if you plan to keep > this config equal to false), then you don't need to add these config lines > as this is already the default value [1]. > > [1] > https://opendev.org/openstack/oslo.messaging/src/branch/master/oslo_messaging/_drivers/amqp.py#L34 > > Le jeu. 9 d?c. 2021 ? 22:40, Albert Braden a ?crit : > > Replying from my home email because I've been asked to not email the list > from my work email anymore, until I get permission from upper management. > > I'm not sure I follow. I was planning to add 2 lines to > etc/kolla/config/global.conf: > > [oslo_messaging_rabbit] > amqp_durable_queues = False > > Is that not sufficient? What is involved in configuring dedicated control > exchanges for each service? What would that look like in the config? > > > From: Herve Beraud > Sent: Thursday, December 9, 2021 2:45 AM > To: Bogdan Dobrelya > Cc: openstack-discuss at lists.openstack.org > Subject: [EXTERNAL] Re: [ops] [kolla] RabbitMQ High Availability > > > > Caution: This email originated from outside the organization. Do not click > links or open attachments unless you recognize the sender and know the > content is safe. > > > > > > Le mer. 8 d?c. 2021 ? 11:48, Bogdan Dobrelya a > ?crit : > > Please see inline > > >> I read this with great interest because we are seeing this issue. > Questions: > >> > >> 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. > Should we be upgrading our Train clusters to use 3.8.x? > >> 2. Document [2] recommends policy > '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible > playbooks, nor in any of the config files in the RMQ container. What would > this look like in Ansible, and what should the resulting container config > look like? > >> 3. It appears that we are not setting "amqp_durable_queues = True". > What does this setting look like in Ansible, and what file does it go into? > > > > Note that even having rabbit HA policies adjusted like that and its HA > > replication factor [0] decreased (e.g. to a 2), there still might be > > high churn caused by a large enough number of replicated durable RPC > > topic queues. And that might cripple the cloud down with the incurred > > I/O overhead because a durable queue requires all messages in it to be > > persisted to a disk (for all the messaging cluster replicas) before they > > are ack'ed by the broker. > > > > Given that said, Oslo messaging would likely require a more granular > > control for topic exchanges and the durable queues flag - to tell it to > > declare as durable only the most critical paths of a service. A single > > config setting and a single control exchange per a service might be not > > enough. > > Also note that therefore, amqp_durable_queue=True requires dedicated > control exchanges configured for each service. Those that use > 'openstack' as a default cannot turn the feature ON. Changing it to a > service specific might also cause upgrade impact, as described in the > topic [3]. > > > > The same is true for `amqp_auto_delete=True`. That requires dedicated > control exchanges else it won't work if each service defines its own policy > on a shared control exchange (e.g `openstack`) and if policies differ from > each other. > > > > [3] https://review.opendev.org/q/topic:scope-config-opts > > > > > There are also race conditions with durable queues enabled, like [1]. A > > solution could be where each service declare its own dedicated control > > exchange with its own configuration. > > > > Finally, openstack components should add perhaps a *.next CI job to test > > it with durable queues, like [2] > > > > [0] https://www.rabbitmq.com/ha.html#replication-factor > > > > [1] > > > https://zuul.opendev.org/t/openstack/build/aa514dd788f34cc1be3800e6d7dba0e8/log/controller/logs/screen-n-cpu.txt > > > > [2] https://review.opendev.org/c/openstack/nova/+/820523 > > > >> > >> Does anyone have a sample set of RMQ config files that they can share? > >> > >> It looks like my Outlook has ruined the link; reposting: > >> [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > > > > > -- > > Best regards, > > Bogdan Dobrelya, > > Irc #bogdando > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando > > > > > -- > > Herv? Beraud > > Senior Software Engineer at Red Hat > > irc: hberaud > > https://github.com/4383/ > > https://twitter.com/4383hberaud > > > > -- > Herv? Beraud > Senior Software Engineer at Red Hat > irc: hberaud > https://github.com/4383/ > https://twitter.com/4383hberaud > > -- Herv? Beraud Senior Software Engineer at Red Hat irc: hberaud https://github.com/4383/ https://twitter.com/4383hberaud -------------- next part -------------- An HTML attachment was scrubbed... URL: From kamil.madac at slovenskoit.sk Mon Dec 13 13:00:58 2021 From: kamil.madac at slovenskoit.sk (=?iso-8859-2?Q?Kamil_Mad=E1=E8?=) Date: Mon, 13 Dec 2021 13:00:58 +0000 Subject: [nova] How to update OS-EXT-SRV-ATTR:hostname In-Reply-To: References: Message-ID: Thanks for suggestion, but changes in /etc/cloud/cloud.cfg are reverted during VM rebuild, so this is probably not the solution for or issue. Kamil ________________________________ From: Thomas Goirand Sent: Tuesday, December 7, 2021 11:53 AM To: Kamil Mad?? ; openstack-discuss Subject: Re: [nova] How to update OS-EXT-SRV-ATTR:hostname On 12/7/21 11:20 AM, Kamil Mad?? wrote: > Hello, > > I would like to ask, how to update OS-EXT-SRV-ATTR:hostname via > openstack CLI? We changed hostnames on bunch of instances, but metadata > service still serves old hostname which causes that every future rebuild > of instance reverts hostname (`hostname` command in instance console) > back to old one. > > Kamil Mad?? > /Slovensko IT a.s./ Looks like what you want to do is disable cloud-init editing of your VM hostname. This is done simply by cloud-init to stop doing this. Simply edit /etc/cloud/cloud.cfg and set: preserve_hostname: true or remove any cloud-init module from the "cloud_init_modules:" list. Simply remove the lines containing: - set_hostname - update_hostname - update_etc_hosts I hope this helps, Cheers, Thomas Goirand (zigo) -------------- next part -------------- An HTML attachment was scrubbed... URL: From kamil.madac at slovenskoit.sk Mon Dec 13 13:06:41 2021 From: kamil.madac at slovenskoit.sk (=?utf-8?B?S2FtaWwgTWFkw6HEjQ==?=) Date: Mon, 13 Dec 2021 13:06:41 +0000 Subject: [nova] How to update OS-EXT-SRV-ATTR:hostname In-Reply-To: References: Message-ID: Thanks for info about Xena API abilities. At the moment we ended up with making changed directly to Nova DB ? ________________________________ From: Stephen Finucane Sent: Tuesday, December 7, 2021 12:30 PM To: Thomas Goirand ; Kamil Mad?? ; openstack-discuss Subject: Re: [nova] How to update OS-EXT-SRV-ATTR:hostname On Tue, 2021-12-07 at 11:53 +0100, Thomas Goirand wrote: > On 12/7/21 11:20 AM, Kamil Mad?? wrote: > > Hello, > > > > I would like to ask, how to update OS-EXT-SRV-ATTR:hostname via > > openstack CLI? We changed hostnames on bunch of instances, but metadata > > service still serves old hostname which causes that every future rebuild > > of instance reverts hostname (`hostname` command in instance console) > > back to old one. > > > > Kamil Mad?? > > /Slovensko IT a.s./ > > Looks like what you want to do is disable cloud-init editing of your VM > hostname. This is done simply by cloud-init to stop doing this. Simply > edit /etc/cloud/cloud.cfg and set: > > preserve_hostname: true > > or remove any cloud-init module from the "cloud_init_modules:" list. > Simply remove the lines containing: > - set_hostname > - update_hostname > - update_etc_hosts > > I hope this helps, > > Cheers, > > Thomas Goirand (zigo) To add to this, there's no way to change the hostname embedded in the instance metadata until Xena, when we added the ability to provide a 'hostname' parameter when creating, updating or rebuilding servers (microversion 2.90). See [1] and [2] for more details. Stephen [1]https://docs.openstack.org/nova/latest/reference/api-microversion-history.html#maximum-in-xena [2]https://specs.openstack.org/openstack/nova-specs/specs/xena/implemented/configurable-instance-hostnames.html -------------- next part -------------- An HTML attachment was scrubbed... URL: From gael.therond at bitswalk.com Mon Dec 13 13:43:52 2021 From: gael.therond at bitswalk.com (=?UTF-8?Q?Ga=C3=ABl_THEROND?=) Date: Mon, 13 Dec 2021 14:43:52 +0100 Subject: [CLOUDKITTY][VICTORIA] - policy prohibit report:get_summary to admin In-Reply-To: References: Message-ID: Hi everyone! Thanks a lot for those information. I?ve fixed it downstream as I was waiting for the bugfix to make it to RDO repos, I see that it should be available soon. Thanks a lot! Le lun. 13 d?c. 2021 ? 11:53, Rafael Weing?rtner < rafaelweingartner at gmail.com> a ?crit : > Hello Pierre, > Releases approved. > > On Fri, Dec 10, 2021 at 11:26 AM Pierre Riteau > wrote: > >> Hello Ga?l, >> >> This fix will not be included in RDO packages until we issue a new >> victoria release, which only then can be packaged as RPM. >> I have proposed release patches for both victoria and wallaby. They >> need to be approved by Rafael. >> >> Using Kolla source images is another option, since in recent versions >> (including Victoria) they use the tip of stable branches. >> >> Best wishes, >> Pierre >> >> On Tue, 23 Nov 2021 at 16:31, Ga?l THEROND >> wrote: >> > >> > ah ah! Was exactly that indeed! >> > >> > So, CentOS cloudkitty common package is not using the latest patch >> fixing the issue -> >> http://mirror.centos.org/centos/8/cloud/x86_64/openstack-victoria/Packages/o/openstack-cloudkitty-common-13.0.0-1.el8.noarch.rpm >> > Thanks a lot for the hint! >> > >> > Will patch it downstream waiting for COS patch. >> > >> > Le mar. 23 nov. 2021 ? 15:15, Ga?l THEROND >> a ?crit : >> >> >> >> aaaaah nice catch! I'll check that out as I use CentOS packages; it >> may actually just be that! >> >> >> >> Thanks a lot! >> >> >> >> Le mar. 23 nov. 2021 ? 15:09, Rafael Weing?rtner < >> rafaelweingartner at gmail.com> a ?crit : >> >>> >> >>> I guess that the rule "context_is_admin" might have some weird >> definition in your version. Can you check it? >> >>> >> >>> On Tue, Nov 23, 2021 at 12:06 PM Rafael Weing?rtner < >> rafaelweingartner at gmail.com> wrote: >> >>>> >> >>>> Can you check this one? >> >>>> https://review.opendev.org/c/openstack/cloudkitty/+/785132 >> >>>> >> >>>> On Tue, Nov 23, 2021 at 12:01 PM Ga?l THEROND < >> gael.therond at bitswalk.com> wrote: >> >>>>> >> >>>>> Hi everyone! >> >>>>> >> >>>>> Today I faced a weird situation with one of our cloud platforms >> using victoria release. >> >>>>> >> >>>>> When trying to get a summary of projects rates would it be through >> Horizon or CLI using the admin user of the platform we've got the following >> error message: >> >>>>> >> >>>>> https://paste.opendev.org/show/bIgG6owrN9B2F3O7iqYG/ >> >>>>> >> >>>>> From my understanding of the default policies of cloudkitty, this >> error seems to be a bit odd as the admin user profile actually match the >> default rules. >> >>>>> >> >>>>> At least as exposed in: >> >>>>> >> >>>>> >> https://opendev.org/openstack/cloudkitty/src/branch/stable/victoria/cloudkitty/common/policies/base.py >> >>>>> and >> >>>>> >> https://opendev.org/openstack/cloudkitty/src/branch/stable/victoria/cloudkitty/common/policies/v1/report.py >> >>>>> >> >>>>> Unless I misunderstood something (please correct me if I'm wrong), >> it's supposed to at least be ok with the matching. >> >>>> >> >>>> >> >>>> >> >>>> -- >> >>>> Rafael Weing?rtner >> >>> >> >>> >> >>> >> >>> -- >> >>> Rafael Weing?rtner >> > > > -- > Rafael Weing?rtner > -------------- next part -------------- An HTML attachment was scrubbed... URL: From ruslanas at lpic.lt Mon Dec 13 14:11:38 2021 From: ruslanas at lpic.lt (=?UTF-8?Q?Ruslanas_G=C5=BEibovskis?=) Date: Mon, 13 Dec 2021 16:11:38 +0200 Subject: [Triple0][Undercloud ] Deployment failed because of Bonds In-Reply-To: References: Message-ID: Is undercloud VM or on baremetal? I do undercloud in VM, so I do bonding on KVM level. That also makes easier to backup. On Mon, 13 Dec 2021 at 09:31, Lokendra Rathour wrote: > Hi Ruslanas, > Thanks for the response. > Yes, it does work with a physical interface. > We are looking forward to configuring if something is needed to use the > bond interface. > any input on this would be helpfull. > > > > > On Mon, Dec 13, 2021 at 12:53 PM Ruslanas G?ibovskis > wrote: > >> Try providing physical interface instead, just to see if it works. >> >> On Mon, 13 Dec 2021, 07:35 Lokendra Rathour, >> wrote: >> >>> Hi Team, >>> we have created a bond with name bond-mgmt for the provisioning n/w and >>> passed that name as local interface: >>> Undercloud.conf: >>> >>> *..* >>> *..* >>> >>> *# Network interface on the Undercloud that will be handling the PXE* >>> *# boots and DHCP for Overcloud instances. (string value)* >>> *local_interface = bond-mgmt* >>> *..* >>> *..* >>> >>> >>> >>> Because of this undercloud deployment is getting failed with the error >>> as below: >>> >>> [root at localhost ~]# ovs-vsctl show >>> b3db7071-58f7-4201-8fd5-9577dec0020e >>> Bridge br-ctlplane >>> fail_mode: standalone >>> Port bond-mgmt >>> Interface bond-mgmt >>> error: "could not open network device bond-mgmt (No such >>> device)" >>> Port br-ctlplane >>> Interface br-ctlplane >>> type: internal >>> ovs_version: "2.12.0" >>> [root at localhost ~]# >>> >>> Query: >>> >>> - is bonding supported in Undercloud installation for provisioning >>> n/w? >>> - if yes then what is the correct way to do so. we are trying to >>> deploy Train on Centos 8. >>> >>> >>> >>> -- >>> ~ Lokendra >>> skype: lokendrarathour >>> >>> >>> > > -- > ~ Lokendra > skype: lokendrarathour > > > -- Ruslanas G?ibovskis +370 6030 7030 -------------- next part -------------- An HTML attachment was scrubbed... URL: From balazs.gibizer at est.tech Mon Dec 13 14:17:41 2021 From: balazs.gibizer at est.tech (Balazs Gibizer) Date: Mon, 13 Dec 2021 15:17:41 +0100 Subject: [nova][cinder][devstack][qa] ceph installs python3-logutils which conflicts with neutron In-Reply-To: References: Message-ID: On Fri, Dec 10 2021 at 05:26:19 PM +0000, Sean Mooney wrote: > hi o/ > > gibi noticed that the nova-ceph-multistore job started failing this > morning > with ERROR: Cannot uninstall 'logutils'. It is a distutils installed > project and thus we cannot accurately determine which files belong to > it which would lead to only a partial uninstall. > > https://bugs.launchpad.net/nova/+bug/1954427 > > we quickly identifed that python3-logutils was beeing pulled in by > the devstack-plugin-ceph as a sideeffect fo installing ceph. > > functions-common:apt_get:1207 : sudo > DEBIAN_FRONTEND=noninteractive http_proxy= https_proxy= no_proxy= > apt-get --option Dpkg::Options::=--force-confold --assume-yes install > ceph libnss3- > tools python3-rados python3-rbd > Reading package lists... > Building dependency tree... > Reading state information... > The following packages were automatically installed and are no longer > required: > libboost-iostreams1.71.0 libboost-thread1.71.0 > Use 'sudo apt autoremove' to remove them. > The following additional packages will be installed: > ceph-base ceph-common ceph-mgr ceph-mgr-modules-core ceph-mon > ceph-osd > libbabeltrace1 libcephfs2 libdw1 libgoogle-perftools4 libjaeger > libleveldb1d > liblttng-ust-ctl4 liblttng-ust0 libnuma1 liboath0 librabbitmq4 > librados2 > libradosstriper1 librbd1 librdkafka1 librgw2 libsnappy1v5 > libtcmalloc-minimal4 liburcu6 logrotate python-pastedeploy-tpl > python3-bcrypt python3-bs4 python3-ceph-argparse python3-ceph-common > python3-cephfs python3-cffi-backend python3-cherrypy3 > python3-cryptography > python3-dateutil python3-jwt python3-logutils python3-mako > python3-markupsafe python3-openssl python3-paste python3-pastedeploy > python3-pecan python3-prettytable python3-rgw python3-simplegeneric > python3-singledispatch python3-soupsieve python3-tempita > python3-waitress > python3-webob python3-webtest python3-werkzeug > > to adress this i have proposed a patch based on how we workaround > unremovable python packages in devstack fixup_tools script > too the devstack-plugin-ceph repo to adress this. > https://review.opendev.org/c/openstack/devstack-plugin-ceph/+/821396/ > > untill that merges expect to see all ceph based jobs to fail when > installing neutron. The fix merged during the weekend and the ceph base jobs seems to be happy. Cheers, gibi > > regards > sean > > From lokendrarathour at gmail.com Mon Dec 13 14:18:56 2021 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Mon, 13 Dec 2021 19:48:56 +0530 Subject: [Triple0][Undercloud ] Deployment failed because of Bonds In-Reply-To: References: Message-ID: Hi, Yes, it is VM, and we are getting the error on VM-based Undercloud only for bonds. although with the physical interface name as the local_interface it runs perfectly. On Mon, Dec 13, 2021 at 7:41 PM Ruslanas G?ibovskis wrote: > Is undercloud VM or on baremetal? > > I do undercloud in VM, so I do bonding on KVM level. > That also makes easier to backup. > > On Mon, 13 Dec 2021 at 09:31, Lokendra Rathour > wrote: > >> Hi Ruslanas, >> Thanks for the response. >> Yes, it does work with a physical interface. >> We are looking forward to configuring if something is needed to use the >> bond interface. >> any input on this would be helpfull. >> >> >> >> >> On Mon, Dec 13, 2021 at 12:53 PM Ruslanas G?ibovskis >> wrote: >> >>> Try providing physical interface instead, just to see if it works. >>> >>> On Mon, 13 Dec 2021, 07:35 Lokendra Rathour, >>> wrote: >>> >>>> Hi Team, >>>> we have created a bond with name bond-mgmt for the provisioning n/w and >>>> passed that name as local interface: >>>> Undercloud.conf: >>>> >>>> *..* >>>> *..* >>>> >>>> *# Network interface on the Undercloud that will be handling the PXE* >>>> *# boots and DHCP for Overcloud instances. (string value)* >>>> *local_interface = bond-mgmt* >>>> *..* >>>> *..* >>>> >>>> >>>> >>>> Because of this undercloud deployment is getting failed with the error >>>> as below: >>>> >>>> [root at localhost ~]# ovs-vsctl show >>>> b3db7071-58f7-4201-8fd5-9577dec0020e >>>> Bridge br-ctlplane >>>> fail_mode: standalone >>>> Port bond-mgmt >>>> Interface bond-mgmt >>>> error: "could not open network device bond-mgmt (No >>>> such device)" >>>> Port br-ctlplane >>>> Interface br-ctlplane >>>> type: internal >>>> ovs_version: "2.12.0" >>>> [root at localhost ~]# >>>> >>>> Query: >>>> >>>> - is bonding supported in Undercloud installation for provisioning >>>> n/w? >>>> - if yes then what is the correct way to do so. we are trying to >>>> deploy Train on Centos 8. >>>> >>>> >>>> >>>> -- >>>> ~ Lokendra >>>> skype: lokendrarathour >>>> >>>> >>>> >> >> -- >> ~ Lokendra >> skype: lokendrarathour >> >> >> > > -- > Ruslanas G?ibovskis > +370 6030 7030 > -- ~ Lokendra skype: lokendrarathour -------------- next part -------------- An HTML attachment was scrubbed... URL: From ruslanas at lpic.lt Mon Dec 13 14:27:34 2021 From: ruslanas at lpic.lt (=?UTF-8?Q?Ruslanas_G=C5=BEibovskis?=) Date: Mon, 13 Dec 2021 16:27:34 +0200 Subject: [Triple0][Undercloud ] Deployment failed because of Bonds In-Reply-To: References: Message-ID: so create bond on kvm/vmware or where ever you are running VM, and add that bond to the bridge, and attach that bonded bridge to VM. On Mon, 13 Dec 2021 at 16:19, Lokendra Rathour wrote: > Hi, > Yes, it is VM, and we are getting the error on VM-based Undercloud only > for bonds. > although with the physical interface name as the local_interface it runs > perfectly. > > > > > On Mon, Dec 13, 2021 at 7:41 PM Ruslanas G?ibovskis > wrote: > >> Is undercloud VM or on baremetal? >> >> I do undercloud in VM, so I do bonding on KVM level. >> That also makes easier to backup. >> >> On Mon, 13 Dec 2021 at 09:31, Lokendra Rathour >> wrote: >> >>> Hi Ruslanas, >>> Thanks for the response. >>> Yes, it does work with a physical interface. >>> We are looking forward to configuring if something is needed to use the >>> bond interface. >>> any input on this would be helpfull. >>> >>> >>> >>> >>> On Mon, Dec 13, 2021 at 12:53 PM Ruslanas G?ibovskis >>> wrote: >>> >>>> Try providing physical interface instead, just to see if it works. >>>> >>>> On Mon, 13 Dec 2021, 07:35 Lokendra Rathour, >>>> wrote: >>>> >>>>> Hi Team, >>>>> we have created a bond with name bond-mgmt for the provisioning n/w >>>>> and passed that name as local interface: >>>>> Undercloud.conf: >>>>> >>>>> *..* >>>>> *..* >>>>> >>>>> *# Network interface on the Undercloud that will be handling the PXE* >>>>> *# boots and DHCP for Overcloud instances. (string value)* >>>>> *local_interface = bond-mgmt* >>>>> *..* >>>>> *..* >>>>> >>>>> >>>>> >>>>> Because of this undercloud deployment is getting failed with the error >>>>> as below: >>>>> >>>>> [root at localhost ~]# ovs-vsctl show >>>>> b3db7071-58f7-4201-8fd5-9577dec0020e >>>>> Bridge br-ctlplane >>>>> fail_mode: standalone >>>>> Port bond-mgmt >>>>> Interface bond-mgmt >>>>> error: "could not open network device bond-mgmt (No >>>>> such device)" >>>>> Port br-ctlplane >>>>> Interface br-ctlplane >>>>> type: internal >>>>> ovs_version: "2.12.0" >>>>> [root at localhost ~]# >>>>> >>>>> Query: >>>>> >>>>> - is bonding supported in Undercloud installation for provisioning >>>>> n/w? >>>>> - if yes then what is the correct way to do so. we are trying >>>>> to deploy Train on Centos 8. >>>>> >>>>> >>>>> >>>>> -- >>>>> ~ Lokendra >>>>> skype: lokendrarathour >>>>> >>>>> >>>>> >>> >>> -- >>> ~ Lokendra >>> skype: lokendrarathour >>> >>> >>> >> >> -- >> Ruslanas G?ibovskis >> +370 6030 7030 >> > > > -- > ~ Lokendra > skype: lokendrarathour > > > -- Ruslanas G?ibovskis +370 6030 7030 -------------- next part -------------- An HTML attachment was scrubbed... URL: From rafaelweingartner at gmail.com Mon Dec 13 14:46:13 2021 From: rafaelweingartner at gmail.com (=?UTF-8?Q?Rafael_Weing=C3=A4rtner?=) Date: Mon, 13 Dec 2021 11:46:13 -0300 Subject: [CLOUDKITTY] Missed CloudKitty meeting today Message-ID: Hello guys, I would like to apologize for missing the CloudKitty meeting today. I was concentrating on some work, and my alarm for the meeting did not ring. Isee that you guys already merged some things, thanks for picking the meeting up. I will add more alarms in my calendar, to avoid this situation in the future. If you need something, just let me know. Again, sorry for the inconvenience; see you guys at our next meeting. -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Mon Dec 13 16:24:23 2021 From: zigo at debian.org (Thomas Goirand) Date: Mon, 13 Dec 2021 17:24:23 +0100 Subject: [all] testtools and support for Python 3.10 (error du to deprecation of distutils) Message-ID: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> Hi, When I try to build rally-openstack, I get this, in the test discovery: /usr/lib/python3/dist-packages/rally/common/cfg.py:19: in from oslo_config import fixture # noqa /usr/lib/python3/dist-packages/oslo_config/fixture.py:18: in import fixtures /usr/lib/python3/dist-packages/fixtures/__init__.py:81: in from fixtures.fixture import ( /usr/lib/python3/dist-packages/fixtures/fixture.py:34: in from fixtures.callmany import ( /usr/lib/python3/dist-packages/fixtures/callmany.py:31: in MultipleExceptions = try_import( /usr/lib/python3/dist-packages/testtools/helpers.py:30: in try_import __import__(module_name) /usr/lib/python3/dist-packages/testtools/__init__.py:99: in from testtools.distutilscmd import TestCommand /usr/lib/python3/dist-packages/testtools/distutilscmd.py:7: in from distutils.core import Command /usr/lib/python3.10/distutils/__init__.py:19: in warnings.warn(_DEPRECATION_MESSAGE, E DeprecationWarning: The distutils package is deprecated and slated for removal in Python 3.12. Use setuptools or check PEP 632 for potential alternatives This warning makes it impossible to discover rally-openstack tests, because of the imports here: https://github.com/testing-cabal/testtools/blob/master/testtools/distutilscmd.py#L7 Can someone work on removing distutils support from Testtools? Or is there a way to trick pytest to not fail? Help would be greatly appreciated here. Cheers, Thomas Goirand (zigo) From smooney at redhat.com Mon Dec 13 16:32:25 2021 From: smooney at redhat.com (Sean Mooney) Date: Mon, 13 Dec 2021 16:32:25 +0000 Subject: [cyborg][nova][qa]Question about Accelerator unbinding In-Reply-To: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> References: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> Message-ID: <99456fa20ca2a93b434608cf313f0047ace47abf.camel@redhat.com> On Mon, 2021-12-13 at 09:54 +0800, Di XiaoLi wrote: > hi, Cyborg and nova team: > I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. > Here are my questions: > 1. What is the function of the arq unbind command in cyborg ? it has two usecases. it call by nova when nova is deleting or moving the vm it can be used by an enduser if they are not using cybrog with nova. > 2. How to unbind the accelerator which bounded to vm? Does nova or cyborg support this function now? > the only way to do that for a nova instance is to resize it to a flavor that does not request the device via the cyborg device profiel in the extra spec. more recently we have started to support using cybrog for neutron nics too. in this case the unbinding can be done by doing a port detach and the device should be remvoed from the vm and unbond in cybrog. > > Here are my steps: > step1: openstack accelerator arq unbind ed205084-f58d-4fad-99b4-327a1398858f this should be treated as an admin/service only oepration wehn using cybrog with nova more on that below. > +---------------------+--------------------------------------+ > > Field | Value | > +---------------------+--------------------------------------+ > > uuid | ed205084-f58d-4fad-99b4-327a1398858f | > > state | Unbound | > > device_profile_name | ssd | > > hostname | None | > > device_rp_uuid | None | > > instance_uuid | None | > > attach_handle_type | | > > attach_handle_info | {} | > +---------------------+--------------------------------------+ > > step2: login vm and check the device, but it still here. > step3: stop vm and start vm, met the following error: > "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" that is what i would expect and is the correct behavior as you violated the contract between nova and cyborg by unbindng the arq. when useing cyborg with nova you shoudl never use the cyborg api driectly expcit to list the devcice profiles. you should treat it like placment in that regard. cybrog when used with nova today is an internal service that end user should at most have read only access too list the profiles. nova does not support cybrog device hot/cold attach or detach. the only way to add or remove a device form cyborg via nova is the device profile request in the flavor. so the only support opartion to change the attach device is resize to a differnt flavor. as i said above if you are attching smartnics using cyborg via neutron then you can also attach/devatch cybrog device by attaching/detaching the neutron port which contains the device profile request. note that if the deivce does not exist on the currnt host the attch will likely fail. detach should more or less always work unless there is an internal errror with one of the 3 service invovled. > > From michal.arbet at ultimum.io Mon Dec 13 16:42:41 2021 From: michal.arbet at ultimum.io (Michal Arbet) Date: Mon, 13 Dec 2021 17:42:41 +0100 Subject: Is heat-cfntools still an active project? In-Reply-To: <2dbe4219-7e93-d588-05ba-5490ddbb6784@debian.org> References: <2dbe4219-7e93-d588-05ba-5490ddbb6784@debian.org> Message-ID: Hmm, Are u sure that it's good idea to remove heat-cfntools ? I think this is used in various user defined SoftareDeployments, WaitConditionals etc ... isn't it ? Kevko Michal Arbet Openstack Engineer Ultimum Technologies a.s. Na Po???? 1047/26, 11000 Praha 1 Czech Republic +420 604 228 897 michal.arbet at ultimum.io *https://ultimum.io * LinkedIn | Twitter | Facebook so 11. 12. 2021 v 21:27 odes?latel Thomas Goirand napsal: > Hi, > > The last release of heat-cfntools was in 2016, and in Debian, we're > considering removing removing boto (as boto is unmaintained upstream, > replaced by boto3). If nobody does anything, we may remove both boto and > heat-cfntools from Debian. > > Thoughts anyone? > > Cheers, > > Thomas Goirand (zigo) > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From fungi at yuggoth.org Mon Dec 13 16:48:16 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Mon, 13 Dec 2021 16:48:16 +0000 Subject: [all] testtools and support for Python 3.10 (error du to deprecation of distutils) In-Reply-To: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> References: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> Message-ID: <20211213164815.ftio25ectty7qb2t@yuggoth.org> On 2021-12-13 17:24:23 +0100 (+0100), Thomas Goirand wrote: > When I try to build rally-openstack, I get this, in the test discovery: [...] > /usr/lib/python3.10/distutils/__init__.py:19: in > warnings.warn(_DEPRECATION_MESSAGE, > E DeprecationWarning: The distutils package is deprecated and slated > for removal in Python 3.12. Use setuptools or check PEP 632 for > potential alternatives [...] > Can someone work on removing distutils support from Testtools? Or is > there a way to trick pytest to not fail? Help would be greatly > appreciated here. PYTHONWARNINGS must be getting set to "error" somewhere. Are you setting it at invocation maybe? In order to be able to run with dependencies that call distutils I currently do something like this: PYTHONWARNINGS = error, ignore:The distutils package is deprecated and slated for removal in Python 3.12. Use setuptools or check PEP 632 for potential alternatives:DeprecationWarning All on one line of course. And yes, ideally someone should fix testtools before 3.11 releases, so people can still use it on 3.12 alpha releases once those are available. -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From sbauza at redhat.com Mon Dec 13 17:14:53 2021 From: sbauza at redhat.com (Sylvain Bauza) Date: Mon, 13 Dec 2021 18:14:53 +0100 Subject: [nova] Next spec review day on Tuesday Dec 14th In-Reply-To: References: Message-ID: Kindly reminder that tomorrow will be our last spec review day for the Yoga release. Please be around in IRC or look at Gerrit if you want to discuss with us and thanks if you also review specs. Thanks, -Sylvain On Tue, Nov 30, 2021 at 6:41 PM Sylvain Bauza wrote: > Hola, > As agreed during today's Nova meeting, we will have another spec review > day on Dec 14th. > Please sharpen your pens and write your specs in advance, and just be > around during this day if you want to discuss with the reviewers. > > Thanks, > -Sylvain > -------------- next part -------------- An HTML attachment was scrubbed... URL: From michal.arbet at ultimum.io Mon Dec 13 17:33:27 2021 From: michal.arbet at ultimum.io (Michal Arbet) Date: Mon, 13 Dec 2021 18:33:27 +0100 Subject: [swift][tempest][kolla] Message-ID: Hello to eveyrone, Please, could someone help me with swift capabilities not working when I'm tempesting openstack test stack ? Tempest : rm -rf /tmp/tempest-lock/; refstack-client test -v -c /opt/tempest/tempest.conf -- --regex tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest (.venv) root at ca6353106d83:/opt/refstack-client# rm -rf /tmp/tempest-lock/; refstack-client test -v -c /opt/tempest/tempest.conf -- --regex tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest 2021-12-13 17:44:33.001 4358 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf 2021-12-13 17:44:33,839 refstack_client:518 INFO Starting Tempest test... 2021-12-13 17:44:33.839 4358 INFO refstack_client [-] Starting Tempest test... {0} setUpClass (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) [0.000000s] ... FAILED Captured traceback: ~~~~~~~~~~~~~~~~~~~ Traceback (most recent call last): File "/opt/refstack-client/.tempest/tempest/test.py", line 181, in setUpClass raise value.with_traceback(trace) File "/opt/refstack-client/.tempest/tempest/test.py", line 174, in setUpClass cls.resource_setup() File "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", line 36, in resource_setup super(AccountQuotasNegativeTest, cls).resource_setup() File "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line 95, in resource_setup body = cls.capabilities_client.list_capabilities() File "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", line 32, in list_capabilities self._error_checker(resp, body) File "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line 799, in _error_checker raise exceptions.Unauthorized(resp_body, resp=resp) tempest.lib.exceptions.Unauthorized: Unauthorized Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The request you have made requires authentication.'} ============================== Failed 1 tests - output below: ============================== setUpClass (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) ---------------------------------------------------------------------------------------------- Captured traceback: ~~~~~~~~~~~~~~~~~~~ Traceback (most recent call last): File "/opt/refstack-client/.tempest/tempest/test.py", line 181, in setUpClass raise value.with_traceback(trace) File "/opt/refstack-client/.tempest/tempest/test.py", line 174, in setUpClass cls.resource_setup() File "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", line 36, in resource_setup super(AccountQuotasNegativeTest, cls).resource_setup() File "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line 95, in resource_setup body = cls.capabilities_client.list_capabilities() File "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", line 32, in list_capabilities self._error_checker(resp, body) File "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line 799, in _error_checker raise exceptions.Unauthorized(resp_body, resp=resp) tempest.lib.exceptions.Unauthorized: Unauthorized Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The request you have made requires authentication.'} ====== Totals ====== Ran: 1 tests in 0.0000 sec. - Passed: 0 - Skipped: 0 - Expected Fail: 0 - Unexpected Success: 0 - Failed: 1 Sum of execute time for each test: 0.0000 sec. TEMPEST LOG : Response - Headers: {'content-type': 'application/json', 'content-length': '114', 'www-authenticate': 'Keystone uri=" http://192.168.205.254:5000"', 'x-trans-id': 'tx815e181e33fb4854b2631-0061b7787a', 'x-openstack-request-id': 'tx815e181e33fb4854b2631-0061b7787a', 'date': 'Mon, 13 Dec 2021 16:44:42 GMT', 'connection': 'close', 'status': '401', 'content-location': ' https://api.refstack.ultimum.cloud:8080/info'} Body: b'{"error": {"code": 401, "title": "Unauthorized", "message": "The request you have made requires authentication."}}' _log_request_full /opt/refstack-client/.tempest/tempest/lib/common/rest_client.py:450 Test from command line and from curl . /etc/kolla/refstack.sh ; curl -H "X-Auth-Token: $(openstack token issue -f value -c id)" https://api.refstack.ultimum.cloud:8080/info {"swift": {"version": "2.27.1.dev9", "strict_cors_mode": true, "policies": [{"name": "Policy-0", "aliases": "Policy-0", "default": true}], "allow_account_management": true, "account_autocreate": true, "max_file_size": 5368709122, "max_meta_name_length": 128, "max_meta_value_length": 256, "max_meta_count": 90, "max_meta_overall_size": 4096, "max_header_size": 8192, "max_object_name_length": 1024, "container_listing_limit": 10000, "account_listing_limit": 10000, "max_account_name_length": 256, "max_container_name_length": 256, "extra_header_count": 0}, "container_sync": {"realms": {}}, "bulk_upload": {"max_containers_per_extraction": 10000, "max_failed_extractions": 1000}, "bulk_delete": {"max_deletes_per_request": 10000, "max_failed_deletes": 1000}, "tempurl": {"methods": ["GET", "HEAD", "PUT", "POST", "DELETE"], "incoming_remove_headers": ["x-timestamp"], "incoming_allow_headers": [], "outgoing_remove_headers": ["x-object-meta-*"], "outgoing_allow_headers": ["x-object-meta-public-*"], "allowed_digests": ["sha1", "sha256", "sha512"]}, "ratelimit": {"account_ratelimit": 0.0, "max_sleep_time_seconds": 60.0, "container_ratelimits": [], "container_listing_ratelimits": []}, "container_quotas": {}, "account_quotas": {}, "slo": {"max_manifest_segments": 1000, "max_manifest_size": 8388608, "yield_frequency": 10, "min_segment_size": 1, "allow_async_delete": false}} Python Swiftclient : ubuntu at deploy:/opt/kolla-ansible$ . /etc/kolla/refstack.sh ; swift --os-auth-url http://192.168.205.254:5000/v3 --auth-version 3 --os-project-name refstack --os-project-domain-name default --os-username refstack --os-user-domain-name default --os-password SECRET capabilities Capabilities GET failed: http://192.168.205.254:8080/info 401 Unauthorized [first 60 chars of response] b'{"error": {"code": 401, "title": "Unauthorized", "message": ' Failed Transaction ID: txc1d8607e26eb4cd587459-0061b7791b It looks like swift client is broken, isn't it ? Or ? Maybe kolla-ansible is creating bad roles and config ? (operator_roles, reselleradmin_roel ..etc ? ) Tempest is from master Thank you very much, Kevko Michal Arbet Openstack Engineer Ultimum Technologies a.s. Na Po???? 1047/26, 11000 Praha 1 Czech Republic +420 604 228 897 michal.arbet at ultimum.io *https://ultimum.io * LinkedIn | Twitter | Facebook -------------- next part -------------- An HTML attachment was scrubbed... URL: From michal.arbet at ultimum.io Mon Dec 13 16:48:04 2021 From: michal.arbet at ultimum.io (Michal Arbet) Date: Mon, 13 Dec 2021 17:48:04 +0100 Subject: [tempest][swift][kolla] Message-ID: Hello to eveyrone, Please, could someone help me with swift capabilities not working when I'm tempesting openstack test stack ? Tempest : rm -rf /tmp/tempest-lock/; refstack-client test -v -c /opt/tempest/tempest.conf -- --regex tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest (.venv) root at ca6353106d83:/opt/refstack-client# rm -rf /tmp/tempest-lock/; refstack-client test -v -c /opt/tempest/tempest.conf -- --regex tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest 2021-12-13 17:44:33.001 4358 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf 2021-12-13 17:44:33,839 refstack_client:518 INFO Starting Tempest test... 2021-12-13 17:44:33.839 4358 INFO refstack_client [-] Starting Tempest test... {0} setUpClass (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) [0.000000s] ... FAILED Captured traceback: ~~~~~~~~~~~~~~~~~~~ Traceback (most recent call last): File "/opt/refstack-client/.tempest/tempest/test.py", line 181, in setUpClass raise value.with_traceback(trace) File "/opt/refstack-client/.tempest/tempest/test.py", line 174, in setUpClass cls.resource_setup() File "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", line 36, in resource_setup super(AccountQuotasNegativeTest, cls).resource_setup() File "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line 95, in resource_setup body = cls.capabilities_client.list_capabilities() File "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", line 32, in list_capabilities self._error_checker(resp, body) File "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line 799, in _error_checker raise exceptions.Unauthorized(resp_body, resp=resp) tempest.lib.exceptions.Unauthorized: Unauthorized Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The request you have made requires authentication.'} ============================== Failed 1 tests - output below: ============================== setUpClass (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) ---------------------------------------------------------------------------------------------- Captured traceback: ~~~~~~~~~~~~~~~~~~~ Traceback (most recent call last): File "/opt/refstack-client/.tempest/tempest/test.py", line 181, in setUpClass raise value.with_traceback(trace) File "/opt/refstack-client/.tempest/tempest/test.py", line 174, in setUpClass cls.resource_setup() File "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", line 36, in resource_setup super(AccountQuotasNegativeTest, cls).resource_setup() File "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line 95, in resource_setup body = cls.capabilities_client.list_capabilities() File "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", line 32, in list_capabilities self._error_checker(resp, body) File "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line 799, in _error_checker raise exceptions.Unauthorized(resp_body, resp=resp) tempest.lib.exceptions.Unauthorized: Unauthorized Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The request you have made requires authentication.'} ====== Totals ====== Ran: 1 tests in 0.0000 sec. - Passed: 0 - Skipped: 0 - Expected Fail: 0 - Unexpected Success: 0 - Failed: 1 Sum of execute time for each test: 0.0000 sec. ============== Worker Balance ============== - Worker 0 (1 tests) => 0:00:00 2021-12-13 17:44:43,144 refstack_client:581 INFO Tempest test complete. 2021-12-13 17:44:43.144 4358 INFO refstack_client [-] Tempest test complete. 2021-12-13 17:44:43,146 refstack_client:582 INFO Subunit results located in: /opt/refstack-client/.tempest/.stestr/137 2021-12-13 17:44:43.146 4358 INFO refstack_client [-] Subunit results located in: /opt/refstack-client/.tempest/.stestr/137 2021-12-13 17:44:43,148 refstack_client:585 INFO Number of passed tests: 0 2021-12-13 17:44:43.148 4358 INFO refstack_client [-] Number of passed tests: 0 2021-12-13 17:44:43,150 refstack_client:597 INFO JSON results saved in: /opt/refstack-client/.tempest/.stestr/137.json 2021-12-13 17:44:43.150 4358 INFO refstack_client [-] JSON results saved in: /opt/refstack-client/.tempest/.stestr/137.json TEMPEST LOG : Response - Headers: {'date': 'Mon, 13 Dec 2021 16:44:41 GMT', 'server': 'Apache', 'content-length': '8013', 'x-subject-token': '', 'vary': 'X-Auth-Token', 'x-openstack-request-id': 'req-00b11c98-7ffa-4a1a-88b1-a054fda5d7e9', 'content-type': 'application/json', 'connection': 'close', 'status': '201', 'content-location': 'http://192.168.205.254:35357/v3/auth/tokens'} Body: b'{"token": {"methods": ["password"], "user": {"domain": {"id": "default", "name": "Default"}, "id": "23ebe05900764afaa29f72519d40693a", "name": "refstack", "password_expires_at": null}, "audit_ids": ["4UgDYqQoSla7SWJOdjHXng"], "expires_at": "2021-12-14T16:44:42.000000Z", "issued_at": "2021-12-13T16:44:42.000000Z", "project": {"domain": {"id": "default", "name": "Default"}, "id": "704cf29c27d045199d11a11fe62206b6", "name": "refstack"}, "is_domain": false, "roles": [{"id": "89125ace2bc945af8f12d889c9f5589b", "name": "_member_"}, {"id": "47da66772503466881afc1ae391141e7", "name": "swiftoperator"}, {"id": "e47f85a3982f4a18a44d24d6c2ed6588", "name": "ResellerAdmin"}], "catalog": [{"endpoints": [{"id": "27daafd92ace480da6cf0fcaf4c054d7", "interface": "internal", "region_id": "RegionOne", "url": "http://192.168.205.254:8000/v1", "region": "RegionOne"}, {"id": "8ffc139d5a6c4d86b06557b01956dcb7", "interface": "admin", "region_id": "RegionOne", "url": " http://192.168.205.254:8000/v1", "region": "RegionOne"}, {"id": "e4b48ef6f9cd4bbea5f2e0fb9bf16af1", "interface": "public", "region_id": "RegionOne", "url": "https://api.refstack.ultimum.cloud:8000/v1", "region": "RegionOne"}], "id": "10b7ca67118d45dca197b8ac17f29349", "type": "cloudformation", "name": "heat-cfn"}, {"endpoints": [{"id": "06788107c3d94af9976b9fb4b3011d4f", "interface": "public", "region_id": "RegionOne", "url": "https://api.refstack.ultimum.cloud:5000", "region": "RegionOne"}, {"id": "0d745366b77546f7bd24259c04810b8b", "interface": "internal", "region_id": "RegionOne", "url": "http://192.168.205.254:5000", "region": "RegionOne"}, {"id": "1771c125743b4485b818a2aaa4d6f913", "interface": "admin", "region_id": "RegionOne", "url": " http://192.168.205.254:35357", "region": "RegionOne"}], "id": "1ee3179487894931a223ed3caaf8cdb5", "type": "identity", "name": "keystone"}, {"endpoints": [{"id": "1ce9a0af0c14454aa6043b5efe32dddb", "interface": "internal", "region_id": "RegionOne", "url": " http://192.168.205.254:9001", "region": "RegionOne"}, {"id": "5bcfbaa0af46423ab3de1e789affa187", "interface": "admin", "region_id": "RegionOne", "url": "http://192.168.205.254:9001", "region": "RegionOne"}, {"id": "cbc613e8806a48ac98cf4cc42cca80f2", "interface": "public", "region_id": "RegionOne", "url": "https://api.refstack.ultimum.cloud:9001", "region": "RegionOne"}], "id": "493ace4e817d4eae9aa4a1341bd01046", "type": "dns", "name": "designate"}, {"endpoints": [{"id": "18bea710b4284f5387081eac6833dc9e", "interface": "public", "region_id": "RegionOne", "url": " https://api.refstack.ultimum.cloud:8774/v2/704cf29c27d045199d11a11fe62206b6", "region": "RegionOne"}, {"id": "7f1bd52d8b154bc7973c0f32c57103f9", "interface": "internal", "region_id": "RegionOne", "url": " http://192.168.205.254:8774/v2/704cf29c27d045199d11a11fe62206b6", "region": "RegionOne"}, {"id": "c28543c29f574cc6b4320fba45394a83", "interface": "admin", "region_id": "RegionOne", "url": " http://192.168.205.254:8774/v2/704cf29c27d045199d11a11fe62206b6", "region": "RegionOne"}], "id": "52d1f52dc8f44e5ba7a40732a660a288", "type": "compute_legacy", "name": "nova_legacy"}, {"endpoints": [{"id": "33d722f6c2cf4865b3487b8c4ce96c14", "interface": "admin", "region_id": "RegionOne", "url": "http://192.168.205.254:9696", "region": "RegionOne"}, {"id": "4744e2aa77cb4ba186752fb0ef25e9d3", "interface": "internal", "region_id": "RegionOne", "url": "http://192.168.205.254:9696", "region": "RegionOne"}, {"id": "4a980da3677c429ebf645a0e1733949c", "interface": "public", "region_id": "RegionOne", "url": " https://api.refstack.ultimum.cloud:9696", "region": "RegionOne"}], "id": "56866ea275ec4f5e8f805a08f040c8e0", "type": "network", "name": "neutron"}, {"endpoints": [{"id": "4550e2b070aa4bd6a07ab00d72ef4266", "interface": "admin", "region_id": "RegionOne", "url": "http://192.168.205.254:8780", "region": "RegionOne"}, {"id": "86fc799751384f709c450db05d523d93", "interface": "public", "region_id": "RegionOne", "url": " https://api.refstack.ultimum.cloud:8780", "region": "RegionOne"}, {"id": "bfb36483e5dc40a5bfbf1b4ae6bf60cb", "interface": "internal", "region_id": _log_request_full /opt/refstack-client/.tempest/tempest/lib/common/rest_client.py:450 2021-12-13 17:44:42.789 4369 INFO tempest.lib.common.rest_client [req-4ee89585-59ff-4f5a-8ba3-9eebb1d5e6aa ] Request (AccountQuotasNegativeTest:setUpClass): 201 POST http://192.168.205.254:35357/v3/auth/tokens 0.548s 2021-12-13 17:44:42.790 4369 DEBUG tempest.lib.common.rest_client [req-4ee89585-59ff-4f5a-8ba3-9eebb1d5e6aa ] Request - Headers: {'Content-Type': 'application/json', 'Accept': 'application/json'} Body: Response - Headers: {'date': 'Mon, 13 Dec 2021 16:44:42 GMT', 'server': 'Apache', 'content-length': '8015', 'x-subject-token': '', 'vary': 'X-Auth-Token', 'x-openstack-request-id': 'req-4ee89585-59ff-4f5a-8ba3-9eebb1d5e6aa', 'content-type': 'application/json', 'connection': 'close', 'status': '201', 'content-location': 'http://192.168.205.254:35357/v3/auth/tokens'} Body: b'{"token": {"methods": ["password"], "user": {"domain": {"id": "default", "name": "Default"}, "id": "adcde758a0d5439ba6f8a1a35d641726", "name": "refstack2", "password_expires_at": null}, "audit_ids": ["0clvtk_FQQWVN-RRLSkCWg"], "expires_at": "2021-12-14T16:44:42.000000Z", "issued_at": "2021-12-13T16:44:42.000000Z", "project": {"domain": {"id": "default", "name": "Default"}, "id": "2aabe1b9860c44a08eabdc3133538089", "name": "refstack2"}, "is_domain": false, "roles": [{"id": "e47f85a3982f4a18a44d24d6c2ed6588", "name": "ResellerAdmin"}, {"id": "47da66772503466881afc1ae391141e7", "name": "swiftoperator"}, {"id": "89125ace2bc945af8f12d889c9f5589b", "name": "_member_"}], "catalog": [{"endpoints": [{"id": "27daafd92ace480da6cf0fcaf4c054d7", "interface": "internal", "region_id": "RegionOne", "url": "http://192.168.205.254:8000/v1", "region": "RegionOne"}, {"id": "8ffc139d5a6c4d86b06557b01956dcb7", "interface": "admin", "region_id": "RegionOne", "url": " http://192.168.205.254:8000/v1", "region": "RegionOne"}, {"id": "e4b48ef6f9cd4bbea5f2e0fb9bf16af1", "interface": "public", "region_id": "RegionOne", "url": "https://api.refstack.ultimum.cloud:8000/v1", "region": "RegionOne"}], "id": "10b7ca67118d45dca197b8ac17f29349", "type": "cloudformation", "name": "heat-cfn"}, {"endpoints": [{"id": "06788107c3d94af9976b9fb4b3011d4f", "interface": "public", "region_id": "RegionOne", "url": "https://api.refstack.ultimum.cloud:5000", "region": "RegionOne"}, {"id": "0d745366b77546f7bd24259c04810b8b", "interface": "internal", "region_id": "RegionOne", "url": "http://192.168.205.254:5000", "region": "RegionOne"}, {"id": "1771c125743b4485b818a2aaa4d6f913", "interface": "admin", "region_id": "RegionOne", "url": " http://192.168.205.254:35357", "region": "RegionOne"}], "id": "1ee3179487894931a223ed3caaf8cdb5", "type": "identity", "name": "keystone"}, {"endpoints": [{"id": "1ce9a0af0c14454aa6043b5efe32dddb", "interface": "internal", "region_id": "RegionOne", "url": " http://192.168.205.254:9001", "region": "RegionOne"}, {"id": "5bcfbaa0af46423ab3de1e789affa187", "interface": "admin", "region_id": "RegionOne", "url": "http://192.168.205.254:9001", "region": "RegionOne"}, {"id": "cbc613e8806a48ac98cf4cc42cca80f2", "interface": "public", "region_id": "RegionOne", "url": "https://api.refstack.ultimum.cloud:9001", "region": "RegionOne"}], "id": "493ace4e817d4eae9aa4a1341bd01046", "type": "dns", "name": "designate"}, {"endpoints": [{"id": "18bea710b4284f5387081eac6833dc9e", "interface": "public", "region_id": "RegionOne", "url": " https://api.refstack.ultimum.cloud:8774/v2/2aabe1b9860c44a08eabdc3133538089", "region": "RegionOne"}, {"id": "7f1bd52d8b154bc7973c0f32c57103f9", "interface": "internal", "region_id": "RegionOne", "url": " http://192.168.205.254:8774/v2/2aabe1b9860c44a08eabdc3133538089", "region": "RegionOne"}, {"id": "c28543c29f574cc6b4320fba45394a83", "interface": "admin", "region_id": "RegionOne", "url": " http://192.168.205.254:8774/v2/2aabe1b9860c44a08eabdc3133538089", "region": "RegionOne"}], "id": "52d1f52dc8f44e5ba7a40732a660a288", "type": "compute_legacy", "name": "nova_legacy"}, {"endpoints": [{"id": "33d722f6c2cf4865b3487b8c4ce96c14", "interface": "admin", "region_id": "RegionOne", "url": "http://192.168.205.254:9696", "region": "RegionOne"}, {"id": "4744e2aa77cb4ba186752fb0ef25e9d3", "interface": "internal", "region_id": "RegionOne", "url": "http://192.168.205.254:9696", "region": "RegionOne"}, {"id": "4a980da3677c429ebf645a0e1733949c", "interface": "public", "region_id": "RegionOne", "url": " https://api.refstack.ultimum.cloud:9696", "region": "RegionOne"}], "id": "56866ea275ec4f5e8f805a08f040c8e0", "type": "network", "name": "neutron"}, {"endpoints": [{"id": "4550e2b070aa4bd6a07ab00d72ef4266", "interface": "admin", "region_id": "RegionOne", "url": "http://192.168.205.254:8780", "region": "RegionOne"}, {"id": "86fc799751384f709c450db05d523d93", "interface": "public", "region_id": "RegionOne", "url": " https://api.refstack.ultimum.cloud:8780", "region": "RegionOne"}, {"id": "bfb36483e5dc40a5bfbf1b4ae6bf60cb", "interface": "internal", "region_id _log_request_full /opt/refstack-client/.tempest/tempest/lib/common/rest_client.py:450 2021-12-13 17:44:42.849 4369 INFO tempest.lib.common.rest_client [tx815e181e33fb4854b2631-0061b7787a ] Request (AccountQuotasNegativeTest:setUpClass): 401 GET https://api.refstack.ultimum.cloud:8080/info 0.049s 2021-12-13 17:44:42.849 4369 DEBUG tempest.lib.common.rest_client [tx815e181e33fb4854b2631-0061b7787a ] Request - Headers: {'Content-Type': 'application/json', 'Accept': 'application/json'} Body: None Response - Headers: {'content-type': 'application/json', 'content-length': '114', 'www-authenticate': 'Keystone uri=" http://192.168.205.254:5000"', 'x-trans-id': 'tx815e181e33fb4854b2631-0061b7787a', 'x-openstack-request-id': 'tx815e181e33fb4854b2631-0061b7787a', 'date': 'Mon, 13 Dec 2021 16:44:42 GMT', 'connection': 'close', 'status': '401', 'content-location': ' https://api.refstack.ultimum.cloud:8080/info'} Body: b'{"error": {"code": 401, "title": "Unauthorized", "message": "The request you have made requires authentication."}}' _log_request_full /opt/refstack-client/.tempest/tempest/lib/common/rest_client.py:450 Test from command line and from curl . /etc/kolla/refstack.sh ; curl -H "X-Auth-Token: $(openstack token issue -f value -c id)" https://api.refstack.ultimum.cloud:8080/info /usr/lib/python3/dist-packages/secretstorage/dhcrypto.py:15: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead from cryptography.utils import int_from_bytes /usr/lib/python3/dist-packages/secretstorage/util.py:19: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead from cryptography.utils import int_from_bytes {"swift": {"version": "2.27.1.dev9", "strict_cors_mode": true, "policies": [{"name": "Policy-0", "aliases": "Policy-0", "default": true}], "allow_account_management": true, "account_autocreate": true, "max_file_size": 5368709122, "max_meta_name_length": 128, "max_meta_value_length": 256, "max_meta_count": 90, "max_meta_overall_size": 4096, "max_header_size": 8192, "max_object_name_length": 1024, "container_listing_limit": 10000, "account_listing_limit": 10000, "max_account_name_length": 256, "max_container_name_length": 256, "extra_header_count": 0}, "container_sync": {"realms": {}}, "bulk_upload": {"max_containers_per_extraction": 10000, "max_failed_extractions": 1000}, "bulk_delete": {"max_deletes_per_request": 10000, "max_failed_deletes": 1000}, "tempurl": {"methods": ["GET", "HEAD", "PUT", "POST", "DELETE"], "incoming_remove_headers": ["x-timestamp"], "incoming_allow_headers": [], "outgoing_remove_headers": ["x-object-meta-*"], "outgoing_allow_headers": ["x-object-meta-public-*"], "allowed_digests": ["sha1", "sha256", "sha512"]}, "ratelimit": {"account_ratelimit": 0.0, "max_sleep_time_seconds": 60.0, "container_ratelimits": [], "container_listing_ratelimits": []}, "container_quotas": {}, "account_quotas": {}, "slo": {"max_manifest_segments": 1000, "max_manifest_size": 8388608, "yield_frequency": 10, "min_segment_size": 1, "allow_async_delete": false}} Swiftclient : ubuntu at deploy:/opt/kolla-ansible$ . /etc/kolla/refstack.sh ; swift --os-auth-url http://192.168.205.254:5000/v3 --auth-version 3 --os-project-name refstack --os-project-domain-name default --os-username refstack --os-user-domain-name default --os-password refstack capabilities /usr/lib/python3/dist-packages/secretstorage/dhcrypto.py:15: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead from cryptography.utils import int_from_bytes /usr/lib/python3/dist-packages/secretstorage/util.py:19: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead from cryptography.utils import int_from_bytes Capabilities GET failed: http://192.168.205.254:8080/info 401 Unauthorized [first 60 chars of response] b'{"error": {"code": 401, "title": "Unauthorized", "message": ' Failed Transaction ID: txc1d8607e26eb4cd587459-0061b7791b It looks like swift client is broken, isn't it ? Thank you very much, Kevko Michal Arbet Openstack Engineer Ultimum Technologies a.s. Na Po???? 1047/26, 11000 Praha 1 Czech Republic +420 604 228 897 michal.arbet at ultimum.io *https://ultimum.io * LinkedIn | Twitter | Facebook -------------- next part -------------- An HTML attachment was scrubbed... URL: From andr.kurilin at gmail.com Mon Dec 13 19:44:47 2021 From: andr.kurilin at gmail.com (Andrey Kurilin) Date: Mon, 13 Dec 2021 21:44:47 +0200 Subject: [all] testtools and support for Python 3.10 (error du to deprecation of distutils) In-Reply-To: <20211213164815.ftio25ectty7qb2t@yuggoth.org> References: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> <20211213164815.ftio25ectty7qb2t@yuggoth.org> Message-ID: Hi folks! ??, 13 ???. 2021 ?. ? 18:53, Jeremy Stanley : > On 2021-12-13 17:24:23 +0100 (+0100), Thomas Goirand wrote: > > When I try to build rally-openstack, I get this, in the test discovery: > [...] > > /usr/lib/python3.10/distutils/__init__.py:19: in > > warnings.warn(_DEPRECATION_MESSAGE, > > E DeprecationWarning: The distutils package is deprecated and slated > > for removal in Python 3.12. Use setuptools or check PEP 632 for > > potential alternatives > [...] > > Can someone work on removing distutils support from Testtools? Or is > > there a way to trick pytest to not fail? Help would be greatly > > appreciated here. > > PYTHONWARNINGS must be getting set to "error" somewhere. Are you > setting it at invocation maybe? It is set via tox.ini at https://github.com/openstack/rally-openstack/blob/2.2.0/tox.ini#L122-L124 It is better to fix deprecated things earlier before they result in the broken project :) > In order to be able to run with > dependencies that call distutils I currently do something like this: > > PYTHONWARNINGS = error, ignore:The distutils package is > deprecated and slated for removal in Python 3.12. Use > setuptools or check PEP 632 for potential > alternatives:DeprecationWarning > > I'm ok to ignore the warning from testtools since it doesn't require any changes from our side. All on one line of course. And yes, ideally someone should fix > testtools before 3.11 releases, so people can still use it on 3.12 > alpha releases once those are available. > -- > Jeremy Stanley > -- Best regards, Andrey Kurilin. -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Mon Dec 13 21:08:48 2021 From: zigo at debian.org (Thomas Goirand) Date: Mon, 13 Dec 2021 22:08:48 +0100 Subject: [all] testtools and support for Python 3.10 (error du to deprecation of distutils) In-Reply-To: References: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> <20211213164815.ftio25ectty7qb2t@yuggoth.org> Message-ID: <8ad35b9d-b0c7-5d8c-77f6-de780ed59077@debian.org> On 12/13/21 8:44 PM, Andrey Kurilin wrote: > Hi folks! > > ??, 13 ???. 2021 ?. ? 18:53, Jeremy Stanley >: > > On 2021-12-13 17:24:23 +0100 (+0100), Thomas Goirand wrote: > > When I try to build rally-openstack, I get this, in the test > discovery: > [...] > > /usr/lib/python3.10/distutils/__init__.py:19: in > >? ? ?warnings.warn(_DEPRECATION_MESSAGE, > > E? ?DeprecationWarning: The distutils package is deprecated and slated > > for removal in Python 3.12. Use setuptools or check PEP 632 for > > potential alternatives > [...] > > Can someone work on removing distutils support from Testtools? Or is > > there a way to trick pytest to not fail? Help would be greatly > > appreciated here. > > PYTHONWARNINGS must be getting set to "error" somewhere. Are you > setting it at invocation maybe? > > > It is set via tox.ini at > https://github.com/openstack/rally-openstack/blob/2.2.0/tox.ini#L122-L124 > > It is better to fix deprecated things earlier before they result in the > broken project :) > ? > > In order to be able to run with > dependencies that call distutils I currently do something like this: > > ? ? PYTHONWARNINGS = error, ignore:The distutils package is > ? ? ? ? deprecated and slated for removal in Python 3.12. Use > ? ? ? ? setuptools or check PEP 632 for potential > ? ? ? ? alternatives:DeprecationWarning > > > I'm ok to ignore the warning from testtools since it doesn't require any > changes from our side. > > All on one line of course. And yes, ideally someone should fix > testtools before 3.11 releases, so people can still use it on 3.12 > alpha releases once those are available. > -- > Jeremy Stanley > > > > -- > Best regards, > Andrey Kurilin. Again, this happens during the *test discovery*. I'm not using Tox (package maintainer never do). FYI, I'm simply doing: python3 -m pytest tests/unit If you know how to disable the errors on warning during test discovery, let me know. But right now, this equals to FTBFS for rally-openstack in Debian (the package was removed from testing because of it...). Cheers, Thomas Goirand (zigo) From fungi at yuggoth.org Mon Dec 13 21:14:10 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Mon, 13 Dec 2021 21:14:10 +0000 Subject: [all] testtools and support for Python 3.10 (error du to deprecation of distutils) In-Reply-To: <8ad35b9d-b0c7-5d8c-77f6-de780ed59077@debian.org> References: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> <20211213164815.ftio25ectty7qb2t@yuggoth.org> <8ad35b9d-b0c7-5d8c-77f6-de780ed59077@debian.org> Message-ID: <20211213211410.taznfxueot7yxcuw@yuggoth.org> On 2021-12-13 22:08:48 +0100 (+0100), Thomas Goirand wrote: > On 12/13/21 8:44 PM, Andrey Kurilin wrote: > > Hi folks! > > > > ??, 13 ???. 2021 ?. ? 18:53, Jeremy Stanley > >: > > > > On 2021-12-13 17:24:23 +0100 (+0100), Thomas Goirand wrote: > > > When I try to build rally-openstack, I get this, in the test > > discovery: > > [...] > > > /usr/lib/python3.10/distutils/__init__.py:19: in > > >? ? ?warnings.warn(_DEPRECATION_MESSAGE, > > > E? ?DeprecationWarning: The distutils package is deprecated and slated > > > for removal in Python 3.12. Use setuptools or check PEP 632 for > > > potential alternatives > > [...] > > > Can someone work on removing distutils support from Testtools? Or is > > > there a way to trick pytest to not fail? Help would be greatly > > > appreciated here. > > > > PYTHONWARNINGS must be getting set to "error" somewhere. Are you > > setting it at invocation maybe? > > > > > > It is set via tox.ini at > > https://github.com/openstack/rally-openstack/blob/2.2.0/tox.ini#L122-L124 > > > > It is better to fix deprecated things earlier before they result in the > > broken project :) [...] > Again, this happens during the *test discovery*. I'm not using Tox > (package maintainer never do). FYI, I'm simply doing: > > python3 -m pytest tests/unit > > If you know how to disable the errors on warning during test discovery, > let me know. But right now, this equals to FTBFS for rally-openstack in > Debian (the package was removed from testing because of it...). To reiterate, PYTHONWARNINGS must be getting set to "error" somewhere. Maybe it's exported in that environment before python3 is invoked? -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From gagehugo at gmail.com Mon Dec 13 21:17:10 2021 From: gagehugo at gmail.com (Gage Hugo) Date: Mon, 13 Dec 2021 15:17:10 -0600 Subject: [security][security sig] Retiring the security-specs repo Message-ID: Hey folks, The security SIG has been going through the current list of repositories that we are managing and one of them, the "security-specs"[0] repository, has been unused for several years now so the consensus was to retire the repo. If there are no objections, we will look to retire the repo here relatively soon. Should anyone have any other information about it, please let the security SIG know. Thanks! [0] https://github.com/openstack/security-specs -------------- next part -------------- An HTML attachment was scrubbed... URL: From jlabarre at redhat.com Mon Dec 13 21:43:19 2021 From: jlabarre at redhat.com (James LaBarre) Date: Mon, 13 Dec 2021 16:43:19 -0500 Subject: [tempest] Extending tempest for mixed-architecture stacks Message-ID: <8f8768fa-3097-6bc2-39ff-f87eb6f91ba1@redhat.com> Recently I had been running Tempest on my setup testing a mixed-architecture deployment (x86_64 ans ppc64le compute nodes at the same time).? It seems that some of the migration and affinity tests will check if there's more than one Compute node before they run.? However, it would seem that's as far as they check, without checking if they are in fact compatible or even of the same architecture.? (my test cluster is very small, and normally includes two ppc64le Compute nodes, and sometimes one x86_64 Compute node.? Currently one ppc64le machine is down for repair). Because the two compute nodes are different architectures, I am getting failures in various migration and affinity tests, maybe more if I tested a larger subset.? Now granted my particular setup is a special case, but it does bring to mind some extensions that may be needed for Tempest in the future.? I could see it being possible to have x86_64 and ARM mixed together in one stack, maybe even tossing in RISC-v someday. I'm thinking we need to start adding in extra test images, flavors, etc into the Tempest configurations (as in defining multiple options so that each architecture can have test images, etc assigned to it, rather than the current primary and alt image for just one architecture)? Additionally, there should be testcases taking into account the architectures involved (such as seeing that an instance on one arch cannot be migrated to the other, as an example).? I know this involves a bit of refactoring, I didn't know if it had even been considered yet. -- James LaBarre Software Engineer, OpenStack MultiArch Red Hat jlabarre at redhat.com -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Mon Dec 13 22:29:05 2021 From: zigo at debian.org (Thomas Goirand) Date: Mon, 13 Dec 2021 23:29:05 +0100 Subject: [all] testtools and support for Python 3.10 (error du to deprecation of distutils) In-Reply-To: <20211213211410.taznfxueot7yxcuw@yuggoth.org> References: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> <20211213164815.ftio25ectty7qb2t@yuggoth.org> <8ad35b9d-b0c7-5d8c-77f6-de780ed59077@debian.org> <20211213211410.taznfxueot7yxcuw@yuggoth.org> Message-ID: <61ee101e-e627-eac1-7c05-4765d8041b67@debian.org> On 12/13/21 10:14 PM, Jeremy Stanley wrote: > To reiterate, PYTHONWARNINGS must be getting set to "error" > somewhere. Maybe it's exported in that environment before python3 is > invoked? > This returns nothing: grep -r PYTHONWARNINGS * So no, it's not what's happening, unless I don't understand... Cheers, Thomas Goirand (zigo) From fungi at yuggoth.org Mon Dec 13 22:34:58 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Mon, 13 Dec 2021 22:34:58 +0000 Subject: [all] testtools and support for Python 3.10 (error du to deprecation of distutils) In-Reply-To: <61ee101e-e627-eac1-7c05-4765d8041b67@debian.org> References: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> <20211213164815.ftio25ectty7qb2t@yuggoth.org> <8ad35b9d-b0c7-5d8c-77f6-de780ed59077@debian.org> <20211213211410.taznfxueot7yxcuw@yuggoth.org> <61ee101e-e627-eac1-7c05-4765d8041b67@debian.org> Message-ID: <20211213223457.wz6r2p5zaylptxic@yuggoth.org> On 2021-12-13 23:29:05 +0100 (+0100), Thomas Goirand wrote: > On 12/13/21 10:14 PM, Jeremy Stanley wrote: > > To reiterate, PYTHONWARNINGS must be getting set to "error" > > somewhere. Maybe it's exported in that environment before python3 is > > invoked? > > > > This returns nothing: > grep -r PYTHONWARNINGS * > > So no, it's not what's happening, unless I don't understand... Are you able to reproduce it locally? Under sbuild or some other way? -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From zigo at debian.org Mon Dec 13 22:50:40 2021 From: zigo at debian.org (Thomas Goirand) Date: Mon, 13 Dec 2021 23:50:40 +0100 Subject: [all] testtools and support for Python 3.10 (error du to deprecation of distutils) In-Reply-To: <20211213223457.wz6r2p5zaylptxic@yuggoth.org> References: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> <20211213164815.ftio25ectty7qb2t@yuggoth.org> <8ad35b9d-b0c7-5d8c-77f6-de780ed59077@debian.org> <20211213211410.taznfxueot7yxcuw@yuggoth.org> <61ee101e-e627-eac1-7c05-4765d8041b67@debian.org> <20211213223457.wz6r2p5zaylptxic@yuggoth.org> Message-ID: On 12/13/21 11:34 PM, Jeremy Stanley wrote: > On 2021-12-13 23:29:05 +0100 (+0100), Thomas Goirand wrote: >> On 12/13/21 10:14 PM, Jeremy Stanley wrote: >>> To reiterate, PYTHONWARNINGS must be getting set to "error" >>> somewhere. Maybe it's exported in that environment before python3 is >>> invoked? >>> >> >> This returns nothing: >> grep -r PYTHONWARNINGS * >> >> So no, it's not what's happening, unless I don't understand... > > Are you able to reproduce it locally? Under sbuild or some other > way? > Yep. Doing this: git clone https://salsa.debian.org/openstack-team/services/rally-openstack.git cd rally-openstack sbuild ... Cheers, Thomas Goirand (zigo) From gmann at ghanshyammann.com Mon Dec 13 23:52:50 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Mon, 13 Dec 2021 17:52:50 -0600 Subject: [swift][tempest][kolla] In-Reply-To: References: Message-ID: <17db63694c8.12444de92127096.1140393958066839556@ghanshyammann.com> ---- On Mon, 13 Dec 2021 11:33:27 -0600 Michal Arbet wrote ---- > Hello to eveyrone, > Please, could someone help me with swift capabilities not working when I'm tempesting openstack test stack ? > Tempest : > rm -rf /tmp/tempest-lock/; refstack-client test -v -c /opt/tempest/tempest.conf -- --regex tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest > > (.venv) root at ca6353106d83:/opt/refstack-client# rm -rf /tmp/tempest-lock/; refstack-client test -v -c /opt/tempest/tempest.conf -- --regex tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest > 2021-12-13 17:44:33.001 4358 INFO tempest [-] Using tempest config file /etc/tempest/tempest.conf > 2021-12-13 17:44:33,839 refstack_client:518 INFO Starting Tempest test... > 2021-12-13 17:44:33.839 4358 INFO refstack_client [-] Starting Tempest test... > {0} setUpClass (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) [0.000000s] ... FAILED > > Captured traceback: > ~~~~~~~~~~~~~~~~~~~ > Traceback (most recent call last): > > File "/opt/refstack-client/.tempest/tempest/test.py", line 181, in setUpClass > raise value.with_traceback(trace) > > File "/opt/refstack-client/.tempest/tempest/test.py", line 174, in setUpClass > cls.resource_setup() > > File "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", line 36, in resource_setup > super(AccountQuotasNegativeTest, cls).resource_setup() > > File "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line 95, in resource_setup > body = cls.capabilities_client.list_capabilities() > > File "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", line 32, in list_capabilities > self._error_checker(resp, body) > > File "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line 799, in _error_checker > raise exceptions.Unauthorized(resp_body, resp=resp) > > tempest.lib.exceptions.Unauthorized: Unauthorized > Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The request you have made requires authentication.'} > > > ============================== > Failed 1 tests - output below: > ============================== > > setUpClass (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) > ---------------------------------------------------------------------------------------------- > > Captured traceback: > ~~~~~~~~~~~~~~~~~~~ > Traceback (most recent call last): > > File "/opt/refstack-client/.tempest/tempest/test.py", line 181, in setUpClass > raise value.with_traceback(trace) > > File "/opt/refstack-client/.tempest/tempest/test.py", line 174, in setUpClass > cls.resource_setup() > > File "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", line 36, in resource_setup > super(AccountQuotasNegativeTest, cls).resource_setup() > > File "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line 95, in resource_setup > body = cls.capabilities_client.list_capabilities() > > File "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", line 32, in list_capabilities > self._error_checker(resp, body) > > File "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line 799, in _error_checker > raise exceptions.Unauthorized(resp_body, resp=resp) > > tempest.lib.exceptions.Unauthorized: Unauthorized > Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The request you have made requires authentication.'} > > > > ====== > Totals > ====== > Ran: 1 tests in 0.0000 sec. > - Passed: 0 > - Skipped: 0 > - Expected Fail: 0 > - Unexpected Success: 0 > - Failed: 1 > Sum of execute time for each test: 0.0000 sec. > > TEMPEST LOG : > Response - Headers: {'content-type': 'application/json', 'content-length': '114', 'www-authenticate': 'Keystone uri="http://192.168.205.254:5000"', 'x-trans-id': 'tx815e181e33fb4854b2631-0061b7787a', 'x-openstack-request-id': 'tx815e181e33fb4854b2631-0061b7787a', 'date': 'Mon, 13 Dec 2021 16:44:42 GMT', 'connection': 'close', 'status': '401', 'content-location': 'https://api.refstack.ultimum.cloud:8080/info'} > Body: b'{"error": {"code": 401, "title": "Unauthorized", "message": "The request you have made requires authentication."}}' _log_request_full /opt/refstack-client/.tempest/tempest/lib/common/rest_client.py:450 > > > > Test from command line and from curl > . /etc/kolla/refstack.sh ; curl -H "X-Auth-Token: $(openstack token issue -f value -c id)" https://api.refstack.ultimum.cloud:8080/info > {"swift": {"version": "2.27.1.dev9", "strict_cors_mode": true, "policies": [{"name": "Policy-0", "aliases": "Policy-0", "default": true}], "allow_account_management": true, "account_autocreate": true, "max_file_size": 5368709122, "max_meta_name_length": 128, "max_meta_value_length": 256, "max_meta_count": 90, "max_meta_overall_size": 4096, "max_header_size": 8192, "max_object_name_length": 1024, "container_listing_limit": 10000, "account_listing_limit": 10000, "max_account_name_length": 256, "max_container_name_length": 256, "extra_header_count": 0}, "container_sync": {"realms": {}}, "bulk_upload": {"max_containers_per_extraction": 10000, "max_failed_extractions": 1000}, "bulk_delete": {"max_deletes_per_request": 10000, "max_failed_deletes": 1000}, "tempurl": {"methods": ["GET", "HEAD", "PUT", "POST", "DELETE"], "incoming_remove_headers": ["x-timestamp"], "incoming_allow_headers": [], "outgoing_remove_headers": ["x-object-meta-*"], "outgoing_allow_headers": ["x-object-meta-public-*"], "allowed_digests": ["sha1", "sha256", "sha512"]}, "ratelimit": {"account_ratelimit": 0.0, "max_sleep_time_seconds": 60.0, "container_ratelimits": [], "container_listing_ratelimits": []}, "container_quotas": {}, "account_quotas": {}, "slo": {"max_manifest_segments": 1000, "max_manifest_size": 8388608, "yield_frequency": 10, "min_segment_size": 1, "allow_async_delete": false}} > > Python Swiftclient : > ubuntu at deploy:/opt/kolla-ansible$ . /etc/kolla/refstack.sh ; swift --os-auth-url http://192.168.205.254:5000/v3 --auth-version 3 --os-project-name refstack --os-project-domain-name default --os-username refstack --os-user-domain-name default --os-password SECRET capabilities > Capabilities GET failed: http://192.168.205.254:8080/info 401 Unauthorized [first 60 chars of response] b'{"error": {"code": 401, "title": "Unauthorized", "message": ' > Failed Transaction ID: txc1d8607e26eb4cd587459-0061b7791b > > > > It looks like swift client is broken, isn't it ? Or ? Maybe kolla-ansible is creating bad roles and config ? (operator_roles, reselleradmin_roel ..etc ? ) Tempest is from master > Thank you very much,Kevko Are you running it with dynamic creds or pre-provisioned creds ? Error is from cls.capabilities_client which is initialized from CONF.object_storage.operator_role[1] which is 'member' role by default. what is your configuration for this? With default CONF.object_storage.operator_role as 'member' role, this test pass in upstream CI/CD so client is not broken but it is configuration issue: https://zuul.opendev.org/t/openstack/build/b29147647370418fb9fbb0182832749d/log/job-output.txt#27292 [1] https://github.com/openstack/tempest/blob/34432dc970d09a55572a68fa007575285e35b550/tempest/api/object_storage/base.py#L77 -gmann > > > Michal Arbet > Openstack Engineer > > Ultimum Technologies a.s. > Na Po???? 1047/26, 11000 Praha 1 > Czech Republic > > +420 604 228 897 > michal.arbet at ultimum.io > https://ultimum.io > > LinkedIn | Twitter | Facebook From gmann at ghanshyammann.com Mon Dec 13 23:54:46 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Mon, 13 Dec 2021 17:54:46 -0600 Subject: [all][tc] Technical Committee next weekly meeting on Dec 16th at 1500 UTC Message-ID: <17db6385d36.ca951ea5127116.4798485549554257708@ghanshyammann.com> Hello Everyone, Technical Committee's next weekly meeting is scheduled for Dec 16th at 1500 UTC. If you would like to add topics for discussion, please add them to the below wiki page by Wednesday, Dec 15th, at 2100 UTC. https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting -gmann From xiaolihope1008 at 163.com Tue Dec 14 02:10:19 2021 From: xiaolihope1008 at 163.com (Di XiaoLi) Date: Tue, 14 Dec 2021 10:10:19 +0800 (GMT+08:00) Subject: [cyborg][nova][qa]Question about Accelerator unbinding In-Reply-To: <99456fa20ca2a93b434608cf313f0047ace47abf.camel@redhat.com> References: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> <99456fa20ca2a93b434608cf313f0047ace47abf.camel@redhat.com> Message-ID: <7f1e8b3.614a.17db6b47675.Coremail.xiaolihope1008@163.com> Well,I see. Thanks to Sean for your detailed answer! On 12/14/2021 00:33?Sean Mooney wrote? On Mon, 2021-12-13 at 09:54 +0800, Di XiaoLi wrote: hi, Cyborg and nova team: I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. Here are my questions: 1. What is the function of the arq unbind command in cyborg ? it has two usecases. it call by nova when nova is deleting or moving the vm it can be used by an enduser if they are not using cybrog with nova. 2. How to unbind the accelerator which bounded to vm? Does nova or cyborg support this function now? the only way to do that for a nova instance is to resize it to a flavor that does not request the device via the cyborg device profiel in the extra spec. more recently we have started to support using cybrog for neutron nics too. in this case the unbinding can be done by doing a port detach and the device should be remvoed from the vm and unbond in cybrog. Here are my steps: step1: openstack accelerator arq unbind ed205084-f58d-4fad-99b4-327a1398858f this should be treated as an admin/service only oepration wehn using cybrog with nova more on that below. +---------------------+--------------------------------------+ Field | Value | +---------------------+--------------------------------------+ uuid | ed205084-f58d-4fad-99b4-327a1398858f | state | Unbound | device_profile_name | ssd | hostname | None | device_rp_uuid | None | instance_uuid | None | attach_handle_type | | attach_handle_info | {} | +---------------------+--------------------------------------+ step2: login vm and check the device, but it still here. step3: stop vm and start vm, met the following error: "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" that is what i would expect and is the correct behavior as you violated the contract between nova and cyborg by unbindng the arq. when useing cyborg with nova you shoudl never use the cyborg api driectly expcit to list the devcice profiles. you should treat it like placment in that regard. cybrog when used with nova today is an internal service that end user should at most have read only access too list the profiles. nova does not support cybrog device hot/cold attach or detach. the only way to add or remove a device form cyborg via nova is the device profile request in the flavor. so the only support opartion to change the attach device is resize to a differnt flavor. as i said above if you are attching smartnics using cyborg via neutron then you can also attach/devatch cybrog device by attaching/detaching the neutron port which contains the device profile request. note that if the deivce does not exist on the currnt host the attch will likely fail. detach should more or less always work unless there is an internal errror with one of the 3 service invovled. -------------- next part -------------- An HTML attachment was scrubbed... URL: From gagehugo at gmail.com Tue Dec 14 05:31:47 2021 From: gagehugo at gmail.com (Gage Hugo) Date: Mon, 13 Dec 2021 23:31:47 -0600 Subject: [openstack-helm] No Meetings Rest of 2021 Message-ID: Hey team, Due to the upcoming holidays and the lack of topics for tomorrow's meeting the next 3 meetings are canceled. We will start back up in the new year on Jan 4th at the usual time. Happy Holidays! -------------- next part -------------- An HTML attachment was scrubbed... URL: From chkumar at redhat.com Tue Dec 14 08:44:23 2021 From: chkumar at redhat.com (Chandan Kumar) Date: Tue, 14 Dec 2021 14:14:23 +0530 Subject: [tripleo] Initial plans around stable/wallaby branch on CentOS Stream 9 In-Reply-To: References: Message-ID: Hello, On Mon, Dec 13, 2021 at 4:59 PM Ronelle Landy wrote: > > Below are the two patches which need review. * https://review.opendev.org/c/openstack/tripleo-ansible/+/821663 * https://review.opendev.org/c/openstack/tripleo-common/+/821511 For tripleo-heat-templates and python-tripleoclient all the cs9 patches are backported to stable/wallaby branch. I will look into other projects which need backport. Thanks, Chandan Kumar From faisalsheikh.cyber at gmail.com Tue Dec 14 09:31:37 2021 From: faisalsheikh.cyber at gmail.com (Faisal Sheikh) Date: Tue, 14 Dec 2021 14:31:37 +0500 Subject: [wallaby][ovn][Open vSwitch] HA for OVN DB servers using pacemaker Message-ID: Hi, I am using Openstack Wallaby release with OVN on Ubuntu 20.04. My environment consists of 2 compute nodes and 1 controller node. ovs-vswitchd (Open vSwitch) 2.15.0 Ubuntu Kernel Version: 5.4.0-88-generic - compute node1 172.16.30.1 - compute node2 172.16.30.3 - controller-khi01/Network node/Primary ovs-db IP 172.16.30.46 - backup ovs-db 172.16.30.47 I want to run OVSDB server to run in Active/Passive mode. Active OVSDB-server is on node "controller-khi01" and Passive OVSDB-server is on "controller02-khi01". I have setup Pacemaker to manage the ovn-northd(Open vSwitch) service between primary and backup ovn-db server. You can see the "pcs status" output below. But I am unable to replicate database state between primary and backup OVSDB server. If i execute "ovsdb-server --sync-from=server" command, getting below error. root at controller-khi01# ovsdb-server --sync-from=172.16.30.47 2021-12-14T09:16:29Z|00001|lockfile|WARN|/var/lib/openvswitch/.conf.db.~lock~: cannot lock file because it is already locked by pid 14121 ovsdb-server: I/O error: /etc/openvswitch/conf.db: failed to lock lockfile (Resource temporarily unavailable) root at controller-khi01# pcs status Cluster name: cluster1 Cluster Summary: * Stack: corosync * Current DC: controller-khi01 (version 2.0.3-4b1f869f0f) - partition with quorum * Last updated: Tue Dec 14 07:53:06 2021 * Last change: Tue Dec 14 06:16:21 2021 by hacluster via crmd on controller02-khi01 * 2 nodes configured * 3 resource instances configured Node List: * Online: [ controller-khi01 controller02-khi01 ] Full List of Resources: * Clone Set: ovndb_servers-clone [ovndb_servers] (promotable): * Masters: [ controller-khi01 ] * Slaves: [ controller02-khi01 ] * ovn-virtual-ip (ocf::heartbeat:IPaddr2): Started controller-khi01 Failed Resource Actions: * ovndb_servers_monitor_10000 on controller02-khi01 'master' (8): call=17, status='complete', exitreason='', last-rc-change='2021-12-14 06:48:26Z', queued=0ms, exec=0ms Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled I would really appreciate any input in this regard. Best regards, Faisal Sheikh -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Tue Dec 14 11:06:11 2021 From: zigo at debian.org (Thomas Goirand) Date: Tue, 14 Dec 2021 12:06:11 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: References: <9a3cf269-6b89-fb32-6dc5-4fd708171cb2@inovex.de> Message-ID: <4fc29f97-c82f-7156-7876-7a9aef15d7ed@debian.org> On 12/12/21 6:38 PM, Cedric Lemarchand wrote: > So I validated what from my point of view seems to be the most > "supported" way to boot an Openstack instance in PXE : > > - add the flavor the extra specs "hw:boot_menu = 1" (thanks Sean) > - spawn the instance > - configure the extra_dhcp_opts of the associated port (thanks > Christian). Its not supported with the CLI, but doable with an sdk, > below a Pythonic example [1] > - reboot the instance, hit esc at the bios prompt then third choice (iPXE) > > This allows the use of the Openstack DHCP and to keep port security > enabled, which is nice. > > Cheers > > [1] > --- > conn.update_port(port_id, extra_dhcp_opts=[ > {"opt_name": "server-ip-address","opt_value": "10.0.0.15"}, > {"opt_name": "bootfile-name", "opt_value": "pxelinux.0"}, > {"opt_name": "tftp-server", "opt_value": "10.0.0.15"} > ] > ) Hi, It really is supported by the cli. I've done this way: openstack port create ${MACHINE} --network oci-boot-network \ --extra-dhcp-option \ name=tftp-server,value=$,ip-version=4 \ --extra-dhcp-option \ name=bootfile-name,value=http:///oci/ipxe.php,ip-version=4 \ --fixed-ip subnet=subname,ip-address= At the present moment, I have a working OCI [1] functional CI that works with non-admin OpenStack credentials (just a normal user). I hope to be able to share this very soon, so that anyone can try without too much hassle. Cheers, Thomas Goirand (zigo) [1] https://salsa.debian.org/openstack-team/debian/openstack-cluster-installer From amotoki at gmail.com Tue Dec 14 13:37:13 2021 From: amotoki at gmail.com (Akihiro Motoki) Date: Tue, 14 Dec 2021 22:37:13 +0900 Subject: bug deputy report (week of Dec 6) Message-ID: Hi, I was a bug deputy last week. Here is a summary of last week. Two needs-attention bugs need what neutron should behave, so they need more investigation. Two OVN bugs and one gate-failure bug also need assignees and investigations. Thanks, Akihiro Motoki (amotoki) # Needs attention - https://bugs.launchpad.net/neutron/+bug/1953478 Resize and shelve server fails in the multinode CI jobs intemittent It was reported to nova by slawek, but gibi moved it to neutron as there is a question on when neutron send vif-plugged event to nova. We need to clarify when the vif-plugged event is sent and what is the expected behavior. - https://bugs.launchpad.net/neutron/+bug/1954384 Neutron doesn't use set dns-domain for Nova instance It would be nice if folks familiar with DNS integration can look into it. # OVN https://bugs.launchpad.net/neutron/+bug/1953510 neutron_ovn_metadata_agent doesn`t work after reconnect to new OVN Southbound leader DB https://bugs.launchpad.net/neutron/+bug/1953295 neutron-ovn-metadata-agent AttributeError: 'MetadataAgent' object has no attribute 'sb_idl' # Gate failures https://bugs.launchpad.net/neutron/+bug/1953479 Timeout in the scenario jobs' execution # Assigned https://bugs.launchpad.net/neutron/+bug/1953716 Improve message in the NetworkInUse exception # In Progress https://bugs.launchpad.net/neutron/+bug/1954316 In VXLAN interface creation, we must provide always a device https://bugs.launchpad.net/neutron/+bug/1954662 Quota driver "DbQuotaNoLockDriver" can lock when removing the expired reservations https://bugs.launchpad.net/neutron/+bug/1953480 Agents API test failed due to not expected "alive" status of the agent From stephenfin at redhat.com Tue Dec 14 14:29:12 2021 From: stephenfin at redhat.com (Stephen Finucane) Date: Tue, 14 Dec 2021 14:29:12 +0000 Subject: [all] testtools and support for Python 3.10 (error du to deprecation of distutils) In-Reply-To: References: <6af95731-d926-74f8-8c16-e8583235a25a@debian.org> <20211213164815.ftio25ectty7qb2t@yuggoth.org> <8ad35b9d-b0c7-5d8c-77f6-de780ed59077@debian.org> <20211213211410.taznfxueot7yxcuw@yuggoth.org> <61ee101e-e627-eac1-7c05-4765d8041b67@debian.org> <20211213223457.wz6r2p5zaylptxic@yuggoth.org> Message-ID: <5a4dba1080a338ee39f8a3f8334f5419f49808e4.camel@redhat.com> On Mon, 2021-12-13 at 23:50 +0100, Thomas Goirand wrote: > On 12/13/21 11:34 PM, Jeremy Stanley wrote: > > On 2021-12-13 23:29:05 +0100 (+0100), Thomas Goirand wrote: > > > On 12/13/21 10:14 PM, Jeremy Stanley wrote: > > > > To reiterate, PYTHONWARNINGS must be getting set to "error" > > > > somewhere. Maybe it's exported in that environment before python3 is > > > > invoked? > > > > > > > > > > This returns nothing: > > > grep -r PYTHONWARNINGS * > > > > > > So no, it's not what's happening, unless I don't understand... > > > > Are you able to reproduce it locally? Under sbuild or some other > > way? > > > > Yep. Doing this: > > git clone > https://salsa.debian.org/openstack-team/services/rally-openstack.git > cd rally-openstack > sbuild > ... You might not be using tox, but pytest will still loading configuration from that file if it exists [1][2][3] Stephen [1] https://docs.pytest.org/en/7.0.x/reference/customize.html#tox-ini [2] https://docs.pytest.org/en/7.0.x/how-to/capture-warnings.html [3] https://github.com/openstack/rally/blob/0f0f20dab/tox.ini#L151-L163 > > Cheers, > > Thomas Goirand (zigo) > From michal.arbet at ultimum.io Tue Dec 14 15:25:54 2021 From: michal.arbet at ultimum.io (Michal Arbet) Date: Tue, 14 Dec 2021 16:25:54 +0100 Subject: [swift][tempest][kolla] In-Reply-To: <17db63694c8.12444de92127096.1140393958066839556@ghanshyammann.com> References: <17db63694c8.12444de92127096.1140393958066839556@ghanshyammann.com> Message-ID: Hi, [filter:authtoken] -delay_auth_decision = True +delay_auth_decision = True Above does the trick and now working :) Thank you Michal Arbet Openstack Engineer Ultimum Technologies a.s. Na Po???? 1047/26, 11000 Praha 1 Czech Republic +420 604 228 897 michal.arbet at ultimum.io *https://ultimum.io * LinkedIn | Twitter | Facebook ?t 14. 12. 2021 v 0:52 odes?latel Ghanshyam Mann napsal: > ---- On Mon, 13 Dec 2021 11:33:27 -0600 Michal Arbet < > michal.arbet at ultimum.io> wrote ---- > > Hello to eveyrone, > > Please, could someone help me with swift capabilities not working when > I'm tempesting openstack test stack ? > > Tempest : > > rm -rf /tmp/tempest-lock/; refstack-client test -v -c > /opt/tempest/tempest.conf -- --regex > tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest > > > > (.venv) root at ca6353106d83:/opt/refstack-client# rm -rf > /tmp/tempest-lock/; refstack-client test -v -c /opt/tempest/tempest.conf -- > --regex > tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest > > 2021-12-13 17:44:33.001 4358 INFO tempest [-] Using tempest config file > /etc/tempest/tempest.conf > > 2021-12-13 17:44:33,839 refstack_client:518 INFO Starting Tempest > test... > > 2021-12-13 17:44:33.839 4358 INFO refstack_client [-] Starting Tempest > test... > > {0} setUpClass > (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) > [0.000000s] ... FAILED > > > > Captured traceback: > > ~~~~~~~~~~~~~~~~~~~ > > Traceback (most recent call last): > > > > File "/opt/refstack-client/.tempest/tempest/test.py", line 181, > in setUpClass > > raise value.with_traceback(trace) > > > > File "/opt/refstack-client/.tempest/tempest/test.py", line 174, > in setUpClass > > cls.resource_setup() > > > > File > "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", > line 36, in resource_setup > > super(AccountQuotasNegativeTest, cls).resource_setup() > > > > File > "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line > 95, in resource_setup > > body = cls.capabilities_client.list_capabilities() > > > > File > "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", > line 32, in list_capabilities > > self._error_checker(resp, body) > > > > File > "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line > 799, in _error_checker > > raise exceptions.Unauthorized(resp_body, resp=resp) > > > > tempest.lib.exceptions.Unauthorized: Unauthorized > > Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The request > you have made requires authentication.'} > > > > > > ============================== > > Failed 1 tests - output below: > > ============================== > > > > setUpClass > (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) > > > ---------------------------------------------------------------------------------------------- > > > > Captured traceback: > > ~~~~~~~~~~~~~~~~~~~ > > Traceback (most recent call last): > > > > File "/opt/refstack-client/.tempest/tempest/test.py", line 181, > in setUpClass > > raise value.with_traceback(trace) > > > > File "/opt/refstack-client/.tempest/tempest/test.py", line 174, > in setUpClass > > cls.resource_setup() > > > > File > "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", > line 36, in resource_setup > > super(AccountQuotasNegativeTest, cls).resource_setup() > > > > File > "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line > 95, in resource_setup > > body = cls.capabilities_client.list_capabilities() > > > > File > "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", > line 32, in list_capabilities > > self._error_checker(resp, body) > > > > File > "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line > 799, in _error_checker > > raise exceptions.Unauthorized(resp_body, resp=resp) > > > > tempest.lib.exceptions.Unauthorized: Unauthorized > > Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The request > you have made requires authentication.'} > > > > > > > > ====== > > Totals > > ====== > > Ran: 1 tests in 0.0000 sec. > > - Passed: 0 > > - Skipped: 0 > > - Expected Fail: 0 > > - Unexpected Success: 0 > > - Failed: 1 > > Sum of execute time for each test: 0.0000 sec. > > > > TEMPEST LOG : > > Response - Headers: {'content-type': 'application/json', > 'content-length': '114', 'www-authenticate': 'Keystone uri=" > http://192.168.205.254:5000"', 'x-trans-id': > 'tx815e181e33fb4854b2631-0061b7787a', 'x-openstack-request-id': > 'tx815e181e33fb4854b2631-0061b7787a', 'date': 'Mon, 13 Dec 2021 16:44:42 > GMT', 'connection': 'close', 'status': '401', 'content-location': ' > https://api.refstack.ultimum.cloud:8080/info'} > > Body: b'{"error": {"code": 401, "title": "Unauthorized", > "message": "The request you have made requires authentication."}}' > _log_request_full > /opt/refstack-client/.tempest/tempest/lib/common/rest_client.py:450 > > > > > > > > Test from command line and from curl > > . /etc/kolla/refstack.sh ; curl -H "X-Auth-Token: $(openstack token > issue -f value -c id)" https://api.refstack.ultimum.cloud:8080/info > > {"swift": {"version": "2.27.1.dev9", "strict_cors_mode": true, > "policies": [{"name": "Policy-0", "aliases": "Policy-0", "default": true}], > "allow_account_management": true, "account_autocreate": true, > "max_file_size": 5368709122, "max_meta_name_length": 128, > "max_meta_value_length": 256, "max_meta_count": 90, > "max_meta_overall_size": 4096, "max_header_size": 8192, > "max_object_name_length": 1024, "container_listing_limit": 10000, > "account_listing_limit": 10000, "max_account_name_length": 256, > "max_container_name_length": 256, "extra_header_count": 0}, > "container_sync": {"realms": {}}, "bulk_upload": > {"max_containers_per_extraction": 10000, "max_failed_extractions": 1000}, > "bulk_delete": {"max_deletes_per_request": 10000, "max_failed_deletes": > 1000}, "tempurl": {"methods": ["GET", "HEAD", "PUT", "POST", "DELETE"], > "incoming_remove_headers": ["x-timestamp"], "incoming_allow_headers": [], > "outgoing_remove_headers": ["x-object-meta-*"], "outgoing_allow_headers": > ["x-object-meta-public-*"], "allowed_digests": ["sha1", "sha256", > "sha512"]}, "ratelimit": {"account_ratelimit": 0.0, > "max_sleep_time_seconds": 60.0, "container_ratelimits": [], > "container_listing_ratelimits": []}, "container_quotas": {}, > "account_quotas": {}, "slo": {"max_manifest_segments": 1000, > "max_manifest_size": 8388608, "yield_frequency": 10, "min_segment_size": 1, > "allow_async_delete": false}} > > > > Python Swiftclient : > > ubuntu at deploy:/opt/kolla-ansible$ . /etc/kolla/refstack.sh ; swift > --os-auth-url http://192.168.205.254:5000/v3 --auth-version 3 > --os-project-name refstack --os-project-domain-name default --os-username > refstack --os-user-domain-name default --os-password SECRET capabilities > > Capabilities GET failed: http://192.168.205.254:8080/info 401 > Unauthorized [first 60 chars of response] b'{"error": {"code": 401, > "title": "Unauthorized", "message": ' > > Failed Transaction ID: txc1d8607e26eb4cd587459-0061b7791b > > > > > > > > It looks like swift client is broken, isn't it ? Or ? Maybe > kolla-ansible is creating bad roles and config ? (operator_roles, > reselleradmin_roel ..etc ? ) Tempest is from master > > Thank you very much,Kevko > > Are you running it with dynamic creds or pre-provisioned creds ? > > Error is from cls.capabilities_client which is initialized from > CONF.object_storage.operator_role[1] which is 'member' > role by default. what is your configuration for this? > > With default CONF.object_storage.operator_role as 'member' role, this test > pass in upstream CI/CD so client is > not broken but it is configuration issue: > https://zuul.opendev.org/t/openstack/build/b29147647370418fb9fbb0182832749d/log/job-output.txt#27292 > > > [1] > https://github.com/openstack/tempest/blob/34432dc970d09a55572a68fa007575285e35b550/tempest/api/object_storage/base.py#L77 > > -gmann > > > > > > > Michal Arbet > > Openstack Engineer > > > > Ultimum Technologies a.s. > > Na Po???? 1047/26, 11000 Praha 1 > > Czech Republic > > > > +420 604 228 897 > > michal.arbet at ultimum.io > > https://ultimum.io > > > > LinkedIn | Twitter | Facebook > -------------- next part -------------- An HTML attachment was scrubbed... URL: From michal.arbet at ultimum.io Tue Dec 14 16:09:31 2021 From: michal.arbet at ultimum.io (Michal Arbet) Date: Tue, 14 Dec 2021 17:09:31 +0100 Subject: [swift][tempest][kolla] In-Reply-To: References: <17db63694c8.12444de92127096.1140393958066839556@ghanshyammann.com> Message-ID: Hi, Sorry I broke a diff because of manual edit, below is the fix. [filter:authtoken] -delay_auth_decision = False +delay_auth_decision = True Thank to Pierre Riteau (priteau) that he catched my fault :) Kevko Michal Arbet Openstack Engineer Ultimum Technologies a.s. Na Po???? 1047/26, 11000 Praha 1 Czech Republic +420 604 228 897 michal.arbet at ultimum.io *https://ultimum.io * LinkedIn | Twitter | Facebook ?t 14. 12. 2021 v 16:25 odes?latel Michal Arbet napsal: > Hi, > > [filter:authtoken] > -delay_auth_decision = True > +delay_auth_decision = True > > Above does the trick and now working :) > > Thank you > > Michal Arbet > Openstack Engineer > > Ultimum Technologies a.s. > Na Po???? 1047/26, 11000 Praha 1 > Czech Republic > > +420 604 228 897 > michal.arbet at ultimum.io > *https://ultimum.io * > > LinkedIn | Twitter > | Facebook > > > > ?t 14. 12. 2021 v 0:52 odes?latel Ghanshyam Mann > napsal: > >> ---- On Mon, 13 Dec 2021 11:33:27 -0600 Michal Arbet < >> michal.arbet at ultimum.io> wrote ---- >> > Hello to eveyrone, >> > Please, could someone help me with swift capabilities not working when >> I'm tempesting openstack test stack ? >> > Tempest : >> > rm -rf /tmp/tempest-lock/; refstack-client test -v -c >> /opt/tempest/tempest.conf -- --regex >> tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest >> > >> > (.venv) root at ca6353106d83:/opt/refstack-client# rm -rf >> /tmp/tempest-lock/; refstack-client test -v -c /opt/tempest/tempest.conf -- >> --regex >> tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest >> > 2021-12-13 17:44:33.001 4358 INFO tempest [-] Using tempest config >> file /etc/tempest/tempest.conf >> > 2021-12-13 17:44:33,839 refstack_client:518 INFO Starting Tempest >> test... >> > 2021-12-13 17:44:33.839 4358 INFO refstack_client [-] Starting Tempest >> test... >> > {0} setUpClass >> (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) >> [0.000000s] ... FAILED >> > >> > Captured traceback: >> > ~~~~~~~~~~~~~~~~~~~ >> > Traceback (most recent call last): >> > >> > File "/opt/refstack-client/.tempest/tempest/test.py", line 181, >> in setUpClass >> > raise value.with_traceback(trace) >> > >> > File "/opt/refstack-client/.tempest/tempest/test.py", line 174, >> in setUpClass >> > cls.resource_setup() >> > >> > File >> "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", >> line 36, in resource_setup >> > super(AccountQuotasNegativeTest, cls).resource_setup() >> > >> > File >> "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line >> 95, in resource_setup >> > body = cls.capabilities_client.list_capabilities() >> > >> > File >> "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", >> line 32, in list_capabilities >> > self._error_checker(resp, body) >> > >> > File >> "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line >> 799, in _error_checker >> > raise exceptions.Unauthorized(resp_body, resp=resp) >> > >> > tempest.lib.exceptions.Unauthorized: Unauthorized >> > Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The >> request you have made requires authentication.'} >> > >> > >> > ============================== >> > Failed 1 tests - output below: >> > ============================== >> > >> > setUpClass >> (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) >> > >> ---------------------------------------------------------------------------------------------- >> > >> > Captured traceback: >> > ~~~~~~~~~~~~~~~~~~~ >> > Traceback (most recent call last): >> > >> > File "/opt/refstack-client/.tempest/tempest/test.py", line 181, >> in setUpClass >> > raise value.with_traceback(trace) >> > >> > File "/opt/refstack-client/.tempest/tempest/test.py", line 174, >> in setUpClass >> > cls.resource_setup() >> > >> > File >> "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", >> line 36, in resource_setup >> > super(AccountQuotasNegativeTest, cls).resource_setup() >> > >> > File >> "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line >> 95, in resource_setup >> > body = cls.capabilities_client.list_capabilities() >> > >> > File >> "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", >> line 32, in list_capabilities >> > self._error_checker(resp, body) >> > >> > File >> "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line >> 799, in _error_checker >> > raise exceptions.Unauthorized(resp_body, resp=resp) >> > >> > tempest.lib.exceptions.Unauthorized: Unauthorized >> > Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The >> request you have made requires authentication.'} >> > >> > >> > >> > ====== >> > Totals >> > ====== >> > Ran: 1 tests in 0.0000 sec. >> > - Passed: 0 >> > - Skipped: 0 >> > - Expected Fail: 0 >> > - Unexpected Success: 0 >> > - Failed: 1 >> > Sum of execute time for each test: 0.0000 sec. >> > >> > TEMPEST LOG : >> > Response - Headers: {'content-type': 'application/json', >> 'content-length': '114', 'www-authenticate': 'Keystone uri=" >> http://192.168.205.254:5000"', 'x-trans-id': >> 'tx815e181e33fb4854b2631-0061b7787a', 'x-openstack-request-id': >> 'tx815e181e33fb4854b2631-0061b7787a', 'date': 'Mon, 13 Dec 2021 16:44:42 >> GMT', 'connection': 'close', 'status': '401', 'content-location': ' >> https://api.refstack.ultimum.cloud:8080/info'} >> > Body: b'{"error": {"code": 401, "title": "Unauthorized", >> "message": "The request you have made requires authentication."}}' >> _log_request_full >> /opt/refstack-client/.tempest/tempest/lib/common/rest_client.py:450 >> > >> > >> > >> > Test from command line and from curl >> > . /etc/kolla/refstack.sh ; curl -H "X-Auth-Token: $(openstack token >> issue -f value -c id)" https://api.refstack.ultimum.cloud:8080/info >> > {"swift": {"version": "2.27.1.dev9", "strict_cors_mode": true, >> "policies": [{"name": "Policy-0", "aliases": "Policy-0", "default": true}], >> "allow_account_management": true, "account_autocreate": true, >> "max_file_size": 5368709122, "max_meta_name_length": 128, >> "max_meta_value_length": 256, "max_meta_count": 90, >> "max_meta_overall_size": 4096, "max_header_size": 8192, >> "max_object_name_length": 1024, "container_listing_limit": 10000, >> "account_listing_limit": 10000, "max_account_name_length": 256, >> "max_container_name_length": 256, "extra_header_count": 0}, >> "container_sync": {"realms": {}}, "bulk_upload": >> {"max_containers_per_extraction": 10000, "max_failed_extractions": 1000}, >> "bulk_delete": {"max_deletes_per_request": 10000, "max_failed_deletes": >> 1000}, "tempurl": {"methods": ["GET", "HEAD", "PUT", "POST", "DELETE"], >> "incoming_remove_headers": ["x-timestamp"], "incoming_allow_headers": [], >> "outgoing_remove_headers": ["x-object-meta-*"], "outgoing_allow_headers": >> ["x-object-meta-public-*"], "allowed_digests": ["sha1", "sha256", >> "sha512"]}, "ratelimit": {"account_ratelimit": 0.0, >> "max_sleep_time_seconds": 60.0, "container_ratelimits": [], >> "container_listing_ratelimits": []}, "container_quotas": {}, >> "account_quotas": {}, "slo": {"max_manifest_segments": 1000, >> "max_manifest_size": 8388608, "yield_frequency": 10, "min_segment_size": 1, >> "allow_async_delete": false}} >> > >> > Python Swiftclient : >> > ubuntu at deploy:/opt/kolla-ansible$ . /etc/kolla/refstack.sh ; swift >> --os-auth-url http://192.168.205.254:5000/v3 --auth-version 3 >> --os-project-name refstack --os-project-domain-name default --os-username >> refstack --os-user-domain-name default --os-password SECRET capabilities >> > Capabilities GET failed: http://192.168.205.254:8080/info 401 >> Unauthorized [first 60 chars of response] b'{"error": {"code": 401, >> "title": "Unauthorized", "message": ' >> > Failed Transaction ID: txc1d8607e26eb4cd587459-0061b7791b >> > >> > >> > >> > It looks like swift client is broken, isn't it ? Or ? Maybe >> kolla-ansible is creating bad roles and config ? (operator_roles, >> reselleradmin_roel ..etc ? ) Tempest is from master >> > Thank you very much,Kevko >> >> Are you running it with dynamic creds or pre-provisioned creds ? >> >> Error is from cls.capabilities_client which is initialized from >> CONF.object_storage.operator_role[1] which is 'member' >> role by default. what is your configuration for this? >> >> With default CONF.object_storage.operator_role as 'member' role, this >> test pass in upstream CI/CD so client is >> not broken but it is configuration issue: >> https://zuul.opendev.org/t/openstack/build/b29147647370418fb9fbb0182832749d/log/job-output.txt#27292 >> >> >> [1] >> https://github.com/openstack/tempest/blob/34432dc970d09a55572a68fa007575285e35b550/tempest/api/object_storage/base.py#L77 >> >> -gmann >> >> > >> > >> > Michal Arbet >> > Openstack Engineer >> > >> > Ultimum Technologies a.s. >> > Na Po???? 1047/26, 11000 Praha 1 >> > Czech Republic >> > >> > +420 604 228 897 >> > michal.arbet at ultimum.io >> > https://ultimum.io >> > >> > LinkedIn | Twitter | Facebook >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From michal.arbet at ultimum.io Tue Dec 14 16:14:26 2021 From: michal.arbet at ultimum.io (Michal Arbet) Date: Tue, 14 Dec 2021 17:14:26 +0100 Subject: [swift][tempest][kolla] In-Reply-To: References: <17db63694c8.12444de92127096.1140393958066839556@ghanshyammann.com> Message-ID: Hi, Additional info from openstack swift team : #openstack-swift IRC: delay_auth_decision is this dangerous in swift ? or why it is default false ? i found that if it is false ..discoverability not working IIRC it defaults to false because tempauth does not need it, so SAIO works without. But basically all it does is letting 2 auths coexist, or have an auth that has 2 middlewares like authtoken (obtains the tokens) and keystone (makes the decision). So it's "delayed" in a sense that it invokes the auth hook after all the middlewares had a chance to execute. But of course it occurs before the request proceeds. hmm, so if I am using keystoneauth, i should set it to True, shouldn't I ? Yes. Well, strictly speaking, if middleware is in the correct order, you can get keystone itself to work. But not things like tempurl. So just set it. There's no security concern with it. So, from my perspective of view, we should set this to True in kolla-ansible and add Release note to inform users default value has changed. Kevko Michal Arbet Openstack Engineer Ultimum Technologies a.s. Na Po???? 1047/26, 11000 Praha 1 Czech Republic +420 604 228 897 michal.arbet at ultimum.io *https://ultimum.io * LinkedIn | Twitter | Facebook ?t 14. 12. 2021 v 17:09 odes?latel Michal Arbet napsal: > Hi, > > Sorry I broke a diff because of manual edit, below is the fix. > > [filter:authtoken] > -delay_auth_decision = False > +delay_auth_decision = True > > Thank to Pierre Riteau (priteau) that he catched my fault :) > > Kevko > > Michal Arbet > Openstack Engineer > > Ultimum Technologies a.s. > Na Po???? 1047/26, 11000 Praha 1 > Czech Republic > > +420 604 228 897 > michal.arbet at ultimum.io > *https://ultimum.io * > > LinkedIn | Twitter > | Facebook > > > > ?t 14. 12. 2021 v 16:25 odes?latel Michal Arbet > napsal: > >> Hi, >> >> [filter:authtoken] >> -delay_auth_decision = True >> +delay_auth_decision = True >> >> Above does the trick and now working :) >> >> Thank you >> >> Michal Arbet >> Openstack Engineer >> >> Ultimum Technologies a.s. >> Na Po???? 1047/26, 11000 Praha 1 >> Czech Republic >> >> +420 604 228 897 >> michal.arbet at ultimum.io >> *https://ultimum.io * >> >> LinkedIn | >> Twitter | Facebook >> >> >> >> ?t 14. 12. 2021 v 0:52 odes?latel Ghanshyam Mann >> napsal: >> >>> ---- On Mon, 13 Dec 2021 11:33:27 -0600 Michal Arbet < >>> michal.arbet at ultimum.io> wrote ---- >>> > Hello to eveyrone, >>> > Please, could someone help me with swift capabilities not working >>> when I'm tempesting openstack test stack ? >>> > Tempest : >>> > rm -rf /tmp/tempest-lock/; refstack-client test -v -c >>> /opt/tempest/tempest.conf -- --regex >>> tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest >>> > >>> > (.venv) root at ca6353106d83:/opt/refstack-client# rm -rf >>> /tmp/tempest-lock/; refstack-client test -v -c /opt/tempest/tempest.conf -- >>> --regex >>> tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest >>> > 2021-12-13 17:44:33.001 4358 INFO tempest [-] Using tempest config >>> file /etc/tempest/tempest.conf >>> > 2021-12-13 17:44:33,839 refstack_client:518 INFO Starting Tempest >>> test... >>> > 2021-12-13 17:44:33.839 4358 INFO refstack_client [-] Starting >>> Tempest test... >>> > {0} setUpClass >>> (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) >>> [0.000000s] ... FAILED >>> > >>> > Captured traceback: >>> > ~~~~~~~~~~~~~~~~~~~ >>> > Traceback (most recent call last): >>> > >>> > File "/opt/refstack-client/.tempest/tempest/test.py", line 181, >>> in setUpClass >>> > raise value.with_traceback(trace) >>> > >>> > File "/opt/refstack-client/.tempest/tempest/test.py", line 174, >>> in setUpClass >>> > cls.resource_setup() >>> > >>> > File >>> "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", >>> line 36, in resource_setup >>> > super(AccountQuotasNegativeTest, cls).resource_setup() >>> > >>> > File >>> "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line >>> 95, in resource_setup >>> > body = cls.capabilities_client.list_capabilities() >>> > >>> > File >>> "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", >>> line 32, in list_capabilities >>> > self._error_checker(resp, body) >>> > >>> > File >>> "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line >>> 799, in _error_checker >>> > raise exceptions.Unauthorized(resp_body, resp=resp) >>> > >>> > tempest.lib.exceptions.Unauthorized: Unauthorized >>> > Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The >>> request you have made requires authentication.'} >>> > >>> > >>> > ============================== >>> > Failed 1 tests - output below: >>> > ============================== >>> > >>> > setUpClass >>> (tempest.api.object_storage.test_account_quotas_negative.AccountQuotasNegativeTest) >>> > >>> ---------------------------------------------------------------------------------------------- >>> > >>> > Captured traceback: >>> > ~~~~~~~~~~~~~~~~~~~ >>> > Traceback (most recent call last): >>> > >>> > File "/opt/refstack-client/.tempest/tempest/test.py", line 181, >>> in setUpClass >>> > raise value.with_traceback(trace) >>> > >>> > File "/opt/refstack-client/.tempest/tempest/test.py", line 174, >>> in setUpClass >>> > cls.resource_setup() >>> > >>> > File >>> "/opt/refstack-client/.tempest/tempest/api/object_storage/test_account_quotas_negative.py", >>> line 36, in resource_setup >>> > super(AccountQuotasNegativeTest, cls).resource_setup() >>> > >>> > File >>> "/opt/refstack-client/.tempest/tempest/api/object_storage/base.py", line >>> 95, in resource_setup >>> > body = cls.capabilities_client.list_capabilities() >>> > >>> > File >>> "/opt/refstack-client/.tempest/tempest/lib/services/object_storage/capabilities_client.py", >>> line 32, in list_capabilities >>> > self._error_checker(resp, body) >>> > >>> > File >>> "/opt/refstack-client/.tempest/tempest/lib/common/rest_client.py", line >>> 799, in _error_checker >>> > raise exceptions.Unauthorized(resp_body, resp=resp) >>> > >>> > tempest.lib.exceptions.Unauthorized: Unauthorized >>> > Details: {'code': 401, 'title': 'Unauthorized', 'message': 'The >>> request you have made requires authentication.'} >>> > >>> > >>> > >>> > ====== >>> > Totals >>> > ====== >>> > Ran: 1 tests in 0.0000 sec. >>> > - Passed: 0 >>> > - Skipped: 0 >>> > - Expected Fail: 0 >>> > - Unexpected Success: 0 >>> > - Failed: 1 >>> > Sum of execute time for each test: 0.0000 sec. >>> > >>> > TEMPEST LOG : >>> > Response - Headers: {'content-type': 'application/json', >>> 'content-length': '114', 'www-authenticate': 'Keystone uri=" >>> http://192.168.205.254:5000"', 'x-trans-id': >>> 'tx815e181e33fb4854b2631-0061b7787a', 'x-openstack-request-id': >>> 'tx815e181e33fb4854b2631-0061b7787a', 'date': 'Mon, 13 Dec 2021 16:44:42 >>> GMT', 'connection': 'close', 'status': '401', 'content-location': ' >>> https://api.refstack.ultimum.cloud:8080/info'} >>> > Body: b'{"error": {"code": 401, "title": "Unauthorized", >>> "message": "The request you have made requires authentication."}}' >>> _log_request_full >>> /opt/refstack-client/.tempest/tempest/lib/common/rest_client.py:450 >>> > >>> > >>> > >>> > Test from command line and from curl >>> > . /etc/kolla/refstack.sh ; curl -H "X-Auth-Token: $(openstack token >>> issue -f value -c id)" https://api.refstack.ultimum.cloud:8080/info >>> > {"swift": {"version": "2.27.1.dev9", "strict_cors_mode": true, >>> "policies": [{"name": "Policy-0", "aliases": "Policy-0", "default": true}], >>> "allow_account_management": true, "account_autocreate": true, >>> "max_file_size": 5368709122, "max_meta_name_length": 128, >>> "max_meta_value_length": 256, "max_meta_count": 90, >>> "max_meta_overall_size": 4096, "max_header_size": 8192, >>> "max_object_name_length": 1024, "container_listing_limit": 10000, >>> "account_listing_limit": 10000, "max_account_name_length": 256, >>> "max_container_name_length": 256, "extra_header_count": 0}, >>> "container_sync": {"realms": {}}, "bulk_upload": >>> {"max_containers_per_extraction": 10000, "max_failed_extractions": 1000}, >>> "bulk_delete": {"max_deletes_per_request": 10000, "max_failed_deletes": >>> 1000}, "tempurl": {"methods": ["GET", "HEAD", "PUT", "POST", "DELETE"], >>> "incoming_remove_headers": ["x-timestamp"], "incoming_allow_headers": [], >>> "outgoing_remove_headers": ["x-object-meta-*"], "outgoing_allow_headers": >>> ["x-object-meta-public-*"], "allowed_digests": ["sha1", "sha256", >>> "sha512"]}, "ratelimit": {"account_ratelimit": 0.0, >>> "max_sleep_time_seconds": 60.0, "container_ratelimits": [], >>> "container_listing_ratelimits": []}, "container_quotas": {}, >>> "account_quotas": {}, "slo": {"max_manifest_segments": 1000, >>> "max_manifest_size": 8388608, "yield_frequency": 10, "min_segment_size": 1, >>> "allow_async_delete": false}} >>> > >>> > Python Swiftclient : >>> > ubuntu at deploy:/opt/kolla-ansible$ . /etc/kolla/refstack.sh ; swift >>> --os-auth-url http://192.168.205.254:5000/v3 --auth-version 3 >>> --os-project-name refstack --os-project-domain-name default --os-username >>> refstack --os-user-domain-name default --os-password SECRET capabilities >>> > Capabilities GET failed: http://192.168.205.254:8080/info 401 >>> Unauthorized [first 60 chars of response] b'{"error": {"code": 401, >>> "title": "Unauthorized", "message": ' >>> > Failed Transaction ID: txc1d8607e26eb4cd587459-0061b7791b >>> > >>> > >>> > >>> > It looks like swift client is broken, isn't it ? Or ? Maybe >>> kolla-ansible is creating bad roles and config ? (operator_roles, >>> reselleradmin_roel ..etc ? ) Tempest is from master >>> > Thank you very much,Kevko >>> >>> Are you running it with dynamic creds or pre-provisioned creds ? >>> >>> Error is from cls.capabilities_client which is initialized from >>> CONF.object_storage.operator_role[1] which is 'member' >>> role by default. what is your configuration for this? >>> >>> With default CONF.object_storage.operator_role as 'member' role, this >>> test pass in upstream CI/CD so client is >>> not broken but it is configuration issue: >>> https://zuul.opendev.org/t/openstack/build/b29147647370418fb9fbb0182832749d/log/job-output.txt#27292 >>> >>> >>> [1] >>> https://github.com/openstack/tempest/blob/34432dc970d09a55572a68fa007575285e35b550/tempest/api/object_storage/base.py#L77 >>> >>> -gmann >>> >>> > >>> > >>> > Michal Arbet >>> > Openstack Engineer >>> > >>> > Ultimum Technologies a.s. >>> > Na Po???? 1047/26, 11000 Praha 1 >>> > Czech Republic >>> > >>> > +420 604 228 897 >>> > michal.arbet at ultimum.io >>> > https://ultimum.io >>> > >>> > LinkedIn | Twitter | Facebook >>> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: From yipikai7 at gmail.com Tue Dec 14 16:23:07 2021 From: yipikai7 at gmail.com (Cedric Lemarchand) Date: Tue, 14 Dec 2021 17:23:07 +0100 Subject: [nova] Spawning instance that will do PXE booting In-Reply-To: <4fc29f97-c82f-7156-7876-7a9aef15d7ed@debian.org> References: <9a3cf269-6b89-fb32-6dc5-4fd708171cb2@inovex.de> <4fc29f97-c82f-7156-7876-7a9aef15d7ed@debian.org> Message-ID: On Tue, Dec 14, 2021 at 12:10 PM Thomas Goirand wrote: > It really is supported by the cli. I've done this way: Right, thing is I was trying to update the port configuration rather than creating it before the instance, good to know thanks. From amy at demarco.com Tue Dec 14 16:39:34 2021 From: amy at demarco.com (Amy Marrich) Date: Tue, 14 Dec 2021 10:39:34 -0600 Subject: [Diversity] D&I WG Meeting Time and IRC vs Video Poll In-Reply-To: References: Message-ID: It seems January 3rd(first Monday in January) is a holiday for most people so we will skip the January meeting. We will also extend the Doodle poll to January 19th to give folks time to participate. I will also send a reminder out in the new year. Please vote for your preferred days and times AND also for IRC and/or Video. https://doodle.com/poll/mnzxkauzwf2isssm?utm_source=poll&utm_medium=link Thanks, Amy (spotz) [Diversity] D&I WG Meeting Time and IRC vs Video Poll Amy Marrich Dec 12, 2021, 5:53 PM (2 days ago) to foundation, openstack-discuss, starlingx-discuss, kata-dev, airship-discuss, zuul-discuss, openinfralabs Hi everyone, The Diversity and Inclusion WG is holding a poll to see if there would be interest in attending if it were on another day and time as well as if IRC vs Video would be preferred. We currently meet the first Monday of the month at 17:00 UTC. And hope to have a planning meeting as our first meeting of the year. Our first meeting of the new year is slated for January 3rd at 17:00 UTC and for IRC. We will take the poll results into consideration and may cancel the January meeting to allow folks to plan on attending in which case February's meeting would be the first of the year. We are keeping to morning hours to encourage participation from EMEA but not overly early for PST folks to attend. We will keep the poll up through the end of the year. Please feel free to reach out with any questions or concerns and remember the Diversity and Inclusion WG is for ALL projects under the Open Infra Foundation! Thanks, Amy (spotz) foundation at lists.openstack.org, openstack-discuss ( openstack-discuss at lists.openstack.org), starlingx-discuss at lists.starlingx.io4 more It seems January 3rd(first Monday in January) is a holiday for most people so we will skip the January meeting. We will also extend the Doodle poll to January 19th to give folks time to participate. I will also send a reminder out in the new year. On Sun, Dec 12, 2021 at 5:53 PM Amy Marrich wrote: > Hi everyone, > > The Diversity and Inclusion WG is holding a poll to see if there would be > interest in attending if it were on another day and time as well as if IRC > vs Video would be preferred. We currently meet the first Monday of the > month at 17:00 UTC. And hope to have a planning meeting as our first > meeting of the year. > > Our first meeting of the new year is slated for January 3rd at 17:00 UTC > and for IRC. We will take the poll results into consideration and may > cancel the January meeting to allow folks to plan on attending in which > case February's meeting would be the first of the year. > > Please vote for your preferred days and times AND also for IRC and/or > Video. > https://doodle.com/poll/mnzxkauzwf2isssm?utm_source=poll&utm_medium=link > > We are keeping to morning hours to encourage participation from EMEA but > not overly early for PST folks to attend. We will keep the poll up through > the end of the year. > > Please feel free to reach out with any questions or concerns and remember > the Diversity and Inclusion WG is for ALL projects under the Open Infra > Foundation! > > Thanks, > > Amy (spotz) > -------------- next part -------------- An HTML attachment was scrubbed... URL: From hiromu.asahina.az at hco.ntt.co.jp Tue Dec 14 16:39:06 2021 From: hiromu.asahina.az at hco.ntt.co.jp (Hiromu Asahina) Date: Wed, 15 Dec 2021 01:39:06 +0900 Subject: [keystone] OAuth2.0 implementation in Yoga Message-ID: <000501d7f109$1cae65b0$560b3110$@hco.ntt.co.jp> Hi, Please, could any of the keystone core members give me some advice on this spec? https://review.opendev.org/c/openstack/keystone-specs/+/813152 We'd like to make the following points clear by the end of this year to forward the implementation. So, please kindly check it and please let me know your opinion. - OAuth2.0 scope [1]: As there are differences between OAuth2.0 scope format and the Application credentials access rule format and we haven't found a good solution to map them, we'd like to omit the implementation of the OAuth2.0 scope in Yoga. Is there any concerns? - Access policy configuration: - Which one is appropriate? (i) End-users can use the OAuth2.0 API if they have permission for the OAuth2.0 API even if they don't have permission for the Application credentials API (ii) End-users can use the OAuth2.0 API only if they have permission for both the OAuth2.0 API and the Application credentials API. - API endpoint: - Which one is appropriate? (i) `/identity/v3/auth/OS-OAUTH2//clients` (ii) `/identity/v3/users/{user_id}/OS-AUTH2/clients` (iii) other [1] https://datatracker.ietf.org/doc/html/rfc6749#page-23 Thanks, Hiromu Asahina (h_asahina) From ozzzo at yahoo.com Tue Dec 14 21:10:26 2021 From: ozzzo at yahoo.com (Albert Braden) Date: Tue, 14 Dec 2021 21:10:26 +0000 (UTC) Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: References: <5dd6d28f-9955-7ca5-0ab8-0c5ce11ceb7e@redhat.com> <14084e87df22458caa7668eea8b496b6@verisign.com> <1147779219.1274196.1639086048233@mail.yahoo.com> <986294621.1553814.1639155002132@mail.yahoo.com> Message-ID: <169252651.2819859.1639516226823@mail.yahoo.com> Following [1] I successfully set "amqp_durable_queues = True" and restricted HA to the appropriate queues, but I'm having trouble with some of the other settings such as "expires" and "message-ttl". Does anyone have an example of a working kolla config that includes these changes? [1] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit On Monday, December 13, 2021, 07:51:32 AM EST, Herve Beraud wrote: So, your config snippet LGTM. Le?ven. 10 d?c. 2021 ??17:50, Albert Braden a ?crit?: Sorry, that was a transcription error. I thought "True" and my fingers typed "False." The correct lines are: [oslo_messaging_rabbit] amqp_durable_queues = True On Friday, December 10, 2021, 02:55:55 AM EST, Herve Beraud wrote: If you plan to let `amqp_durable_queues = False` (i.e if you plan to keep this config equal to false), then you don't need to add these config lines as this is already the default value [1]. [1] https://opendev.org/openstack/oslo.messaging/src/branch/master/oslo_messaging/_drivers/amqp.py#L34 Le?jeu. 9 d?c. 2021 ??22:40, Albert Braden a ?crit?: Replying from my home email because I've been asked to not email the list from my work email anymore, until I get permission from upper management. I'm not sure I follow. I was planning to add 2 lines to etc/kolla/config/global.conf: [oslo_messaging_rabbit] amqp_durable_queues = False Is that not sufficient? What is involved in configuring dedicated control exchanges for each service? What would that look like in the config? From: Herve Beraud Sent: Thursday, December 9, 2021 2:45 AM To: Bogdan Dobrelya Cc: openstack-discuss at lists.openstack.org Subject: [EXTERNAL] Re: [ops] [kolla] RabbitMQ High Availability ? Caution:?This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe.? ? ? Le?mer. 8 d?c. 2021 ??11:48, Bogdan Dobrelya a ?crit?: Please see inline >> I read this with great interest because we are seeing this issue. Questions: >> >> 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? >> 2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? >> 3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? > > Note that even having rabbit HA policies adjusted like that and its HA > replication factor [0] decreased (e.g. to a 2), there still might be > high churn caused by a large enough number of replicated durable RPC > topic queues. And that might cripple the cloud down with the incurred > I/O overhead because a durable queue requires all messages in it to be > persisted to a disk (for all the messaging cluster replicas) before they > are ack'ed by the broker. > > Given that said, Oslo messaging would likely require a more granular > control for topic exchanges and the durable queues flag - to tell it to > declare as durable only the most critical paths of a service. A single > config setting and a single control exchange per a service might be not > enough. Also note that therefore, amqp_durable_queue=True requires dedicated control exchanges configured for each service. Those that use 'openstack' as a default cannot turn the feature ON. Changing it to a service specific might also cause upgrade impact, as described in the topic [3]. ? The same is true for `amqp_auto_delete=True`. That requires dedicated control exchanges else it won't work if each service defines its own policy on a shared control exchange (e.g `openstack`) and if policies differ from each other. ? [3] https://review.opendev.org/q/topic:scope-config-opts > > There are also race conditions with durable queues enabled, like [1]. A > solution could be where each service declare its own dedicated control > exchange with its own configuration. > > Finally, openstack components should add perhaps a *.next CI job to test > it with durable queues, like [2] > > [0] https://www.rabbitmq.com/ha.html#replication-factor > > [1] > https://zuul.opendev.org/t/openstack/build/aa514dd788f34cc1be3800e6d7dba0e8/log/controller/logs/screen-n-cpu.txt > > [2] https://review.opendev.org/c/openstack/nova/+/820523 > >> >> Does anyone have a sample set of RMQ config files that they can share? >> >> It looks like my Outlook has ruined the link; reposting: >> [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando -- Best regards, Bogdan Dobrelya, Irc #bogdando -- Herv? Beraud Senior Software Engineer at Red Hat irc: hberaud https://github.com/4383/ https://twitter.com/4383hberaud -- Herv? BeraudSenior Software Engineer at Red Hatirc: hberaudhttps://github.com/4383/https://twitter.com/4383hberaud -- Herv? BeraudSenior Software Engineer at Red Hatirc: hberaudhttps://github.com/4383/https://twitter.com/4383hberaud -------------- next part -------------- An HTML attachment was scrubbed... URL: From zaitcev at redhat.com Wed Dec 15 00:56:19 2021 From: zaitcev at redhat.com (Pete Zaitcev) Date: Tue, 14 Dec 2021 18:56:19 -0600 Subject: [swift][tempest][kolla] In-Reply-To: References: <17db63694c8.12444de92127096.1140393958066839556@ghanshyammann.com> Message-ID: <20211214185619.6296d956@niphredil.zaitcev.lan> On Tue, 14 Dec 2021 17:14:26 +0100 Michal Arbet wrote: > hmm, so if I am using keystoneauth, i should set it to True, > shouldn't I ? > Yes. > So, from my perspective of view, we should set this to True in > kolla-ansible and add Release note to inform users default value has > changed. Please file a ticket in Launchpad or what is kolla-ansible's preferred tracker is. Otherwise it's guaranteed to be forgotten. -- Pete From arne.wiebalck at cern.ch Wed Dec 15 07:45:01 2021 From: arne.wiebalck at cern.ch (Arne Wiebalck) Date: Wed, 15 Dec 2021 08:45:01 +0100 Subject: [baremetal-sig][ironic] Tue Dec 14, 2021, 2pm UTC: Metal3 Overview In-Reply-To: <9b43905a-e407-6574-d720-7b0bf0ce73b8@cern.ch> References: <9b43905a-e407-6574-d720-7b0bf0ce73b8@cern.ch> Message-ID: <11e4a041-f962-2843-ae33-4073d803be97@cern.ch> Dear all, The recording of Dmitry's "Metal3 Overview" in the Bare Metal SIG meeting yesterday is already available from https://www.youtube.com/watch?v=rjSC6cJ9YY8 All previous presentations from the Bare Metal SIG, e.g. on Prometheus, Redfish, RBAC, Bifrost, Burn-in, Multi-tenancy, deployment show cases and more are available from the Bare Metal SIG channel: https://www.youtube.com/playlist?list=PLKqaoAnDyfgoBFAjUvZGjKXQjogWZBLL_ The SIG will take a break in January and resume meetings in February 2022. Have a good break, see you next year! Arne On 13.12.21 09:55, Arne Wiebalck wrote: > Dear all, > > The Bare Metal SIG will meet tomorrow > > ? Tue Dec 14, 2021, at 2pm UTC on zoom. > > The meeting will feature a "topic-of-the-day" presentation > by Dmitry Tantsur (dtantsur) with an > > ? "Overview of Metal3" > > (and I hear there might be a demo even!) > > As usual, all details on https://etherpad.opendev.org/p/bare-metal-sig > > Everyone is welcome: come along, learn from the expert and > bring your questions! > > Cheers, > ?Arne > From tkajinam at redhat.com Wed Dec 15 10:47:43 2021 From: tkajinam at redhat.com (Takashi Kajinami) Date: Wed, 15 Dec 2021 19:47:43 +0900 Subject: [all] Pending patches proposed by release bot Message-ID: (I'm using all tag instead of project specific tags because the list can be too long) Hi all, I noticed that many patches proposed by release bot are still pending on reviews. Because of this, I see several projects like; - unit tests are executed in old test runtimes, - release note for specific release is missing ... https://review.opendev.org/q/owner:infra-root%2540openstack.org+status:open I'd like to encourage each project to take a moment to move these forward, or abandon them if these are unnecessary. I'm trying to list up all projects affected but I might be missing some of them. So please check the above query to find anything maintained by your project. - heat - keystone - mistral - sahara - magnum - adjutant - monasca - murano - barbican - freezer - openstacksdk - zaqar - watcher - telemetry(ceilometer/aodh/panko) Thank you, Takashi -------------- next part -------------- An HTML attachment was scrubbed... URL: From tkajinam at redhat.com Wed Dec 15 10:48:40 2021 From: tkajinam at redhat.com (Takashi Kajinami) Date: Wed, 15 Dec 2021 19:48:40 +0900 Subject: [all] Pending patches proposed by release bot In-Reply-To: References: Message-ID: On Wed, Dec 15, 2021 at 7:47 PM Takashi Kajinami wrote: > (I'm using all tag instead of project specific tags because the list can > be too long) > Hi all, > > > I noticed that many patches proposed by release bot are still pending on > reviews. > Because of this, I see several projects like; > I meant to say; Because of this, I see several *problems* like; > - unit tests are executed in old test runtimes, > - release note for specific release is missing > ... > > https://review.opendev.org/q/owner:infra-root%2540openstack.org+status:open > > I'd like to encourage each project to take a moment to move these forward, > or > abandon them if these are unnecessary. > > I'm trying to list up all projects affected but I might be missing some of > them. > So please check the above query to find anything maintained by your > project. > > - heat > - keystone > - mistral > - sahara > - magnum > - adjutant > - monasca > - murano > - barbican > - freezer > - openstacksdk > - zaqar > - watcher > - telemetry(ceilometer/aodh/panko) > > > Thank you, > Takashi > -------------- next part -------------- An HTML attachment was scrubbed... URL: From anbanerj at redhat.com Wed Dec 15 11:39:16 2021 From: anbanerj at redhat.com (Ananya Banerjee) Date: Wed, 15 Dec 2021 17:09:16 +0530 Subject: [gate][tripleo] gate blocker Message-ID: Hi, We have a tripleo gate blocker. Patches 816991,16 and 821778 which fixes bugs needs to go first to unblock the rest. Can someone please get these two patches to the top of the queue? https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/821699/ https://review.opendev.org/c/openstack/tripleo-ci/+/821778/ Thanks, Ananya -- Ananya Banerjee, RHCSA, RHCE-OSP Software Engineer Red Hat EMEA anbanerj at redhat.com M: +491784949931 IM: frenzy_friday @RedHat Red Hat Red Hat -------------- next part -------------- An HTML attachment was scrubbed... URL: From anbanerj at redhat.com Wed Dec 15 12:09:15 2021 From: anbanerj at redhat.com (Ananya Banerjee) Date: Wed, 15 Dec 2021 17:39:15 +0530 Subject: [gate][tripleo] gate blocker In-Reply-To: References: Message-ID: On Wed, Dec 15, 2021 at 5:09 PM Ananya Banerjee wrote: > Hi, > > We have a tripleo gate blocker. Patches 816991,16 and 821778 which fixes > bugs needs to go first to unblock the rest. Can someone please get these > two patches to the top of the queue? > > https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/821699/ > https://review.opendev.org/c/openstack/tripleo-ci/+/821778/ > Sorry pls ignore the previous patches. The correct patches that have to go first to unblock the gate are: 821538, 821778, 821699 in order. 821538 - fixes ceph failures on sc001 004 and 010 821778 - fixes wallaby container builds 821699 - fixes pcs version issue in ussuri/victoria upgrades > > > Thanks, > Ananya > -- > > Ananya Banerjee, RHCSA, RHCE-OSP > > Software Engineer > > Red Hat EMEA > > anbanerj at redhat.com > M: +491784949931 IM: frenzy_friday > @RedHat Red Hat > Red Hat > > > -- Ananya Banerjee, RHCSA, RHCE-OSP Software Engineer Red Hat EMEA anbanerj at redhat.com M: +491784949931 IM: frenzy_friday @RedHat Red Hat Red Hat -------------- next part -------------- An HTML attachment was scrubbed... URL: From dtantsur at redhat.com Wed Dec 15 12:30:48 2021 From: dtantsur at redhat.com (Dmitry Tantsur) Date: Wed, 15 Dec 2021 13:30:48 +0100 Subject: [all] Pending patches proposed by release bot In-Reply-To: References: Message-ID: On Wed, Dec 15, 2021 at 11:53 AM Takashi Kajinami wrote: > (I'm using all tag instead of project specific tags because the list can > be too long) > Hi all, > > > I noticed that many patches proposed by release bot are still pending on > reviews. > Because of this, I see several projects like; > - unit tests are executed in old test runtimes, > - release note for specific release is missing > ... > > https://review.opendev.org/q/owner:infra-root%2540openstack.org+status:open > > I'd like to encourage each project to take a moment to move these forward, > or > abandon them if these are unnecessary. > > I'm trying to list up all projects affected but I might be missing some of > them. > So please check the above query to find anything maintained by your > project. > > - heat > - keystone > - mistral > - sahara > - magnum > - adjutant > - monasca > - murano > - barbican > - freezer > - openstacksdk > I've taken care of 2 outstanding patches, let me know if I missed anything. Dmitry > - zaqar > - watcher > - telemetry(ceilometer/aodh/panko) > > > Thank you, > Takashi > -- Red Hat GmbH, https://de.redhat.com/ , Registered seat: Grasbrunn, Commercial register: Amtsgericht Muenchen, HRB 153243, Managing Directors: Charles Cachera, Brian Klemm, Laurie Krebs, Michael O'Neill -------------- next part -------------- An HTML attachment was scrubbed... URL: From senrique at redhat.com Wed Dec 15 13:31:51 2021 From: senrique at redhat.com (Sofia Enriquez) Date: Wed, 15 Dec 2021 10:31:51 -0300 Subject: [cinder] Bug deputy report for week of 12-15-2021 Message-ID: This is a bug report from 12-08-2021 to 12-15-2021. Agenda: https://etherpad.opendev.org/p/cinder-bug-squad-meeting ----------------------------------------------------------------------------------------- Medium - https://bugs.launchpad.net/cinder/+bug/1953704 "Volume Type deletion error (image_volume_cache_enabled = True) - stable/xena". Unassigned. Incomplete - https://bugs.launchpad.net/cinder/+bug/1954613 "Blank volume's attach_status causes the volume list to fail". Unassigned. - https://bugs.launchpad.net/cinder/+bug/1954664 "Customer has issues in removing the volumes on the hosts , volumes from 3par". Unassigned. Cheers -- Sof?a Enriquez she/her Software Engineer Red Hat PnT IRC: @enriquetaso @RedHat Red Hat Red Hat -------------- next part -------------- An HTML attachment was scrubbed... URL: From fungi at yuggoth.org Wed Dec 15 14:27:18 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Wed, 15 Dec 2021 14:27:18 +0000 Subject: [gate][tripleo] gate blocker In-Reply-To: References: Message-ID: <20211215142717.xfdz6zjnorc7p6us@yuggoth.org> On 2021-12-15 17:39:15 +0530 (+0530), Ananya Banerjee wrote: > On Wed, Dec 15, 2021 at 5:09 PM Ananya Banerjee wrote: > > > Hi, > > > > We have a tripleo gate blocker. Patches 816991,16 and 821778 which fixes > > bugs needs to go first to unblock the rest. Can someone please get these > > two patches to the top of the queue? > > > > https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/821699/ > > https://review.opendev.org/c/openstack/tripleo-ci/+/821778/ > > > > Sorry pls ignore the previous patches. The correct patches that have to go > first to unblock the gate are: 821538, 821778, 821699 in order. > 821538 - fixes ceph failures on sc001 004 and 010 > 821778 - fixes wallaby container builds > 821699 - fixes pcs version issue in ussuri/victoria upgrades Due to the length of the tripleo shared gate queue, I've moved these three items (in order) to the front of the list in order to minimize further resource thrashing due to gate resets. -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From pkliczew at redhat.com Wed Dec 15 16:15:53 2021 From: pkliczew at redhat.com (Piotr Kliczewski) Date: Wed, 15 Dec 2021 17:15:53 +0100 Subject: [Openstack][FOSDEM][CFP] Virtualization & IaaS Devroom In-Reply-To: References: Message-ID: All, A friendly reminder that the submission deadline is in 5 days. Thanks, Piotr On Tue, Nov 30, 2021 at 2:42 PM Piotr Kliczewski wrote: > We are excited to announce that the call for proposals is now open for the > Virtualization & IaaS devroom at the upcoming FOSDEM 2022, to be hosted > virtually on February 5th 2022. > > This year will mark FOSDEM?s 22nd anniversary as one of the > longest-running free and open source software developer events, attracting > thousands of developers and users from all over the world. Due to Covid-19, > FOSDEM will be held virtually this year on February 5th & 6th, 2022. > > About the Devroom > > The Virtualization & IaaS devroom will feature session topics such as open > source hypervisors and virtual machine managers such as Xen Project, KVM, > bhyve, and VirtualBox, and Infrastructure-as-a-Service projects such as > KubeVirt, Apache CloudStack, Foreman, OpenStack, oVirt, QEMU and OpenNebula. > > This devroom will host presentations that focus on topics of shared > interest, such as KVM; libvirt; shared storage; virtualized networking; > cloud security; clustering and high availability; interfacing with multiple > hypervisors; hyperconverged deployments; and scaling across hundreds or > thousands of servers. > > Presentations in this devroom will be aimed at users or developers working > on these platforms who are looking to collaborate and improve shared > infrastructure or solve common problems. We seek topics that encourage > dialog between projects and continued work post-FOSDEM. > > Important Dates > > Submission deadline: 20th of December > > Acceptance notifications: 25th of December > > Final schedule announcement: 31st of December > > Recorded presentations upload deadline: 15th of January > > Devroom: 6th February 2022 > > Submit Your Proposal > > All submissions must be made via the Pentabarf event planning site[1]. If > you have not used Pentabarf before, you will need to create an account. If > you submitted proposals for FOSDEM in previous years, you can use your > existing account. > > After creating the account, select Create Event to start the submission > process. Make sure to select Virtualization and IaaS devroom from the Track > list. Please fill out all the required fields, and provide a meaningful > abstract and description of your proposed session. > > Submission Guidelines > > We expect more proposals than we can possibly accept, so it is vitally > important that you submit your proposal on or before the deadline. Late > submissions are unlikely to be considered. > > All presentation slots are 30 minutes, with 20 minutes planned for > presentations, and 10 minutes for Q&A. > > All presentations will need to be pre-recorded and put into our system at > least a couple of weeks before the event. > > The presentations should be uploaded by 15th of January and made > available under Creative > > Commons licenses. In the Submission notes field, please indicate that you > agree that your presentation will be licensed under the CC-By-SA-4.0 or > CC-By-4.0 license and that you agree to have your presentation recorded. > For example: > > "If my presentation is accepted for FOSDEM, I hereby agree to license all > recordings, slides, and other associated materials under the Creative > Commons Attribution Share-Alike 4.0 International License. Sincerely, > ." > > In the Submission notes field, please also confirm that if your talk is > accepted, you will be able to attend the virtual FOSDEM event for the Q&A. > We will not consider proposals from prospective speakers who are unsure > whether they will be able to attend the FOSDEM virtual event. > > If you are experiencing problems with Pentabarf, the proposal submission > interface, or have other questions, you can email our devroom mailing > list[2] and we will try to help you. > > > Code of Conduct > > Following the release of the updated code of conduct for FOSDEM, we'd like > to remind all speakers and attendees that all of the presentations and > discussions in our devroom are held under the guidelines set in the CoC and > we expect attendees, speakers, and volunteers to follow the CoC at all > times. > > If you submit a proposal and it is accepted, you will be required to > confirm that you accept the FOSDEM CoC. If you have any questions about the > CoC or wish to have one of the devroom organizers review your presentation > slides or any other content for CoC compliance, please email us and we will > do our best to assist you. > > Call for Volunteers > > We are also looking for volunteers to help run the devroom. We need > assistance with helping speakers to record the presentation as well as > helping with streaming and chat moderation for the devroom. Please contact > devroom mailing list [2] for more information. > > Questions? > > If you have any questions about this devroom, please send your questions > to our devroom mailing list. You can also subscribe to the list to receive > updates about important dates, session announcements, and to connect with > other attendees. > > See you all at FOSDEM! > > [1] https://penta.fosdem.org/submission/FOSDEM22 > > [2] iaas-virt-devroom at lists.fosdem.org > -------------- next part -------------- An HTML attachment was scrubbed... URL: From massimo.sgaravatto at gmail.com Wed Dec 15 17:17:09 2021 From: massimo.sgaravatto at gmail.com (Massimo Sgaravatto) Date: Wed, 15 Dec 2021 18:17:09 +0100 Subject: [ops] [nova] problems with some default nova config values after update to Xena Message-ID: Hi After trying to update a CentOS8-stream testbed to Xena, there is a mess with some default values of nova In particular I see that: - everything (for all services) is being logged to syslog (it looks like that log_dir is using "none") - state_path for nova is using /usr/lib/python3.6/site-packages instead of /var/lib/nova Any hints on how to debug this? Thanks, Massimo -------------- next part -------------- An HTML attachment was scrubbed... URL: From dmendiza at redhat.com Wed Dec 15 17:20:31 2021 From: dmendiza at redhat.com (Douglas Mendizabal) Date: Wed, 15 Dec 2021 11:20:31 -0600 Subject: [barbican][keystone] Weekly meetings to resume in 2022 Message-ID: <3054000a-0462-1082-5859-0e96257da6ee@redhat.com> Hi Barbican and Keystone contributors! I'll be taking some time off for the holidays starting next week and won't be available to chair the weekly meetings. I'll be back on January 11 to resume the weekly meetings for both Barbican and Keystone teams. Happy holidays, y'all! See you in 2022! - Douglas Mendiz?bal (redrobot) From katonalala at gmail.com Wed Dec 15 17:21:59 2021 From: katonalala at gmail.com (Lajos Katona) Date: Wed, 15 Dec 2021 18:21:59 +0100 Subject: [neutron] PTL on PTO Message-ID: Hi, I will be on PTO, back in Tuesday, 04. 01. 2022. Don't hesitate to write, I will be reachable. Regards Lajos Katona (lajoskatona) -------------- next part -------------- An HTML attachment was scrubbed... URL: From skaplons at redhat.com Wed Dec 15 18:30:51 2021 From: skaplons at redhat.com (Slawek Kaplonski) Date: Wed, 15 Dec 2021 19:30:51 +0100 Subject: [neutron] PTL on PTO In-Reply-To: References: Message-ID: <3143668.aeNJFYEL58@p1> Hi, On ?roda, 15 grudnia 2021 18:21:59 CET Lajos Katona wrote: > Hi, > I will be on PTO, back in Tuesday, 04. 01. 2022. > Don't hesitate to write, I will be reachable. > > Regards > Lajos Katona (lajoskatona) Merry Christmas and Happy New Year! Take your time and rest during that holiday season! -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From viroel at gmail.com Wed Dec 15 18:33:21 2021 From: viroel at gmail.com (Douglas) Date: Wed, 15 Dec 2021 15:33:21 -0300 Subject: [gate][tripleo] gate blocker In-Reply-To: <20211215142717.xfdz6zjnorc7p6us@yuggoth.org> References: <20211215142717.xfdz6zjnorc7p6us@yuggoth.org> Message-ID: Hi all, lots of changes have merged from tripleo queue in the past minutes, gate blockers are clear at this moment. Jeremy, thanks for the help with this. o/ On Wed, Dec 15, 2021 at 11:28 AM Jeremy Stanley wrote: > On 2021-12-15 17:39:15 +0530 (+0530), Ananya Banerjee wrote: > > On Wed, Dec 15, 2021 at 5:09 PM Ananya Banerjee > wrote: > > > > > Hi, > > > > > > We have a tripleo gate blocker. Patches 816991,16 and 821778 which > fixes > > > bugs needs to go first to unblock the rest. Can someone please get > these > > > two patches to the top of the queue? > > > > > > > https://review.opendev.org/c/openstack/tripleo-quickstart-extras/+/821699/ > > > https://review.opendev.org/c/openstack/tripleo-ci/+/821778/ > > > > > > > Sorry pls ignore the previous patches. The correct patches that have to > go > > first to unblock the gate are: 821538, 821778, 821699 in order. > > 821538 - fixes ceph failures on sc001 004 and 010 > > 821778 - fixes wallaby container builds > > 821699 - fixes pcs version issue in ussuri/victoria upgrades > > Due to the length of the tripleo shared gate queue, I've moved these > three items (in order) to the front of the list in order to minimize > further resource thrashing due to gate resets. > -- > Jeremy Stanley > -- Douglas Salles Viroel - dviroel -------------- next part -------------- An HTML attachment was scrubbed... URL: From emilien at redhat.com Wed Dec 15 21:05:22 2021 From: emilien at redhat.com (Emilien Macchi) Date: Wed, 15 Dec 2021 16:05:22 -0500 Subject: [stable][requirements][zuul] unpinned setuptools dependency on stable In-Reply-To: <0861d9e7-0dc3-683f-ad65-120b156d03a0@est.tech> References: <6J4UZQ.VOBD0LVDTPUX1@est.tech> <827e99c6-99b2-54c8-a627-5153e3b84e6b@est.tech> <0861d9e7-0dc3-683f-ad65-120b156d03a0@est.tech> Message-ID: On Thu, Oct 14, 2021 at 10:03 AM El?d Ill?s wrote: > Hi, > > First, sorry for the slow response. > > I think pinning setuptools in requirements for stable branches is also a > good idea (up till wallaby). I can accept that. > > Another thing is that the openstack projects that I've checked don't have > issues in their CI regarding the unpinned setuptools. Mostly I saw/see the > problem in unit test, static code check and similar tox targets. > > Anyway, if this issue is there for devstack for others then I think we can > cap setuptools, too, in requirements repository, if it is OK for everyone. > My only concern is to cap it from the newest relevant stable branch where > we need it. If I'm not mistaken most of the projects have fixed their > related issue in Xena, so I guess Wallaby should be the first branch to cap > setuptools. > I don't know if someone was working on it already but I proposed the patch on Ussuri: https://review.opendev.org/c/openstack/requirements/+/821878 If this gets accepted, I'll backport it to stable/train as well. Without that patch, it's impossible to use Devstack outside of the gate to deploy Ussuri or Train. Thanks, Thanks, > > El?d > > On 2021. 10. 04. 20:16, Neil Jerram wrote: > > I can now confirm that > https://review.opendev.org/c/openstack/requirements/+/810859 fixes my CI > use case. (By temporarily using a fork of the requirements repo that > includes that change.) > > (Fix detail if needed here: > https://github.com/projectcalico/networking-calico/pull/64/commits/cbed6282405957f7d60b6e0790c91fb852afe84c > ) > > Best wishes. > Neil > > > On Mon, Oct 4, 2021 at 6:28 PM Neil Jerram wrote: > >> Is anyone helping to progress this? I just checked that stable/ussuri >> devstack is still broken. >> >> Best wishes, >> Neil >> >> >> On Tue, Sep 28, 2021 at 9:20 AM Neil Jerram wrote: >> >>> But I don't think that solution works for devstack, does it? Is there a >>> way to pin setuptools in a stable/ussuri devstack run, except by changing >>> the stable branch of the requirements project? >>> >>> >>> On Mon, Sep 27, 2021 at 7:50 PM El?d Ill?s >>> wrote: >>> >>>> Hi again, >>>> >>>> as I see there is no objection yet about using gibi's solution [1] (as >>>> I >>>> already summarized the situation in my previous mail [2]) for a fix for >>>> similar cases, so with a general stable core hat on, I *suggest* >>>> everyone to use that solution to pin the setuptools in tox for every >>>> failing cases (so that to avoid similar future errors as well). >>>> >>>> [1] https://review.opendev.org/810461 >>>> [2] >>>> >>>> http://lists.openstack.org/pipermail/openstack-discuss/2021-September/025059.html >>>> >>>> El?d >>>> >>>> >>>> On 2021. 09. 27. 14:47, Balazs Gibizer wrote: >>>> > >>>> > >>>> > On Fri, Sep 24 2021 at 10:21:33 PM +0200, Thomas Goirand >>>> > wrote: >>>> >> Hi Gibi! >>>> >> >>>> >> Thanks for bringing this up. >>>> >> >>>> >> As a distro package maintainer, here's my view. >>>> >> >>>> >> On 9/22/21 2:11 PM, Balazs Gibizer wrote: >>>> >>> Option 1: Bump the major version of the decorator dependency on >>>> >>> stable. >>>> >> >>>> >> Decorator 4.0.11 is even in Debian Stretch (currently oldoldstable), >>>> for >>>> >> which I don't even maintain OpenStack anymore (that's OpenStack >>>> >> Newton...). So I don't see how switching to decorator 4.0.0 is a >>>> >> problem, and I don't understand how OpenStack could be using 3.4.0 >>>> which >>>> >> is in Jessie (ie: 6 years old Debian release). >>>> >> >>>> >> PyPi says Decorator 3.4.0 is from 2012: >>>> >> https://pypi.org/project/decorator/#history >>>> >> >>>> >> Do you have your release numbers correct? If so, then switching to >>>> >> Decorator 4.4.2 (available in Debian Bullseye (shipped with Victoria) >>>> >> and Ubuntu >=Focal) looks like reasonable to me... Sticking with >>>> 3.4.0 >>>> >> feels a bit crazy (and I wasn't aware of it). >>>> > >>>> > Thanks for the info. So from Debian perspective it is OK to bump the >>>> > decorator version on stable. As others noted in this thread it seems >>>> > to be more than just decorator that broke. :/ >>>> > >>>> >> >>>> >>> Option 2: Pin the setuptools version during tox installation >>>> >> >>>> >> Please don't do this for the master branch, we need OpenStack to stay >>>> >> current with setuptools (yeah, even if this means breaking >>>> changes...). >>>> > >>>> > I've no intention to pin it on master. Master needs to work with the >>>> > latest and greatest. Also on master it is easier to fix / replace the >>>> > dependencies that become broken with new setuptools. >>>> > >>>> >> >>>> >> For already released OpenStack: I don't mind much if this is done (I >>>> >> could backport fixes if something breaks). >>>> > >>>> > ack >>>> > >>>> >> >>>> >>> Option 3: turn off lower-constraints testing >>>> >> >>>> >> I already expressed myself about this: this is dangerous as distros >>>> rely >>>> >> on it for setting lower bounds as low as possible (which is always >>>> >> preferred from a distro point of view). >>>> >> >>>> >>> Option 4: utilize pyproject.toml[6] to specify build-time >>>> requirements >>>> >> >>>> >> I don't know about pyproject.toml. >>>> >> >>>> >> Just my 2 cents, hoping it's useful, >>>> > >>>> > Thanks! >>>> > >>>> > Cheers, >>>> > gibi >>>> > >>>> >> Cheers, >>>> >> >>>> >> Thomas Goirand (zigo) >>>> >> >>>> > >>>> > >>>> > >>>> >>>> >>>> -- Emilien Macchi -------------- next part -------------- An HTML attachment was scrubbed... URL: From syedammad83 at gmail.com Wed Dec 15 21:19:23 2021 From: syedammad83 at gmail.com (Ammad Syed) Date: Thu, 16 Dec 2021 02:19:23 +0500 Subject: [neutron][ovn] VLAN Ranges Message-ID: Hi, I am using neutron xena release with OVN. I have currently defined the vlan range "network_vlan_ranges = vlannet1:3500:3600" in ml2_conf.ini. I have set bridge mapping via ovs with the below command. ovs-vsctl set open . external-ids:ovn-bridge-mappings=vlannet1:br-vlan Now I have to add a new vlan range. Can anyone confirm the syntax of ml2_conf.ini and bridge mapping. network_vlan_ranges = vlannet1:3500:3600,vlannet2:300:400 ovs-vsctl set open . external-ids:ovn-bridge-mappings=vlannet1:br-vlan,vlannet2:br-vlan Need to confirm the correct way to add new vlan range if the server is currently live ? Can we use the same bridge with two different mapping i.e vlannet1 and vlannet2 ? Ammad -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Wed Dec 15 22:45:13 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Wed, 15 Dec 2021 16:45:13 -0600 Subject: [all][tc] Technical Committee next weekly meeting on Dec 16th at 1500 UTC In-Reply-To: <17db6385d36.ca951ea5127116.4798485549554257708@ghanshyammann.com> References: <17db6385d36.ca951ea5127116.4798485549554257708@ghanshyammann.com> Message-ID: <17dc04566a3.da7d9b8430738.4767033272397735638@ghanshyammann.com> Hello Everyone, Below is the agenda for Tomorrow's TC IRC meeting schedule at 1500 UTC. https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting == Agenda for tomorrow's TC meeting == * Roll call * Follow up on past action items * Gate health check ** Fixing Zuul config error in OpenStack *** https://etherpad.opendev.org/p/zuul-config-error-openstack * Skyline as an official project ** https://review.opendev.org/c/openstack/governance/+/814037 ** http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026206.html * SIG i18n status check ** Xena translation missing *** http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026244.html ** Translation bug *** https://review.opendev.org/c/openstack/contributor-guide/+/821371 * Adjutant need PTLs and maintainers ** http://lists.openstack.org/pipermail/openstack-discuss/2021-October/025555.html * Open Reviews ** https://review.opendev.org/q/projects:openstack/governance+is:open -gmann ---- On Mon, 13 Dec 2021 17:54:46 -0600 Ghanshyam Mann wrote ---- > Hello Everyone, > > Technical Committee's next weekly meeting is scheduled for Dec 16th at 1500 UTC. > > If you would like to add topics for discussion, please add them to the below wiki page by > Wednesday, Dec 15th, at 2100 UTC. > > https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting > > -gmann > > > From miguel at mlavalle.com Wed Dec 15 23:49:19 2021 From: miguel at mlavalle.com (Miguel Lavalle) Date: Wed, 15 Dec 2021 17:49:19 -0600 Subject: [neutron] PTL on PTO In-Reply-To: <3143668.aeNJFYEL58@p1> References: <3143668.aeNJFYEL58@p1> Message-ID: Hi Lajos, I hope you enjoy your time off. Are the Neutron and Drivers meetings canceled for the rest of 2021? Best regards On Wed, Dec 15, 2021 at 12:41 PM Slawek Kaplonski wrote: > Hi, > > On ?roda, 15 grudnia 2021 18:21:59 CET Lajos Katona wrote: > > Hi, > > I will be on PTO, back in Tuesday, 04. 01. 2022. > > Don't hesitate to write, I will be reachable. > > > > Regards > > Lajos Katona (lajoskatona) > > Merry Christmas and Happy New Year! > Take your time and rest during that holiday season! > > -- > Slawek Kaplonski > Principal Software Engineer > Red Hat -------------- next part -------------- An HTML attachment was scrubbed... URL: From zhangbailin at inspur.com Thu Dec 16 07:00:04 2021 From: zhangbailin at inspur.com (=?utf-8?B?QnJpbiBaaGFuZyjlvKDnmb7mnpcp?=) Date: Thu, 16 Dec 2021 07:00:04 +0000 Subject: =?utf-8?B?562U5aSNOiBbY3lib3JnXVtub3ZhXVtxYV1RdWVzdGlvbiBhYm91dCBBY2Nl?= =?utf-8?Q?lerator_unbinding?= In-Reply-To: <99456fa20ca2a93b434608cf313f0047ace47abf.camel@redhat.com> References: <3574a43d.ce9.17db17f9935.Coremail.xiaolihope1008@163.com> <99456fa20ca2a93b434608cf313f0047ace47abf.camel@redhat.com> Message-ID: <4bd0df96724f457a92d13eef2ad5523a@inspur.com> The current nova using device profile relies on flavor extra_specs, so uninstalling can only be achieved by replacing the flavor, as Sean said. In order to better improve the operability of the accelerator, we need to give a warning in the unbind arq API document of Cyborg, which will be improved later. thanks. brinzhang -----????----- ???: Sean Mooney [mailto:smooney at redhat.com] ????: 2021?12?14? 0:32 ???: Di XiaoLi ; openstack-discuss ??: Re: [cyborg][nova][qa]Question about Accelerator unbinding On Mon, 2021-12-13 at 09:54 +0800, Di XiaoLi wrote: > hi, Cyborg and nova team: > I am using cyborg with "Wallaby" release to manage my accelerator devices, while I'm trying to unbind the accelerator I found that the device was not actually unbound from the virtual machine. > Here are my questions: > 1. What is the function of the arq unbind command in cyborg ? it has two usecases. it call by nova when nova is deleting or moving the vm it can be used by an enduser if they are not using cybrog with nova. > 2. How to unbind the accelerator which bounded to vm? Does nova or > cyborg support this function now? > the only way to do that for a nova instance is to resize it to a flavor that does not request the device via the cyborg device profiel in the extra spec. more recently we have started to support using cybrog for neutron nics too. in this case the unbinding can be done by doing a port detach and the device should be remvoed from the vm and unbond in cybrog. > > Here are my steps: > step1: openstack accelerator arq unbind > ed205084-f58d-4fad-99b4-327a1398858f this should be treated as an admin/service only oepration wehn using cybrog with nova more on that below. > +---------------------+--------------------------------------+ > > Field | Value | > +---------------------+--------------------------------------+ > > uuid | ed205084-f58d-4fad-99b4-327a1398858f | > > state | Unbound | > > device_profile_name | ssd | > > hostname | None | > > device_rp_uuid | None | > > instance_uuid | None | > > attach_handle_type | | > > attach_handle_info | {} | > +---------------------+--------------------------------------+ > > step2: login vm and check the device, but it still here. > step3: stop vm and start vm, met the following error: > "nova.exception.AcceleratorRequestOpFailed: Failed to get accelerator requests: Cyborg returned no accelerator requests for instance ca77ef4e-421c-4c6c-9d76-7618a90ec921" that is what i would expect and is the correct behavior as you violated the contract between nova and cyborg by unbindng the arq. when useing cyborg with nova you shoudl never use the cyborg api driectly expcit to list the devcice profiles. you should treat it like placment in that regard. cybrog when used with nova today is an internal service that end user should at most have read only access too list the profiles. nova does not support cybrog device hot/cold attach or detach. the only way to add or remove a device form cyborg via nova is the device profile request in the flavor. so the only support opartion to change the attach device is resize to a differnt flavor. as i said above if you are attching smartnics using cyborg via neutron then you can also attach/devatch cybrog device by attaching/detaching the neutron port which contains the device profile request. note that if the deivce does not exist on the currnt host the attch will likely fail. detach should more or less always work unless there is an internal errror with one of the 3 service invovled. > > From skaplons at redhat.com Thu Dec 16 07:42:26 2021 From: skaplons at redhat.com (Slawek Kaplonski) Date: Thu, 16 Dec 2021 08:42:26 +0100 Subject: [neutron][ovn] VLAN Ranges In-Reply-To: References: Message-ID: <11886699.O9o76ZdvQC@p1> Hi, On ?roda, 15 grudnia 2021 22:19:23 CET Ammad Syed wrote: > Hi, > > I am using neutron xena release with OVN. I have currently defined the vlan > range "network_vlan_ranges = vlannet1:3500:3600" in ml2_conf.ini. > > I have set bridge mapping via ovs with the below command. > > ovs-vsctl set open . external-ids:ovn-bridge-mappings=vlannet1:br-vlan > > Now I have to add a new vlan range. Can anyone confirm the syntax of > ml2_conf.ini and bridge mapping. > > network_vlan_ranges = vlannet1:3500:3600,vlannet2:300:400 That syntax is correct. This config option is comma-separated list of physnets and ranges as You did above. > > ovs-vsctl set open . > external-ids:ovn-bridge-mappings=vlannet1:br-vlan,vlannet2:br-vlan > > Need to confirm the correct way to add new vlan range if the server is > currently live ? Can we use the same bridge with two different mapping i.e > vlannet1 and vlannet2 ? No. On bridge can be mapped to one physical network. We even had proposal to change it recently: https://bugs.launchpad.net/neutron/+bug/1952867 but finally we decided that we don't want to allow that. > > Ammad -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From syedammad83 at gmail.com Thu Dec 16 08:06:26 2021 From: syedammad83 at gmail.com (Ammad Syed) Date: Thu, 16 Dec 2021 13:06:26 +0500 Subject: [neutron][ovn] VLAN Ranges In-Reply-To: <11886699.O9o76ZdvQC@p1> References: <11886699.O9o76ZdvQC@p1> Message-ID: Thanks Slawek. Is it possible to define two vlan ranges for one physnet for example 200 to 300 and 3500 to 3600? Ammad On Thu, Dec 16, 2021 at 12:42 PM Slawek Kaplonski wrote: > Hi, > > On ?roda, 15 grudnia 2021 22:19:23 CET Ammad Syed wrote: > > Hi, > > > > I am using neutron xena release with OVN. I have currently defined the > vlan > > range "network_vlan_ranges = vlannet1:3500:3600" in ml2_conf.ini. > > > > I have set bridge mapping via ovs with the below command. > > > > ovs-vsctl set open . external-ids:ovn-bridge-mappings=vlannet1:br-vlan > > > > Now I have to add a new vlan range. Can anyone confirm the syntax of > > ml2_conf.ini and bridge mapping. > > > > network_vlan_ranges = vlannet1:3500:3600,vlannet2:300:400 > > That syntax is correct. This config option is comma-separated list of > physnets > and ranges as You did above. > > > > > ovs-vsctl set open . > > external-ids:ovn-bridge-mappings=vlannet1:br-vlan,vlannet2:br-vlan > > > > Need to confirm the correct way to add new vlan range if the server is > > currently live ? Can we use the same bridge with two different mapping > i.e > > vlannet1 and vlannet2 ? > > No. On bridge can be mapped to one physical network. We even had proposal > to > change it recently: https://bugs.launchpad.net/neutron/+bug/1952867 but > finally > we decided that we don't want to allow that. > > > > > Ammad > > > > -- > Slawek Kaplonski > Principal Software Engineer > Red Hat -- Regards, Syed Ammad Ali -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Thu Dec 16 12:56:47 2021 From: zigo at debian.org (Thomas Goirand) Date: Thu, 16 Dec 2021 13:56:47 +0100 Subject: [tc][all] Follow up pain point discussion In-Reply-To: <17da028f2c6.f53fb4e822720.4072888890953590866@ghanshyammann.com> References: <17da028f2c6.f53fb4e822720.4072888890953590866@ghanshyammann.com> Message-ID: <7285a21e-2d2a-1c4f-d3ba-d74f65880ea6@debian.org> On 12/9/21 6:06 PM, Ghanshyam Mann wrote: > ---- On Thu, 09 Dec 2021 04:00:05 -0600 Thomas Goirand wrote ---- > > > > Hi, > > > > I missed this thread and the discussion, though I've added to the etherpad: > > > > Glance > > > > FIX NEEDED - Glance doesn't work well over uwsgi, it's the only > > service with this state (appart from Swfit). We would need Glance to > > fully gate using uwsgi in the CI. > > We run glance in uwsgi mode by default since wallaby [1] (GLANCE_STANDALONE=False means uwsgi) > and from this patch onwards[2], even import workflow is run in uwsgi by default in CI. > > [1] https://github.com/openstack/devstack/blob/6c849e371384e468679d3d030fe494a36587c505/lib/glance#L85 > [2] https://review.opendev.org/c/openstack/devstack/+/742332 > > -gmann Last time I checked in Victoria, Glance was broken when we had: Transfer-Encoding: chucked Has this been fixed? If yes, can someone point at the patches? Cheers, Thomas Goirand (zigo) From rosmaita.fossdev at gmail.com Thu Dec 16 14:42:14 2021 From: rosmaita.fossdev at gmail.com (Brian Rosmaita) Date: Thu, 16 Dec 2021 09:42:14 -0500 Subject: [cinder] festival of XS reviews 17 December 2021 Message-ID: <0ed59283-93b3-f5df-9c58-fe1c5cef3ce8@gmail.com> Hello Argonauts [0], Apologies for the lateness of this reminder, but hopefully everyone has already used the handy ICS file [1] to have this event on their personal calendars. The most recent edition of the Cinder Festival of XS Reviews will be held tomorrow (Friday 17 December). who: Everyone! what: The Cinder Festival of XS Reviews when: Friday 17 December 2021 from 1400-1600 UTC where: https://meetpad.opendev.org/cinder-festival-of-reviews etherpad: https://etherpad.opendev.org/p/cinder-festival-of-reviews See you there! brian [0] At the Yoga PTG, the Cinder mascot was named "Argo" and the Cinder project team thereby became the "Argonauts". [1] http://eavesdrop.openstack.org/calendars/cinder-festival-of-reviews.ics From gmann at ghanshyammann.com Thu Dec 16 14:55:44 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Thu, 16 Dec 2021 08:55:44 -0600 Subject: [tc][all] Follow up pain point discussion In-Reply-To: <7285a21e-2d2a-1c4f-d3ba-d74f65880ea6@debian.org> References: <17da028f2c6.f53fb4e822720.4072888890953590866@ghanshyammann.com> <7285a21e-2d2a-1c4f-d3ba-d74f65880ea6@debian.org> Message-ID: <17dc3bdef3c.c0f9d0e042778.4126473039134888292@ghanshyammann.com> ---- On Thu, 16 Dec 2021 06:56:47 -0600 Thomas Goirand wrote ---- > On 12/9/21 6:06 PM, Ghanshyam Mann wrote: > > ---- On Thu, 09 Dec 2021 04:00:05 -0600 Thomas Goirand wrote ---- > > > > > > Hi, > > > > > > I missed this thread and the discussion, though I've added to the etherpad: > > > > > > Glance > > > > > > FIX NEEDED - Glance doesn't work well over uwsgi, it's the only > > > service with this state (appart from Swfit). We would need Glance to > > > fully gate using uwsgi in the CI. > > > > We run glance in uwsgi mode by default since wallaby [1] (GLANCE_STANDALONE=False means uwsgi) > > and from this patch onwards[2], even import workflow is run in uwsgi by default in CI. > > > > [1] https://github.com/openstack/devstack/blob/6c849e371384e468679d3d030fe494a36587c505/lib/glance#L85 > > [2] https://review.opendev.org/c/openstack/devstack/+/742332 > > > > -gmann > > Last time I checked in Victoria, Glance was broken when we had: > > Transfer-Encoding: chucked > > Has this been fixed? If yes, can someone point at the patches? Import workflow fixes I mentioned above were in wallaby, I am not sure about victoria. Maybe it is better to check on or after wallaby and see if there is still issues or glance team can give more insights (Dan is on PTO until Jan 2nd, he is the one who made it work on uwsgi). -gmann > > Cheers, > > Thomas Goirand (zigo) > > From pierre at stackhpc.com Thu Dec 16 15:11:50 2021 From: pierre at stackhpc.com (Pierre Riteau) Date: Thu, 16 Dec 2021 16:11:50 +0100 Subject: [tc][all] Follow up pain point discussion In-Reply-To: References: Message-ID: On Sun, 12 Dec 2021 at 22:08, Sean Mooney wrote: > > On Fri, 2021-12-10 at 22:02 +0100, Pierre Riteau wrote: > > On Mon, 6 Dec 2021 at 11:59, Stephen Finucane wrote: > > > novaclient ? full feature parity since 5.5.0 > > > > > > We finally introduced a proper '--block-device' option for 'server create' in 5.5.0, which closed our remaining RFE. There have been bugfixes since but it's feature complete, if not bug free. > > > > What about `nova host-evacuate-live` and a few other commands > > identified in doc/source/cli/data/nova.csv? > this has deliberatlly not being ported to OSC and we never intend to add it. > > we advise people to impelemtnt this them selves as it purly a client side impleantion and discurage > using the exsiting host-eveacuate-live as its error handeling if some live mirations fail is undefiend. > > you can do a better job yourself. > > > > > aggregate-cache-images,,Request images be cached. (Supported by API > > versions '2.81' - '2.latest') [hint: use '-- os-compute-api-version' > > flag to show help message for proper version] > > host-evacuate,,Evacuate all instances from failed host. > > host-evacuate-live,,Live migrate all instances off the specified host > > to other available hosts. > > all the host eveacutate apis were not ported for the same reasons as host-evacuate-live. > this is something the nova team nolonger wants to maintain and we deliberatly chose to not port those to > osc. > > > host-meta,,Set or Delete metadata on all instances of a host. > again this is not something we want to maintain. > its not part of the nova api and can be implented yourself in a few lines of bash. > > > host-servers-migrate,,Cold migrate all instances off the specified > as with the other host-evacuate command this is condiers depercated and was intentionally not ported > > host to other available hosts. > > hypervisor-servers,,List servers belonging to specific hypervisors. > openstack server list --all-projects --host > > by the way all the host-evacuate,host-evacuate-live and host-servers-migrate commands do is call > the list endpoint as show aboove to get vms on a give host and loop over them calling evacuate, live migrate or cold migrate > there is no error handeling and for hsot-server-migrate i dont think it actully confirms the migration. > it may but the dosc dont say one way our another. it certenly wont do the right thing in most caes if there is an error so you should not use it > if you really care about the vms. > > > > hypervisor-uptime,,Display the uptime of the specified hypervisor. > the uptime of a hyperivor is part of a deprectated api that was mainly used by xen > that is technially avaiable as part of the hyperviors api now but os-host has been deprecated for years > and it will eventraully be removed. > > > instance-action,,Show an action. > > instance-action-list,,List actions on a server. > > the instance action are aviable via > openstack server event {list,show} > > > > instance-usage-audit-log,,List/Get server usage audits > > > usage is aviabel via "openstack usage {list,show}" > > > > . > > migration-list,,Print a list of migrations. > migration info is aviable at > > "openstack server migration {abort,confirm,force complete,list,rever,show}" > and you can also confrim or revert a migrate via > "openstack server migrate {conrim,revert}" > > > version-list,,List all API versions. > > volume-update,,Update volume attachment. > volume updates shoudl be done via cinder all nova proxy apis have been deprecated for 3+ year proably even longer so no clinet > shoudl still be using those > > the equivalent woudl openstack volume migrate i belvie. > the volume-update api is really swap volume i belive > > > I didn't realise the host-evacuate-live command was not planned to be reimplemented. Thanks for the details. From ozzzo at yahoo.com Thu Dec 16 18:43:56 2021 From: ozzzo at yahoo.com (Albert Braden) Date: Thu, 16 Dec 2021 18:43:56 +0000 (UTC) Subject: [ops] [kolla] RabbitMQ High Availability In-Reply-To: <169252651.2819859.1639516226823@mail.yahoo.com> References: <5dd6d28f-9955-7ca5-0ab8-0c5ce11ceb7e@redhat.com> <14084e87df22458caa7668eea8b496b6@verisign.com> <1147779219.1274196.1639086048233@mail.yahoo.com> <986294621.1553814.1639155002132@mail.yahoo.com> <169252651.2819859.1639516226823@mail.yahoo.com> Message-ID: <1335760337.3548170.1639680236968@mail.yahoo.com> I tried these policies in ansible/roles/rabbitmq/templates/definitions.json.j2: "policies":[ {"vhost": "/", "name": "ha-all", "pattern": '^(?!(amq\.)|(.*_fanout_)|(reply_)).*', "apply-to": "all", "definition": {"ha-mode":"all"}, "priority":0}{% if project_name == 'outward_rabbitmq' %}, {"vhost": "/", "name": "notifications-ttl", "pattern": "^(notifications|versioned_notifications)\\.", "apply-to": "queues", "definition": {"message-ttl":600}, "priority":0} {"vhost": "/", "name": "notifications-expire", "pattern": "^(notifications|versioned_notifications)\\.", "apply-to": "queues", "definition": {"expire":3600}, "priority":0} {"vhost": "{{ murano_agent_rabbitmq_vhost }}", "name": "ha-all", "pattern": ".*", "apply-to": "all", "definition": {"ha-mode":"all"}, "priority":0} {% endif %} But I still see unconsumed messages lingering in notifications_extractor.info. From reading the docs I think this setting should cause messages to expire after 600 seconds, and unused queues to be deleted after 3600 seconds. What am I missing? On Tuesday, December 14, 2021, 04:18:09 PM EST, Albert Braden wrote: Following [1] I successfully set "amqp_durable_queues = True" and restricted HA to the appropriate queues, but I'm having trouble with some of the other settings such as "expires" and "message-ttl". Does anyone have an example of a working kolla config that includes these changes? [1] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit On Monday, December 13, 2021, 07:51:32 AM EST, Herve Beraud wrote: So, your config snippet LGTM. Le?ven. 10 d?c. 2021 ??17:50, Albert Braden a ?crit?: Sorry, that was a transcription error. I thought "True" and my fingers typed "False." The correct lines are: [oslo_messaging_rabbit] amqp_durable_queues = True On Friday, December 10, 2021, 02:55:55 AM EST, Herve Beraud wrote: If you plan to let `amqp_durable_queues = False` (i.e if you plan to keep this config equal to false), then you don't need to add these config lines as this is already the default value [1]. [1] https://opendev.org/openstack/oslo.messaging/src/branch/master/oslo_messaging/_drivers/amqp.py#L34 Le?jeu. 9 d?c. 2021 ??22:40, Albert Braden a ?crit?: Replying from my home email because I've been asked to not email the list from my work email anymore, until I get permission from upper management. I'm not sure I follow. I was planning to add 2 lines to etc/kolla/config/global.conf: [oslo_messaging_rabbit] amqp_durable_queues = False Is that not sufficient? What is involved in configuring dedicated control exchanges for each service? What would that look like in the config? From: Herve Beraud Sent: Thursday, December 9, 2021 2:45 AM To: Bogdan Dobrelya Cc: openstack-discuss at lists.openstack.org Subject: [EXTERNAL] Re: [ops] [kolla] RabbitMQ High Availability ? Caution:?This email originated from outside the organization. Do not click links or open attachments unless you recognize the sender and know the content is safe.? ? ? Le?mer. 8 d?c. 2021 ??11:48, Bogdan Dobrelya a ?crit?: Please see inline >> I read this with great interest because we are seeing this issue. Questions: >> >> 1. We are running kola-ansible Train, and our RMQ version is 3.7.23. Should we be upgrading our Train clusters to use 3.8.x? >> 2. Document [2] recommends policy '^(?!(amq\.)|(.*_fanout_)|(reply_)).*'. I don't see this in our ansible playbooks, nor in any of the config files in the RMQ container. What would this look like in Ansible, and what should the resulting container config look like? >> 3. It appears that we are not setting "amqp_durable_queues = True". What does this setting look like in Ansible, and what file does it go into? > > Note that even having rabbit HA policies adjusted like that and its HA > replication factor [0] decreased (e.g. to a 2), there still might be > high churn caused by a large enough number of replicated durable RPC > topic queues. And that might cripple the cloud down with the incurred > I/O overhead because a durable queue requires all messages in it to be > persisted to a disk (for all the messaging cluster replicas) before they > are ack'ed by the broker. > > Given that said, Oslo messaging would likely require a more granular > control for topic exchanges and the durable queues flag - to tell it to > declare as durable only the most critical paths of a service. A single > config setting and a single control exchange per a service might be not > enough. Also note that therefore, amqp_durable_queue=True requires dedicated control exchanges configured for each service. Those that use 'openstack' as a default cannot turn the feature ON. Changing it to a service specific might also cause upgrade impact, as described in the topic [3]. ? The same is true for `amqp_auto_delete=True`. That requires dedicated control exchanges else it won't work if each service defines its own policy on a shared control exchange (e.g `openstack`) and if policies differ from each other. ? [3] https://review.opendev.org/q/topic:scope-config-opts > > There are also race conditions with durable queues enabled, like [1]. A > solution could be where each service declare its own dedicated control > exchange with its own configuration. > > Finally, openstack components should add perhaps a *.next CI job to test > it with durable queues, like [2] > > [0] https://www.rabbitmq.com/ha.html#replication-factor > > [1] > https://zuul.opendev.org/t/openstack/build/aa514dd788f34cc1be3800e6d7dba0e8/log/controller/logs/screen-n-cpu.txt > > [2] https://review.opendev.org/c/openstack/nova/+/820523 > >> >> Does anyone have a sample set of RMQ config files that they can share? >> >> It looks like my Outlook has ruined the link; reposting: >> [2] https://wiki.openstack.org/wiki/Large_Scale_Configuration_Rabbit > > > -- > Best regards, > Bogdan Dobrelya, > Irc #bogdando -- Best regards, Bogdan Dobrelya, Irc #bogdando -- Herv? Beraud Senior Software Engineer at Red Hat irc: hberaud https://github.com/4383/ https://twitter.com/4383hberaud -- Herv? BeraudSenior Software Engineer at Red Hatirc: hberaudhttps://github.com/4383/https://twitter.com/4383hberaud -- Herv? BeraudSenior Software Engineer at Red Hatirc: hberaudhttps://github.com/4383/https://twitter.com/4383hberaud -------------- next part -------------- An HTML attachment was scrubbed... URL: From arnaud.morin at gmail.com Thu Dec 16 19:29:25 2021 From: arnaud.morin at gmail.com (Arnaud Morin) Date: Thu, 16 Dec 2021 19:29:25 +0000 Subject: [nova] instance device_metadata Message-ID: Hey all, How is filled the device_metadata in instance_extra db? It's empty for my instances. It's supposed to be accessible then through metadata api, and contains at least network and block device info, but it's empty here. Thanks for any hint! From gmann at ghanshyammann.com Thu Dec 16 19:37:49 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Thu, 16 Dec 2021 13:37:49 -0600 Subject: [all][tc][policy] Canceling policy popup team next meeting Message-ID: <17dc4c030be.11259ce4362468.6818934354049028609@ghanshyammann.com> Hello Everyone, Due to holiday weeks, we are cancelling the policy pop-up next meeting scheduled for 30th Dec. We will continue this biweekly meeting from 13th Jan. https://wiki.openstack.org/wiki/Consistent_and_Secure_Default_Policies_Popup_Team#Meeting -gmann From kamil.madac at slovenskoit.sk Thu Dec 16 20:05:18 2021 From: kamil.madac at slovenskoit.sk (=?iso-8859-2?Q?Kamil_Mad=E1=E8?=) Date: Thu, 16 Dec 2021 20:05:18 +0000 Subject: [neutron] missing dhcp ports after reboot and neutron-ovs-cleanup Message-ID: Hello We have deployed Victoria release with kolla-ansible (openvswitch neutron driver), and after the reboot of control nodes we are experiencing issues with dhcp on some networks. I found out that there were missing dhcp tap interfaces in br-int openvswitch. After restart of dhcp agetn and openvswitch agent, interface was added to br-int and dhcp works again in that network. I found bug from 2016 (https://bugs.launchpad.net/kolla/+bug/1617188), that there is a script neutron-ovs-cleanup (it is in each neutron docker container), which should be run after reboot of dhcp/L3 nodes otherwise issues with dhcp/metadata/l3 functionalities could arise. I do not see mentioned this in kolla documentation, bug is quite old and expired, so I would like to ask, is that still valid and should it be run manually after each reboot of dhcp/L3 node? If yes, what is best what how to achieve it when docker starts all container automatically after reboot. Should I stop neutron containers, run neutron-ovs-cleanup and start them again? Kamil Mad?? Slovensko IT a.s. -------------- next part -------------- An HTML attachment was scrubbed... URL: From gouthampravi at gmail.com Thu Dec 16 20:21:44 2021 From: gouthampravi at gmail.com (Goutham Pacha Ravi) Date: Thu, 16 Dec 2021 12:21:44 -0800 Subject: [manila] No IRC meeting on 30th Dec 2021 Message-ID: Hi Zorillas, As we discussed during our meeting today [1], we'll have our last IRC meeting for 2021 next week (23rd Dec 2021 at 1500 UTC) in OFTC's #openstack-meeting-alt. We'll skip the meeting on 30th Dec 2021 due to the year end holidays. As always, please feel free to add to the meeting agenda here: https://wiki.openstack.org/wiki/Manila/Meetings#Next_meeting Thanks, Goutham Pacha Ravi [1] https://meetings.opendev.org/meetings/manila/2021/manila.2021-12-16-15.01.log.html#l-124 From arnaud.morin at gmail.com Thu Dec 16 20:37:46 2021 From: arnaud.morin at gmail.com (Arnaud Morin) Date: Thu, 16 Dec 2021 20:37:46 +0000 Subject: [nova] instance device_metadata In-Reply-To: References: Message-ID: Ok, got it, we need to tag the interface (or block device) on boot. I found how to tag using nova boot, is it possible to tag using openstack server create? Cheers, On 16.12.21 - 19:29, Arnaud Morin wrote: > Hey all, > > How is filled the device_metadata in instance_extra db? > It's empty for my instances. > > It's supposed to be accessible then through metadata api, and contains > at least network and block device info, but it's empty here. > > Thanks for any hint! From katonalala at gmail.com Thu Dec 16 20:55:42 2021 From: katonalala at gmail.com (Lajos Katona) Date: Thu, 16 Dec 2021 21:55:42 +0100 Subject: [neutron] No more meetings in 2021 :-) Message-ID: Hi Neutrinos! Due to the holiday season let's cancel the next 2 Team and Drivers meeting. See you all in January, and have great Holidays! Lajos Katona (lajoskatona) -------------- next part -------------- An HTML attachment was scrubbed... URL: From skaplons at redhat.com Thu Dec 16 21:04:01 2021 From: skaplons at redhat.com (Slawek Kaplonski) Date: Thu, 16 Dec 2021 22:04:01 +0100 Subject: [neutron] missing dhcp ports after reboot and neutron-ovs-cleanup In-Reply-To: References: Message-ID: <10003347.nUPlyArG6x@p1> Hi, On czwartek, 16 grudnia 2021 21:05:18 CET Kamil Mad?? wrote: > Hello > > We have deployed Victoria release with kolla-ansible (openvswitch neutron > driver), and after the reboot of control nodes we are experiencing issues > with dhcp on some networks. I found out that there were missing dhcp tap > interfaces in br-int openvswitch. After restart of dhcp agetn and > openvswitch agent, interface was added to br-int and dhcp works again in > that network. I found bug from 2016 > (https://bugs.launchpad.net/kolla/+bug/1617188), that there is a script > neutron-ovs-cleanup (it is in each neutron docker container), which should > be run after reboot of dhcp/L3 nodes otherwise issues with dhcp/metadata/l3 > functionalities could arise. Basically what this script is doing is the same for both L3 and DHCP agents as both of them are plugging things into br-int in the same way. > > I do not see mentioned this in kolla documentation, bug is quite old and > expired, so I would like to ask, is that still valid and should it be run > manually after each reboot of dhcp/L3 node? If yes, what is best what how to > achieve it when docker starts all container automatically after reboot. > Should I stop neutron containers, run neutron-ovs-cleanup and start them > again? You must be sure that this script runs (if runs) before agents (and nova- compute) will start. The reason for that is that this script is pretty dummy and will simply remove everything from the br-int and other e.g. physical bridges. So if You will have some race between e.g. this script and DHCP agents, it may happen that dhcp agent will already create tap ports for some networks and then this script will remove them. > > > Kamil Mad?? > Slovensko IT a.s. -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From james.page at canonical.com Fri Dec 17 09:10:14 2021 From: james.page at canonical.com (James Page) Date: Fri, 17 Dec 2021 09:10:14 +0000 Subject: [charms] review abandonment policy Message-ID: Hi Charms Contributors As part of this cycles work to put in more proactive general maintenance of the Gerrit review queue for the OpenStack Charms projects we have introduced an abandonment policy for reviews: https://docs.openstack.org/charm-guide/latest/community/rotas.html#abandonment-policy Pruning the set of open reviews on a regular basis against this policy will help reviewers focus on active reviews. We'll start doing this in January - you can follow the blocked/failing links in the charm-guide policy section to see which reviews will be abandoned at any given point in time. Cheers James -------------- next part -------------- An HTML attachment was scrubbed... URL: From manchandavishal143 at gmail.com Fri Dec 17 10:09:18 2021 From: manchandavishal143 at gmail.com (vishal manchanda) Date: Fri, 17 Dec 2021 15:39:18 +0530 Subject: [horizon] Cancelling next two weekly meetings Message-ID: Hello Team, As agreed, during the last weekly meeting, we are canceling our weekly meeting on 22nd and 29th December. The next weekly meeting will be on 5th December. Happy holidays and happy new year in advance! Thanks & Regards, Vishal Manchanda -------------- next part -------------- An HTML attachment was scrubbed... URL: From pierre at stackhpc.com Fri Dec 17 10:20:51 2021 From: pierre at stackhpc.com (Pierre Riteau) Date: Fri, 17 Dec 2021 11:20:51 +0100 Subject: [blazar] Next IRC meeting cancelled Message-ID: Hello, As decided during yesterday's meeting, we are cancelling the Blazar IRC meeting scheduled for December 30. The next meeting will be on January 13. Merry Christmas and Happy New Year! Best wishes, Pierre Riteau (priteau) From skaplons at redhat.com Fri Dec 17 11:33:31 2021 From: skaplons at redhat.com (Slawek Kaplonski) Date: Fri, 17 Dec 2021 12:33:31 +0100 Subject: [neutron][ovn] VLAN Ranges In-Reply-To: References: <11886699.O9o76ZdvQC@p1> Message-ID: <4477813.LvFx2qVVIh@p1> Hi, On czwartek, 16 grudnia 2021 09:06:26 CET Ammad Syed wrote: > Thanks Slawek. Is it possible to define two vlan ranges for one physnet > for example 200 to 300 and 3500 to 3600? Looking at https://github.com/openstack/neutron-lib/blob/master/neutron_lib/ plugins/utils.py#L204 which is used to parse that config option, it is possible. I don't know what version of Neutron You are using, but since few releases we have possbility to configure segment ranges by API. Please check https:// docs.openstack.org/api-ref/network/v2/index.html?expanded=create-network- segment-range-detail#network-segment-ranges for details. > > Ammad > > On Thu, Dec 16, 2021 at 12:42 PM Slawek Kaplonski > > wrote: > > Hi, > > > > On ?roda, 15 grudnia 2021 22:19:23 CET Ammad Syed wrote: > > > Hi, > > > > > > I am using neutron xena release with OVN. I have currently defined the > > > > vlan > > > > > range "network_vlan_ranges = vlannet1:3500:3600" in ml2_conf.ini. > > > > > > I have set bridge mapping via ovs with the below command. > > > > > > ovs-vsctl set open . external-ids:ovn-bridge-mappings=vlannet1:br-vlan > > > > > > Now I have to add a new vlan range. Can anyone confirm the syntax of > > > ml2_conf.ini and bridge mapping. > > > > > > network_vlan_ranges = vlannet1:3500:3600,vlannet2:300:400 > > > > That syntax is correct. This config option is comma-separated list of > > physnets > > and ranges as You did above. > > > > > ovs-vsctl set open . > > > external-ids:ovn-bridge-mappings=vlannet1:br-vlan,vlannet2:br-vlan > > > > > > Need to confirm the correct way to add new vlan range if the server is > > > currently live ? Can we use the same bridge with two different mapping > > > > i.e > > > > > vlannet1 and vlannet2 ? > > > > No. On bridge can be mapped to one physical network. We even had proposal > > to > > change it recently: https://bugs.launchpad.net/neutron/+bug/1952867 but > > finally > > we decided that we don't want to allow that. > > > > > Ammad > > > > -- > > Slawek Kaplonski > > Principal Software Engineer > > Red Hat -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From ruslanas at lpic.lt Fri Dec 17 13:31:46 2021 From: ruslanas at lpic.lt (=?utf-8?Q?Ruslanas_G=C5=BEibovskis_LPIC?=) Date: Fri, 17 Dec 2021 15:31:46 +0200 Subject: [tripleo][ussuri] Log4j add protection to haproxy {even it is not impacted Message-ID: An HTML attachment was scrubbed... URL: From amonster369 at gmail.com Fri Dec 17 14:04:42 2021 From: amonster369 at gmail.com (A Monster) Date: Fri, 17 Dec 2021 15:04:42 +0100 Subject: No network access for openstack instances Message-ID: I have used kolla ansible to deploy openstack on two machine servers, but after setting up the images and the virtual networks, the vm instances don't get the ip address that neutron provides for them, so I had to manually set the addresses inside the vm's, but even with that I can't seem to connect vm's to neither internet nor any other network. -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Fri Dec 17 15:46:48 2021 From: zigo at debian.org (Thomas Goirand) Date: Fri, 17 Dec 2021 16:46:48 +0100 Subject: [neutron] Routing fake "floating IPs" of VMs running inside VMs of an OpenStack Message-ID: Hi, I've been designing a CI for OCI [1] to make it more robust and easier to contribute. After a week and a half, I've been able to make OCI deploy itself in an OpenStack itself. OCI itself pops with a public IP I can reach from the outside, and it setups OpenStack inside OpenStack VMs on a flat, private subnet (currently 192.168.100.0/24). This works pretty well, I have a fully working setup, and the deployment is completely automated. This even works without admin credentials, so all one needs, is a non-privileged account in any OpenStack. However, at this point, I'd like to setup the network for VMs inside the VMs. And that's where it becomes tricky. The VM that runs tempest will need to reach the floating IPs of the VMs inside the VMs that are running nova-compute, in order to do ssh tests. How can I do that? Is this even possible? Cheers, Thomas Goirand (zigo) [1] https://salsa.debian.org/openstack-team/debian/openstack-cluster-installer From fungi at yuggoth.org Fri Dec 17 16:11:47 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Fri, 17 Dec 2021 16:11:47 +0000 Subject: [neutron] Routing fake "floating IPs" of VMs running inside VMs of an OpenStack In-Reply-To: References: Message-ID: <20211217161147.3gqof3r7y23dmly4@yuggoth.org> On 2021-12-17 16:46:48 +0100 (+0100), Thomas Goirand wrote: [...] > I'd like to setup the network for VMs inside the > VMs. And that's where it becomes tricky. The VM that runs tempest will > need to reach the floating IPs of the VMs inside the VMs that are > running nova-compute, in order to do ssh tests. How can I do that? Is > this even possible? [...] In the upstream CI system we do that with bridge interfaces on each machine connected to layer 2 tunnels (using VXLAN, though GRE should also work if the cloud you're running in supports it): https://zuul-ci.org/docs/zuul-jobs/general-roles.html#role-multi-node-bridge While the implementation is in Ansible, it's really just a pile of shell commands you should be able to use to reproduce with just about anything. You can find the multi-node-* roles in the roles directory of the zuul-jobs repo: https://opendev.org/zuul/zuul-jobs Hope that helps! -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From cboylan at sapwetik.org Fri Dec 17 16:42:09 2021 From: cboylan at sapwetik.org (Clark Boylan) Date: Fri, 17 Dec 2021 08:42:09 -0800 Subject: [tripleo][ussuri] Log4j add protection to haproxy {even it is not impacted In-Reply-To: References: Message-ID: On Fri, Dec 17, 2021, at 5:31 AM, Ruslanas G?ibovskis LPIC wrote: > Hi team, > > Thanks to this Log4j I finally found time to read around, how to add > additional settings/options to haproxy config, especially, I would like > to apply haproxy steps to hide log4j vulns. > > I know I know, OSP looks not to be impacted, unless we have some > components such as opendaylight which might have Log4j applied. > > Does anyone has example for yaml file? > > Thanks in advance. It is my understanding that the log messages magic syntax in log4j is sophisticated enough that filtering via proxies is problematic and unlikely to catch everything. You'll be filtering simple attacks and letting sophisticated actors through. You are much better off upgrading log4j or disabling the JNDI class entirely in the jar. I wouldn't rely on haproxy for this. From ralonsoh at redhat.com Fri Dec 17 17:09:52 2021 From: ralonsoh at redhat.com (Rodolfo Alonso Hernandez) Date: Fri, 17 Dec 2021 18:09:52 +0100 Subject: No network access for openstack instances In-Reply-To: References: Message-ID: Hi: Can you provide more info/logs about this problem? What version are you using? What network backend are you using? You can also open a bug in https://bugs.launchpad.net. If you are using OVN, can you be hitting https://bugs.launchpad.net/kolla-ansible/+bug/1952550? However, you should at least receive the IP assigned from the DHCP server. Regards. On Fri, Dec 17, 2021 at 3:14 PM A Monster wrote: > I have used kolla ansible to deploy openstack on two machine servers, but > after setting up the images and the virtual networks, the vm instances > don't get the ip address that neutron provides for them, so I had to > manually set the addresses inside the vm's, but even with that I can't seem > to connect vm's to neither internet nor any other network. > -------------- next part -------------- An HTML attachment was scrubbed... URL: From amonster369 at gmail.com Fri Dec 17 18:53:45 2021 From: amonster369 at gmail.com (A Monster) Date: Fri, 17 Dec 2021 19:53:45 +0100 Subject: No network access for openstack instances In-Reply-To: References: Message-ID: I'm on centos 7 and trying to deploy openstack train using kolla ansible 9.3.1 Ansible 2.9.25 And neutron open vSwitch. On Fri, Dec 17, 2021, 18:10 Rodolfo Alonso Hernandez wrote: > Hi: > > Can you provide more info/logs about this problem? What version are you > using? What network backend are you using? You can also open a bug in > https://bugs.launchpad.net. > > If you are using OVN, can you be hitting > https://bugs.launchpad.net/kolla-ansible/+bug/1952550?? However, you > should at least receive the IP assigned from the DHCP server. > > Regards. > > On Fri, Dec 17, 2021 at 3:14 PM A Monster wrote: > >> I have used kolla ansible to deploy openstack on two machine servers, but >> after setting up the images and the virtual networks, the vm instances >> don't get the ip address that neutron provides for them, so I had to >> manually set the addresses inside the vm's, but even with that I can't seem >> to connect vm's to neither internet nor any other network. >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Fri Dec 17 19:00:31 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Fri, 17 Dec 2021 13:00:31 -0600 Subject: [all][tc] What's happening in Technical Committee: summary 17th Dec, 21: Reading: 10 min Message-ID: <17dc9c466cd.d43bdc19137374.6473452875851451460@ghanshyammann.com> Hello Everyone, Here is this week's summary of the Technical Committee activities. 1. TC Meetings: ============ * Most of the meeting discussions are summarized below (Completed or in-progress activities section). Meeting full logs are available @ https://meetings.opendev.org/meetings/tc/2021/tc.2021-12-16-15.00.log.html * Next week's meeting is on IRC on 23rd Dec (it is scheduled as of now but i will confirm it on Monday), Thursday 15:00 UTC, feel free the topic on the agenda[1] by 22nd Dec. 2. What we completed this week: ========================= * Skyline is accepted as an official project [2] * Updated Yoga testing runtime and job templates[3] 3. Activities In progress: ================== TC Tracker for Yoga cycle ------------------------------ * This etherpad includes the Yoga cycle targets/working items for TC[4]. Open Reviews ----------------- * 3 open reviews for ongoing activities[5]. SIG i18n status check -------------------------- As you might have seen in the email from Amotoki on missing Xena translation[6] or ID translation bug[7], we checked this SIG status in the TC meeting. We did not get any concrete answer about the status of this SIG, but I have an action item to reach out to the current chair of this SIG (Ian). If anyone would like to help or know who can help especially those who need the translation, please reply to this thread or thread started by Amotoki[6]. OpenStack Pain points discussion ---------------------------------------- No updates in this week. Stay tuned on the ML thread[8] and Rico will inform you about the next meeting details. TC position on release cadence ------------------------------------- Discussion is in-progress in ML thread[9]. No other updates from TC on this and we will set up a separate call to continue the discussion sometime in Jan (after the Christmas holidays). Fixing Zuul config error -------------------------- There is some progress by projects but still, we need to fix a lot of config errors. If your project is listed in zuul config error, please start planning to fix those[10] Community-wide goal updates ------------------------------------ You might have seen my email[11] about goal status and planning the work, below is the current status: * Selected (active) goals[12]: ** Migrate from oslo.rootwrap to oslo.privsep ** Consistent and Secure Default RBAC * Goal proposal under review: ** FIPS compatibility and compliance [13]. Adjutant need maintainers and PTLs ------------------------------------------- We are still waiting to hear from Braden, Albert on permission to work on this project[14]. Project updates ------------------- * Retire js-openstack-lib (waiting on Adjutant to have new PTL/maintainer) [15] 4. How to contact the TC: ==================== If you would like to discuss or give feedback to TC, you can reach out to us in multiple ways: 1. Email: you can send the email with tag [tc] on openstack-discuss ML[16]. 2. Weekly meeting: The Technical Committee conduct a weekly meeting every Thursday 15 UTC [17] 3. Ping us using 'tc-members' nickname on #openstack-tc IRC channel. [1] https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting [2] https://review.opendev.org/c/openstack/governance/+/814037 [3] https://review.opendev.org/c/openstack/governance/+/820195 [4] https://etherpad.opendev.org/p/tc-yoga-tracker [5] https://review.opendev.org/q/projects:openstack/governance+status:open [6] http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026244.html [7] https://review.opendev.org/c/openstack/contributor-guide/+/821371 [8] http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026245.html [9] http://lists.openstack.org/pipermail/openstack-discuss/2021-November/025684.html [10] https://etherpad.opendev.org/p/zuul-config-error-openstack [11] http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026218.html [12] https://governance.openstack.org/tc/goals/selected/index.html [13] https://review.opendev.org/c/openstack/governance/+/816587 [14] http://lists.openstack.org/pipermail/openstack-discuss/2021-November/025786.html [15] https://review.opendev.org/c/openstack/governance/+/798540 [16] http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-discuss [17] http://eavesdrop.openstack.org/#Technical_Committee_Meeting -gmann From haiwu.us at gmail.com Sat Dec 18 19:45:02 2021 From: haiwu.us at gmail.com (hai wu) Date: Sat, 18 Dec 2021 13:45:02 -0600 Subject: Networks for different availability zones in Horizon Message-ID: In Horizon "Launch Instance" workflow, in its 4th step of selecting networks for the new instance, it would always show all the available networks from all availability zones, regardless of which "Availability Zone" got picked in 1st step of "Details". I tried to update some DB field for availability zone hint for the relevant network, and that did not help. Any way to ensure in Horizon "Launch Instance" GUI workflow, after a user picking one availability zone in step 1, it would only show the related networks in that availability zone as available networks in step 4? From haiwu.us at gmail.com Sat Dec 18 20:22:30 2021 From: haiwu.us at gmail.com (hai wu) Date: Sat, 18 Dec 2021 14:22:30 -0600 Subject: [horizon] Networks for different availability zones in Horizon In-Reply-To: <1D49E522-36B5-4A4A-8166-7A48A8BB56D5@getmailspring.com> References: <1D49E522-36B5-4A4A-8166-7A48A8BB56D5@getmailspring.com> Message-ID: I was not aware of this. I could send another new email with [horizon] tag in the subject line. On Sat, Dec 18, 2021 at 1:56 PM Jimmy McArthur wrote: > > Note that you might want to tag the project in the subject field like so. > > On Dec 18 2021, at 1:45 pm, hai wu wrote: > > In Horizon "Launch Instance" workflow, in its 4th step of selecting > networks for the new instance, it would always show all the available > networks from all availability zones, regardless of which > "Availability Zone" got picked in 1st step of "Details". > > I tried to update some DB field for availability zone hint for the > relevant network, and that did not help. > > Any way to ensure in Horizon "Launch Instance" GUI workflow, after a > user picking one availability zone in step 1, it would only show the > related networks in that availability zone as available networks in > step 4? From haiwu.us at gmail.com Sat Dec 18 20:23:35 2021 From: haiwu.us at gmail.com (hai wu) Date: Sat, 18 Dec 2021 14:23:35 -0600 Subject: [horizon] Networks for different availability zones in Horizon Message-ID: In Horizon "Launch Instance" workflow, in its 4th step of selecting networks for the new instance, it would always show all the available networks from all availability zones, regardless of which "Availability Zone" got picked in 1st step of "Details". I tried to update some DB field for availability zone hint for the relevant network, and that did not help. Any way to ensure in Horizon "Launch Instance" GUI workflow, after a user picking one availability zone in step 1, it would only show the related networks in that availability zone as available networks in step 4? From smooney at redhat.com Sun Dec 19 00:31:13 2021 From: smooney at redhat.com (Sean Mooney) Date: Sun, 19 Dec 2021 00:31:13 +0000 Subject: Networks for different availability zones in Horizon In-Reply-To: References: Message-ID: <040f6fdaeca76a854ee70b238e6e63c553d290eb.camel@redhat.com> On Sat, 2021-12-18 at 13:45 -0600, hai wu wrote: > In Horizon "Launch Instance" workflow, in its 4th step of selecting > networks for the new instance, it would always show all the available > networks from all availability zones, regardless of which > "Availability Zone" got picked in 1st step of "Details". > > I tried to update some DB field for availability zone hint for the > relevant network, and that did not help. > > Any way to ensure in Horizon "Launch Instance" GUI workflow, after a > user picking one availability zone in step 1, it would only show the > related networks in that availability zone as available networks in > step 4? networks do not have any affinity to Avaiablity Zones. there is no mapping between neutron physnets and nova host aggreates which are use to model aviablityitz zones. when using tunneled networks like vxlan it is assuems that all hosts in a cloud acorss all avaiablity zones can access any tenant vxlan network. the same is also ture of vlan or flat network the only excption being that htey have physnet assocaited with them. physnets may not be aviable on all host but there is corralation between phsynets and aviaablity zones. unless you manually algin them. > From haiwu.us at gmail.com Sun Dec 19 02:44:58 2021 From: haiwu.us at gmail.com (hai wu) Date: Sat, 18 Dec 2021 20:44:58 -0600 Subject: Networks for different availability zones in Horizon In-Reply-To: <040f6fdaeca76a854ee70b238e6e63c553d290eb.camel@redhat.com> References: <040f6fdaeca76a854ee70b238e6e63c553d290eb.camel@redhat.com> Message-ID: Not using vxlan, so there's no network span across availability zones. It seems there's no built-in way to configure Horizon to do so. The problem with this is that sometimes if user picks the wrong network and availability zone combo, the VM will end up in the wrong state. This is why it would be ideal to only show available networks for the availability zone that the user already picked. I am wondering how to modify the horizon source code to achieve this, but I am not familiar with Angular, and the workflow code is pretty complex for newbie .. On Sat, Dec 18, 2021 at 6:31 PM Sean Mooney wrote: > > On Sat, 2021-12-18 at 13:45 -0600, hai wu wrote: > > In Horizon "Launch Instance" workflow, in its 4th step of selecting > > networks for the new instance, it would always show all the available > > networks from all availability zones, regardless of which > > "Availability Zone" got picked in 1st step of "Details". > > > > I tried to update some DB field for availability zone hint for the > > relevant network, and that did not help. > > > > Any way to ensure in Horizon "Launch Instance" GUI workflow, after a > > user picking one availability zone in step 1, it would only show the > > related networks in that availability zone as available networks in > > step 4? > > networks do not have any affinity to Avaiablity Zones. > there is no mapping between neutron physnets and nova host aggreates which are use to > model aviablityitz zones. > > when using tunneled networks like vxlan it is assuems that all hosts in a cloud acorss all avaiablity zones can access > any tenant vxlan network. the same is also ture of vlan or flat network the only excption being that htey have physnet > assocaited with them. physnets may not be aviable on all host but there is corralation between phsynets and aviaablity zones. > unless you manually algin them. > > > From jimmy at openinfra.dev Sat Dec 18 19:56:38 2021 From: jimmy at openinfra.dev (Jimmy McArthur) Date: Sat, 18 Dec 2021 13:56:38 -0600 Subject: [horizon] Networks for different availability zones in Horizon In-Reply-To: References: Message-ID: <1D49E522-36B5-4A4A-8166-7A48A8BB56D5@getmailspring.com> Note that you might want to tag the project in the subject field like so. On Dec 18 2021, at 1:45 pm, hai wu wrote: > In Horizon "Launch Instance" workflow, in its 4th step of selecting > networks for the new instance, it would always show all the available > networks from all availability zones, regardless of which > "Availability Zone" got picked in 1st step of "Details". > > I tried to update some DB field for availability zone hint for the > relevant network, and that did not help. > > Any way to ensure in Horizon "Launch Instance" GUI workflow, after a > user picking one availability zone in step 1, it would only show the > related networks in that availability zone as available networks in > step 4? > -------------- next part -------------- An HTML attachment was scrubbed... URL: From miguel at mlavalle.com Sun Dec 19 23:34:17 2021 From: miguel at mlavalle.com (Miguel Lavalle) Date: Sun, 19 Dec 2021 17:34:17 -0600 Subject: [neutron] bug deputy report December 13th to 19th Message-ID: Hi, Here's this week's bugs deputy report: High ==== https://bugs.launchpad.net/neutron/+bug/1954666 "LoggingPlugin._clean_logs_by_target_id" failing in stable releases. Assigned to Rodolfo Alonso. Fixes: https://review.opendev.org/c/openstack/neutron/+/821563, https://review.opendev.org/c/openstack/neutron/+/821564, https://review.opendev.org/c/openstack/neutron/+/821565 Medium ====== https://bugs.launchpad.net/neutron/+bug/1954763 Creating ports in bulks can be very slow due to many IPAM module conflicts. Assigned to Slawek Kaplonski. Proposed fix: https://review.opendev.org/c/openstack/neutron/+/821727 https://bugs.launchpad.net/neutron/+bug/1954785 Session semantic violated when sending FloatingIP disassociation update during port deletion. Assigned to Szymon Wr?blewski https://bugs.launchpad.net/neutron/+bug/1954942 [fullstack] Failure of "test_reschedule_network_on_new_agent". Needs owner https://bugs.launchpad.net/neutron/+bug/1955008 [functional] Failure of "test_floatingip_mac_bindings". Needs owner https://bugs.launchpad.net/neutron/+bug/1955010 [stable] py27 is not supported in "python-lazy-object-proxy" release 1.7.0. Assigned to Rodolfo Alonso Low ==== https://bugs.launchpad.net/neutron/+bug/1954903 [OVN] Add floating IP pools L3 extension to OVN L3 service. Assigned to Rodolfo Alonso. Proposed fix https://review.opendev.org/c/openstack/neutron/+/821818 https://bugs.launchpad.net/neutron/+bug/1955121 Update subnet with same gateway_ip as is already used should be allowed. Assigned to Slawek Kaplonski. Proposed fix: https://review.opendev.org/c/openstack/neutron/+/822098 Waiting for response from submitter =========================== https://bugs.launchpad.net/neutron/+bug/1954771 [OVN]MAC of lrp has not been updated in MAC_Binding when it re-associated. https://bugs.launchpad.net/neutron/+bug/1954777 setting same static route with subnet already exists as direct connected network Cheers Miguel -------------- next part -------------- An HTML attachment was scrubbed... URL: From katonalala at gmail.com Mon Dec 20 07:15:54 2021 From: katonalala at gmail.com (Lajos Katona) Date: Mon, 20 Dec 2021 08:15:54 +0100 Subject: [neutron] PTL on PTO In-Reply-To: References: <3143668.aeNJFYEL58@p1> Message-ID: Hi Miguel, Yes, they are canceled, sorry if I was not clear. Miguel Lavalle ezt ?rta (id?pont: 2021. dec. 16., Cs, 0:59): > Hi Lajos, > > I hope you enjoy your time off. Are the Neutron and Drivers meetings > canceled for the rest of 2021? > > Best regards > > On Wed, Dec 15, 2021 at 12:41 PM Slawek Kaplonski > wrote: > >> Hi, >> >> On ?roda, 15 grudnia 2021 18:21:59 CET Lajos Katona wrote: >> > Hi, >> > I will be on PTO, back in Tuesday, 04. 01. 2022. >> > Don't hesitate to write, I will be reachable. >> > >> > Regards >> > Lajos Katona (lajoskatona) >> >> Merry Christmas and Happy New Year! >> Take your time and rest during that holiday season! >> >> -- >> Slawek Kaplonski >> Principal Software Engineer >> Red Hat > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From gthiemonge at redhat.com Mon Dec 20 08:38:55 2021 From: gthiemonge at redhat.com (Gregory Thiemonge) Date: Mon, 20 Dec 2021 09:38:55 +0100 Subject: [octavia] cancelling weekly meetings until 2022 Message-ID: Hi, As discussed during our last weekly meeting, the next 2 meetings (Dec 22nd and 29th) are cancelled. Our next meeting will be on January 5th, Greg -------------- next part -------------- An HTML attachment was scrubbed... URL: From mkopec at redhat.com Mon Dec 20 09:46:56 2021 From: mkopec at redhat.com (Martin Kopec) Date: Mon, 20 Dec 2021 10:46:56 +0100 Subject: [qa] Cancelling office hour on Dec 28th due to holidays Message-ID: Hi all, we'll have the last office hour in 2021 tomorrow (Dec. 21st). Our first office hour in 2022 will be held on Dec. 4th. Merry Christmas and Happy New Year! Regards, -- Martin Kopec Senior Software Quality Engineer Red Hat EMEA -------------- next part -------------- An HTML attachment was scrubbed... URL: From radoslaw.piliszek at gmail.com Mon Dec 20 10:09:55 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Mon, 20 Dec 2021 11:09:55 +0100 Subject: [qa] Cancelling office hour on Dec 28th due to holidays In-Reply-To: References: Message-ID: On Mon, 20 Dec 2021 at 10:53, Martin Kopec wrote: > we'll have the last office hour in 2021 tomorrow (Dec. 21st). Our first office hour in 2022 will be held on Dec. 4th. Whoa, that's almost a year from now! > Merry Christmas and Happy New Year! Merry Christmas and Happy New Year! -yoctozepto From syedammad83 at gmail.com Mon Dec 20 10:10:34 2021 From: syedammad83 at gmail.com (Ammad Syed) Date: Mon, 20 Dec 2021 15:10:34 +0500 Subject: [nova] random unauthorized errors Message-ID: Hi, I am using nova 23.0.0 and I am randomly facing unauthorized error sometime when nova uses neutron or cinder phython3 client during listing of server and creation of server. When I list server again, it works fine. I have found below logs in nova-api. 2021-12-20 09:10:24.920 2689471 DEBUG nova.api.openstack.wsgi [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Calling method '>' _process_stack /usr/lib/python3/dist-packages/nova/api/openstack/wsgi.py:513 2021-12-20 09:10:24.925 2689471 DEBUG nova.compute.api [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Searching by: {'project_id': '890eb2b7d1b8488aa88de7c34d08817a', 'deleted': False} get_all /usr/lib/python3/dist-packages/nova/compute/api.py:2857 2021-12-20 09:10:24.930 2689471 DEBUG oslo_concurrency.lockutils [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s inner /usr/lib/python3/dist-packages/oslo_concurrency/lockutils.py:355 2021-12-20 09:10:24.931 2689471 DEBUG oslo_concurrency.lockutils [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default def2021-12-20 09:10:24.548 2689471 INFO nova.osapi_compute.wsgi.server [req-c5da1a3c-3c99-46c2-84e9-b504ccc5d156 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] 172.16.30.56 "GET /v2.1/ HTTP/1.1" status: 200 len: 800 time: 0.0048089 2021-12-20 09:10:24.697 2689471 DEBUG nova.osapi_compute.wsgi.server [req-572eca81-8d3b-4524-89e5-e4e1124a7ac5 - - - - -] (2689471) accepted ('172.16.30.56', 52016) server /usr/lib/pytho n3/dist-packages/eventlet/wsgi.py:992 2021-12-20 09:10:24.920 2689471 DEBUG nova.api.openstack.wsgi [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default defaul t] Calling method '>' _process_stack /usr/lib/python3/dist-packag es/nova/api/openstack/wsgi.py:513 2021-12-20 09:10:24.925 2689471 DEBUG nova.compute.api [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Sear ching by: {'project_id': '890eb2b7d1b8488aa88de7c34d08817a', 'deleted': False} get_all /usr/lib/python3/dist-packages/nova/compute/api.py:2857 2021-12-20 09:10:24.930 2689471 DEBUG oslo_concurrency.lockutils [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default def ault] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s inner /usr/lib/python3/d ist-packages/oslo_concurrency/lockutils.py:355 2021-12-20 09:10:24.931 2689471 DEBUG oslo_concurrency.lockutils [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default def ault] Lock "00000000-0000-0000-0000-000000000000" released by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s inner /usr/lib/python3/dist-packages/oslo_concurrency/lockutils.py:367 2021-12-20 09:10:24.933 2689471 DEBUG oslo_concurrency.lockutils [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Lock "fc805771-7891-4095-8162-491cf824b02f" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s inner /usr/lib/python3/dist-packages/oslo_concurrency/lockutils.py:355 2021-12-20 09:10:24.934 2689471 DEBUG oslo_concurrency.lockutils [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Lock "fc805771-7891-4095-8162-491cf824b02f" released by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s inner /usr/lib/python3/dist-packages/oslo_concurrency/lockutils.py:367 2021-12-20 09:10:25.019 2689471 DEBUG nova.compute.multi_cell_list [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Listed batch of 13 results from cell out of 21 limit. Returned 13 total so far. do_query /usr/lib/python3/dist-packages/nova/compute/multi_cell_list.py:371 2021-12-20 09:10:25.093 2689471 DEBUG oslo_concurrency.lockutils [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Lock "fc805771-7891-4095-8162-491cf824b02f" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s inner /usr/lib/python3/dist-packages/oslo_concurrency/lockutils.py:355 2021-12-20 09:10:25.094 2689471 DEBUG oslo_concurrency.lockutils [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Lock "fc805771-7891-4095-8162-491cf824b02f" released by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s inner /usr/lib/python3/dist-packages/oslo_concurrency/lockutils.py:367 *2021-12-20 09:10:25.393 2689471 DEBUG neutronclient.v2_0.client [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Error message: {"error": {"code": 401, "title": "Unauthorized", "message": "The request you have made requires authentication."}} _handle_fault_response /usr/lib/python3/dist-packages/neutronclient/v2_0/client.py:258* 2021-12-20 09:10:25.394 2689471 ERROR nova.network.neutron [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Neutron client was not able to generate a valid admin token, please verify Neutron admin credential located in nova.conf: neutronclient.common.exceptions.Unauthorized: 401-{'error': {'code': 401, 'title': 'Unauthorized', 'message': 'The request you have made requires authentication.'}} 2021-12-20 09:10:25.395 2689471 INFO nova.api.openstack.wsgi [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] HTTP exception thrown: Networking client is experiencing an unauthorized exception. 2021-12-20 09:10:25.396 2689471 DEBUG nova.api.openstack.wsgi [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] Returning 400 to user: Networking client is experiencing an unauthorized exception. __call__ /usr/lib/python3/dist-packages/nova/api/openstack/wsgi.py:927 2021-12-20 09:10:25.398 2689471 INFO nova.osapi_compute.wsgi.server [req-91ee9c88-e8ab-4e9b-8e1a-ca8a807c7e45 2af528fdf3244e15b4f3f8fcfc0889c5 890eb2b7d1b8488aa88de7c34d08817a - default default] 172.16.30.56 "GET /v2.1/890eb2b7d1b8488aa88de7c34d08817a/servers/detail?limit=21&project_id=890eb2b7d1b8488aa88de7c34d08817a&sort_dir=desc&sort_dir=desc&sort_dir=desc&sort_key=created_at&sort_key=display_name&sort_key=uuid HTTP/1.1" status: 400 len: 529 time: 0.7002470 Is there anything that I need to change ? -- Regards, Syed Ammad Ali -------------- next part -------------- An HTML attachment was scrubbed... URL: From mkopec at redhat.com Mon Dec 20 13:06:50 2021 From: mkopec at redhat.com (Martin Kopec) Date: Mon, 20 Dec 2021 14:06:50 +0100 Subject: [qa] Cancelling office hour on Dec 28th due to holidays In-Reply-To: References: Message-ID: oh, that was supposed to be January 4th :D, I'm sorry for the mistake. On Mon, 20 Dec 2021 at 11:10, Rados?aw Piliszek wrote: > On Mon, 20 Dec 2021 at 10:53, Martin Kopec wrote: > > we'll have the last office hour in 2021 tomorrow (Dec. 21st). Our first > office hour in 2022 will be held on Dec. 4th. > > Whoa, that's almost a year from now! > > > Merry Christmas and Happy New Year! > > Merry Christmas and Happy New Year! > > -yoctozepto > > -- Martin -------------- next part -------------- An HTML attachment was scrubbed... URL: From sbauza at redhat.com Mon Dec 20 13:08:16 2021 From: sbauza at redhat.com (Sylvain Bauza) Date: Mon, 20 Dec 2021 14:08:16 +0100 Subject: [nova][placement] PTL on PTO until Jan 3rd Message-ID: Hi folks, I'll be off the keyboard from tonight and only be back on Monday January the 3rd. Have good holidays if you have them and good rest. For others, thanks for continuing to help Nova and see you again in 2022 ! -Sylvain -------------- next part -------------- An HTML attachment was scrubbed... URL: From akekane at redhat.com Mon Dec 20 14:10:44 2021 From: akekane at redhat.com (Abhishek Kekane) Date: Mon, 20 Dec 2021 19:40:44 +0530 Subject: [Glance] No weekly meeting until New Year Message-ID: Hello All, Christmas and New year holidays are approaching and most of the glance team is on vacation so we are canceling our next two weekly meetings. Next meeting will be on 06th January, 2022. Merry Christmas and Happy New year in advance!!! Cheers, Abhishek Kekane -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Mon Dec 20 17:33:56 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Mon, 20 Dec 2021 11:33:56 -0600 Subject: [gate][tempest][plugins][stable] Stable/train tempest plugins jobs are broken Message-ID: <17dd8e834d9.c0af5eee46883.3204950283589713289@ghanshyammann.com> Hello Everyone, gthiemonge reported the issue for tempest plugin stable/train jobs are broken. I have explained the issue in https://bugs.launchpad.net/tempest/+bug/1955418 and fix is up (it is under testing). This is applicable for any tempest plugin job on stable/train. Please hold the recheck until the below fix is merged: https://review.opendev.org/c/openstack/tempest/+/822339 -gmann From d.gweon at samsung.com Mon Dec 20 06:45:59 2021 From: d.gweon at samsung.com (=?UTF-8?B?6raM64uk7KeE?=) Date: Mon, 20 Dec 2021 15:45:59 +0900 Subject: Inquiry on FWaaS Deprecation References: Message-ID: <20211220064559epcms3p142fa295de7db1d587e2c8295c5c02fae@epcms3p1> An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: not available Type: image/gif Size: 13402 bytes Desc: not available URL: From radoslaw.piliszek at gmail.com Mon Dec 20 18:35:27 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Mon, 20 Dec 2021 19:35:27 +0100 Subject: [kolla] Dropping support for vmtp in Yoga In-Reply-To: References: Message-ID: Dears, As there have been no replies, I have proposed two patches that drop support for vmtp in kolla [1] and kolla-ansible [2]. [1] https://review.opendev.org/c/openstack/kolla/+/822351 [2] https://review.opendev.org/c/openstack/kolla-ansible/+/822349 Kind regards, -yoctozepto On Wed, 8 Dec 2021 at 16:39, Rados?aw Piliszek wrote: > > Dear Users of Kolla! > > The Kolla project, so far, provided vmtp [1] image and ansible role. > However, the status of those has been unknown (the ansible role looks > pretty stale, still referencing ubuntu 16.04!) and now the vmtp image > is failing to build so we are forced to disable it. [2] > The project itself [1] looks rather unmaintained and it's outside of > the openstack namespace so our plan is to drop support for it entirely > in this cycle (Yoga). > > Please reply to this mail if you are using vmtp with Kolla in some > recent release and are willing to contribute to let us keep supporting > it. > > [1] https://opendev.org/x/vmtp > [2] https://review.opendev.org/c/openstack/kolla/+/820043 > > Kind regards, > -yoctozepto From radoslaw.piliszek at gmail.com Mon Dec 20 19:09:54 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Mon, 20 Dec 2021 20:09:54 +0100 Subject: [kolla] The impact of CentOS Linux 8 EOL and OpenDev's drop of the respective image Message-ID: Dears, As you may recall, the CentOS Linux 8 (as opposed to CentOS Stream 8) is reaching its EOL (End Of Life) Dec 31st, 2021 (that is, in eleven days). [1] Due to that, the OpenDev team (that hosts kolla's repos and the CI/CD that we rely on) has decided to drop the EOLed images January 14th, 2022. [2] This impacts both Kolla and Kolla Ansible in subtle ways - both operator-wise and contributor-wise. The EM (Extended Maintenance) branches, Train and Ussuri, in both projects, rely solely on CentOS Linux 8, the Stream variant is *not* supported there. This means that in 2022 we are forced to cease to test these branches against CentOS Linux 8 (we will still test the two other distros - Ubuntu and Debian). Operators are advised to upgrade to Victoria or (preferably) later (psst, go Xena). Furthermore, CentOS Linux 8 is currently supported in Victoria but this support will also end naturally with 2022 on the horizon. Operators are advised to migrate to CentOS Stream 8 and images based on it. [1] https://www.centos.org/centos-linux-eol/ [2] http://lists.opendev.org/pipermail/service-announce/2021-December/000029.html Kind regards, -yoctozepto From james.slagle at gmail.com Mon Dec 20 20:25:58 2021 From: james.slagle at gmail.com (James Slagle) Date: Mon, 20 Dec 2021 15:25:58 -0500 Subject: [TripleO] Tomorrow's (Dec 21) meeting cancelled, next meeting Jan 3 Message-ID: As we have no items on the agenda[1] for tomorrow's meeting, let's cancel it and hold our next meeting on Tuesday, January 3rd. Thanks! [1] https://etherpad.opendev.org/p/tripleo-meeting-items -- -- James Slagle -- -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Mon Dec 20 23:31:10 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Mon, 20 Dec 2021 17:31:10 -0600 Subject: [all][tc] Technical Committee next weekly meeting on Dec 23rd at 1500 UTC Message-ID: <17dda2f42bc.dbd5816756818.7976500778946876675@ghanshyammann.com> Hello Everyone, Technical Committee's next weekly meeting is scheduled for Dec 23rd at 1500 UTC. If you would like to add topics for discussion, please add them to the below wiki page by Wednesday, Dec 22nd, at 2100 UTC. https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting -gmann From jonmills at gmail.com Tue Dec 21 15:58:47 2021 From: jonmills at gmail.com (Jonathan Mills) Date: Tue, 21 Dec 2021 10:58:47 -0500 Subject: Networks for different availability zones in Horizon In-Reply-To: References: <040f6fdaeca76a854ee70b238e6e63c553d290eb.camel@redhat.com> Message-ID: <3863DFA9-9573-431F-B91A-51C4421926DD@me.com> Thank you, Hai Wu, for raising the visibility of this issue. We at NASA Goddard want to echo this sentiment wholeheartedly. It?s a serious pain point! Our production OpenStack (Wallaby) cloud spans two AZs (different datacenters) right now, and before we?re done we may have 3 or 4 AZs. We use non-overlapping VLAN-based networks in different datacenters. Moreover, we have different kinds of server hardware in each datacenter (often recycled HPC compute nodes). We also have different Cinder storage backends in each datacenter. What we end up having is AZs in Nova, AZs in Neutron, AZs in Cinder, and Image Flavors for Glance that are rather AZ dependent (in order to best fit the node geometry). We instantiate these objects with explicit AZs when they support it. In the case of Neutron networks, we can use scheduler hints. The real crux of the problem is the end-user education though. Because OpenStack clients (Horizon, OSC) are perfectly happy to let the end-user try to build impossible combinations. We strongly feel that if the individual services aren?t going to communicate AZ data to each other via RPC, that the clients at least should do some kind of filtering on AZs or scheduler hints about AZs. I'm not entirely sure how you'd solve this problem on CLI easily without RPC communication, but Horizon could (should?) dynamically filter the different menus as you select options: e.g. the first selection screen ?Details? has you choose an AZ. Once the user makes a selection there, the flavors, storage, and network options should adjust accordingly. Jonathan > On Dec 18, 2021, at 9:44 PM, hai wu wrote: > > Not using vxlan, so there's no network span across availability zones. > It seems there's no built-in way to configure Horizon to do so. The > problem with this is that sometimes if user picks the wrong network > and availability zone combo, the VM will end up in the wrong state. > This is why it would be ideal to only show available networks for the > availability zone that the user already picked. > > I am wondering how to modify the horizon source code to achieve this, > but I am not familiar with Angular, and the workflow code is pretty > complex for newbie .. > > On Sat, Dec 18, 2021 at 6:31 PM Sean Mooney wrote: >> >> On Sat, 2021-12-18 at 13:45 -0600, hai wu wrote: >>> In Horizon "Launch Instance" workflow, in its 4th step of selecting >>> networks for the new instance, it would always show all the available >>> networks from all availability zones, regardless of which >>> "Availability Zone" got picked in 1st step of "Details". >>> >>> I tried to update some DB field for availability zone hint for the >>> relevant network, and that did not help. >>> >>> Any way to ensure in Horizon "Launch Instance" GUI workflow, after a >>> user picking one availability zone in step 1, it would only show the >>> related networks in that availability zone as available networks in >>> step 4? >> >> networks do not have any affinity to Avaiablity Zones. >> there is no mapping between neutron physnets and nova host aggreates which are use to >> model aviablityitz zones. >> >> when using tunneled networks like vxlan it is assuems that all hosts in a cloud acorss all avaiablity zones can access >> any tenant vxlan network. the same is also ture of vlan or flat network the only excption being that htey have physnet >> assocaited with them. physnets may not be aviable on all host but there is corralation between phsynets and aviaablity zones. >> unless you manually algin them. >>> >> > From lokendrarathour at gmail.com Tue Dec 21 07:45:25 2021 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Tue, 21 Dec 2021 13:15:25 +0530 Subject: [Triple0] Unable to pxeBoot overcloud Nodes via ipv6 Message-ID: Hi Team, we have installed Undercloud with Train release on Centos8 using ipv6 as loca_ip in undercloud.conf file. Undercloud seems to be installed successfully and we are able to see IPv6 endpoints as well. Also, the OpenStack overcloud images are uploaded and well visible on http://:8088 tftp link. Problem Statement: During introspection start, we see that ipv6 is allocated to the node, further no action is being performed. Please refer the screen attached. [image: image.png] As we can see ipv6 bbbb:bbbb:bbbb::9 is allocated to the desired machine. Furthermore, we see that the link is shown as down for the same mac, which also becomes up as can be seen but the configuring on that NIC eventually fails for some reason. Any idea/inputs on this please? -- ~ Lokendra -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: image.png Type: image/png Size: 135673 bytes Desc: not available URL: From qinhaizhong01 at inspur.com Tue Dec 21 12:09:47 2021 From: qinhaizhong01 at inspur.com (=?gb2312?B?RGF6aG9uZyBRaW4gKMfYuqPW0Ckt1MbK/b7d1tDQxLyvzcU=?=) Date: Tue, 21 Dec 2021 12:09:47 +0000 Subject: Can neutron-fwaas project be revived? Message-ID: Hi? The firewall project is a necessary function when the project is delivered. The lack of firewall function after switching OVN is not acceptable to customers. We intend to maintain this project and develop the fwaas driver based on ovn. Whether the neutron-fwaas project can be reactivate? What should I do ? -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3615 bytes Desc: not available URL: From rosmaita.fossdev at gmail.com Tue Dec 21 16:27:41 2021 From: rosmaita.fossdev at gmail.com (Brian Rosmaita) Date: Tue, 21 Dec 2021 11:27:41 -0500 Subject: [cinder] weekly meetings for the remainder of december Message-ID: As agreed at last week's cinder meeting, our meetings for the remainder of 2021 will be: * Wednesday, December 22: no meeting * Wednesday, December 29: no meeting The next weekly cinder meeting will be Wednesday, January 5 at the usual time (1400 UTC) in the usual place (#openstack-meeting-alt on OFTC). The meeting agenda is here: https://etherpad.opendev.org/p/cinder-yoga-meetings cheers, brian From erin at openstack.org Tue Dec 21 17:05:02 2021 From: erin at openstack.org (Erin Disney) Date: Tue, 21 Dec 2021 11:05:02 -0600 Subject: OpenInfra Summit Berlin 2022 Programming Committee Nominations Now Open Message-ID: <4A25C8ED-9FAE-4844-99FF-EE9368EE45F0@openstack.org> Hey everyone- Programming Committee nominations for the 2022 OpenInfra Summit in Berlin (June 7-9, 2022) are now open! Programming Committees for each Track will help build the Summit schedule, and are made up of individuals working in open infrastructure. Responsibilities include: Help the Summit team put together the best possible content based on your subject matter expertise Promote the individual Tracks within your networks Review the submissions and Community voting results in your particular Track Determine if there are any major content gaps in your Track, and if so, potentially solicit additional speakers directly to submit Ensure diversity of speakers and companies represented in your Track Avoid vendor sales pitches, focusing more on real-world user stories and technical, in-the-trenches experiences Identify which submissions would make good content for OpenInfra Live, the virtual show hosted by the OpenInfra Foundation that encourages live questions from a global audience 2022 Summit Tracks: 5G, NFV & Edge AI, Machine Learning & HPC CI/CD Container Infrastructure Getting Started Hands-on Workshops NEW: Hardware Enablement Open Development Private & Hybrid Cloud Public Cloud Security Full track descriptions are available here . If you?re interested in nominating yourself or someone else to be a member of the Summit Programming Committee for a specific Track, please fill out the nomination form . Nominations will close on January 19, 2022. Programming Committee selections will occur before we close the Call for Presentations (CFP) so that the Committees can host office hours to consult on submissions, and help promote the event. CFP will be opening in January, registration and sponsorship information are already available. Please email speakersupport at openinfra.dev with any questions or feedback. Cheers, Erin Disney Event Marketing Open Infrastructure Foundation -------------- next part -------------- An HTML attachment was scrubbed... URL: From dev.faz at gmail.com Wed Dec 22 07:24:37 2021 From: dev.faz at gmail.com (Fabian Zimmermann) Date: Wed, 22 Dec 2021 08:24:37 +0100 Subject: Networks for different availability zones in Horizon In-Reply-To: <3863DFA9-9573-431F-B91A-51C4421926DD@me.com> References: <040f6fdaeca76a854ee70b238e6e63c553d290eb.camel@redhat.com> <3863DFA9-9573-431F-B91A-51C4421926DD@me.com> Message-ID: Hi, maybe this helps: https://docs.openstack.org/neutron/latest/admin/config-routed-networks.html You should be able to define one-network-to-used-for-all and openstack is doing the magic in the background. Fabian Am Di., 21. Dez. 2021 um 17:15 Uhr schrieb Jonathan Mills : > > Thank you, Hai Wu, for raising the visibility of this issue. We at NASA Goddard want to echo this sentiment wholeheartedly. It?s a serious pain point! > > Our production OpenStack (Wallaby) cloud spans two AZs (different datacenters) right now, and before we?re done we may have 3 or 4 AZs. We use non-overlapping VLAN-based networks in different datacenters. Moreover, we have different kinds of server hardware in each datacenter (often recycled HPC compute nodes). We also have different Cinder storage backends in each datacenter. > > What we end up having is AZs in Nova, AZs in Neutron, AZs in Cinder, and Image Flavors for Glance that are rather AZ dependent (in order to best fit the node geometry). We instantiate these objects with explicit AZs when they support it. In the case of Neutron networks, we can use scheduler hints. > > The real crux of the problem is the end-user education though. Because OpenStack clients (Horizon, OSC) are perfectly happy to let the end-user try to build impossible combinations. We strongly feel that if the individual services aren?t going to communicate AZ data to each other via RPC, that the clients at least should do some kind of filtering on AZs or scheduler hints about AZs. I'm not entirely sure how you'd solve this problem on CLI easily without RPC communication, but Horizon could (should?) dynamically filter the different menus as you select options: e.g. the first selection screen ?Details? has you choose an AZ. Once the user makes a selection there, the flavors, storage, and network options should adjust accordingly. > > > Jonathan > > > > On Dec 18, 2021, at 9:44 PM, hai wu wrote: > > > > Not using vxlan, so there's no network span across availability zones. > > It seems there's no built-in way to configure Horizon to do so. The > > problem with this is that sometimes if user picks the wrong network > > and availability zone combo, the VM will end up in the wrong state. > > This is why it would be ideal to only show available networks for the > > availability zone that the user already picked. > > > > I am wondering how to modify the horizon source code to achieve this, > > but I am not familiar with Angular, and the workflow code is pretty > > complex for newbie .. > > > > On Sat, Dec 18, 2021 at 6:31 PM Sean Mooney wrote: > >> > >> On Sat, 2021-12-18 at 13:45 -0600, hai wu wrote: > >>> In Horizon "Launch Instance" workflow, in its 4th step of selecting > >>> networks for the new instance, it would always show all the available > >>> networks from all availability zones, regardless of which > >>> "Availability Zone" got picked in 1st step of "Details". > >>> > >>> I tried to update some DB field for availability zone hint for the > >>> relevant network, and that did not help. > >>> > >>> Any way to ensure in Horizon "Launch Instance" GUI workflow, after a > >>> user picking one availability zone in step 1, it would only show the > >>> related networks in that availability zone as available networks in > >>> step 4? > >> > >> networks do not have any affinity to Avaiablity Zones. > >> there is no mapping between neutron physnets and nova host aggreates which are use to > >> model aviablityitz zones. > >> > >> when using tunneled networks like vxlan it is assuems that all hosts in a cloud acorss all avaiablity zones can access > >> any tenant vxlan network. the same is also ture of vlan or flat network the only excption being that htey have physnet > >> assocaited with them. physnets may not be aviable on all host but there is corralation between phsynets and aviaablity zones. > >> unless you manually algin them. > >>> > >> > > > > From skaplons at redhat.com Wed Dec 22 07:31:33 2021 From: skaplons at redhat.com (Slawek Kaplonski) Date: Wed, 22 Dec 2021 08:31:33 +0100 Subject: [neutron] CI meeting 28.12.2021 cancelled Message-ID: <13096112.O9o76ZdvQC@p1> Hi, Next week Neutron CI meeting is cancelled. Happy holidays for all of You and see You on the meetings in 2022! We will start with CI meeting 4.01.2022 :) -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From bkslash at poczta.onet.pl Wed Dec 22 08:29:41 2021 From: bkslash at poczta.onet.pl (Adam Tomas) Date: Wed, 22 Dec 2021 09:29:41 +0100 Subject: [kolla-ansible][cinder] Message-ID: Hi, I?d like to use MPIO in connection to storage array, so I have two physical NIC?s in different VLANS on each node and also 4 NIC?s in the storage array (2 in each storage VLAN) so it should result in 4 MPIO paths from every node (2 paths in each VLAN). But how to set it in kolla-ansible? Can I point to more than one interface using storage_interface variable? Or maybe I should configure it in different way? Best regards Adam Tomas From ruslanas at lpic.lt Wed Dec 22 11:22:04 2021 From: ruslanas at lpic.lt (=?UTF-8?Q?Ruslanas_G=C5=BEibovskis?=) Date: Wed, 22 Dec 2021 13:22:04 +0200 Subject: [tripleo][ussuri] Log4j add protection to haproxy {even it is not impacted In-Reply-To: References: Message-ID: yes, but I am just curious how to add it during deployment step. Log4j is just an example. And in our case, OSP does not have Log4j, our sec team is very new, so they are excited, as it is the first zeroday for them ;) On Fri, 17 Dec 2021 at 18:49, Clark Boylan wrote: > On Fri, Dec 17, 2021, at 5:31 AM, Ruslanas G?ibovskis LPIC wrote: > > Hi team, > > > > Thanks to this Log4j I finally found time to read around, how to add > > additional settings/options to haproxy config, especially, I would like > > to apply haproxy steps to hide log4j vulns. > > > > I know I know, OSP looks not to be impacted, unless we have some > > components such as opendaylight which might have Log4j applied. > > > > Does anyone has example for yaml file? > > > > Thanks in advance. > > It is my understanding that the log messages magic syntax in log4j is > sophisticated enough that filtering via proxies is problematic and unlikely > to catch everything. You'll be filtering simple attacks and letting > sophisticated actors through. You are much better off upgrading log4j or > disabling the JNDI class entirely in the jar. I wouldn't rely on haproxy > for this. > > -- Ruslanas G?ibovskis +370 6030 7030 -------------- next part -------------- An HTML attachment was scrubbed... URL: From flux.adam at gmail.com Wed Dec 22 13:36:05 2021 From: flux.adam at gmail.com (Adam Harwell) Date: Wed, 22 Dec 2021 22:36:05 +0900 Subject: Networks for different availability zones in Horizon In-Reply-To: <3863DFA9-9573-431F-B91A-51C4421926DD@me.com> References: <040f6fdaeca76a854ee70b238e6e63c553d290eb.camel@redhat.com> <3863DFA9-9573-431F-B91A-51C4421926DD@me.com> Message-ID: +1 from Yahoo. We have a number of compute AZs in each of our clouds, and networks that are specifically crafted for those AZs (mostly due to security zones). Our users very commonly pick incompatible combinations. We have a user manual with a table that shows valid combinations in an attempt to educate them, but this is really something that would be better handled directly in the tooling somehow. Maybe this kind of setup is more common than people think, at least for large private cloud operators? ?Adam PS: I think routed networks may fix part of our configuration woes, but certainly not all. I wonder if there is a sane way to link networks to AZs like this? On Wed, Dec 22, 2021 at 1:08 Jonathan Mills wrote: > Thank you, Hai Wu, for raising the visibility of this issue. We at NASA > Goddard want to echo this sentiment wholeheartedly. It?s a serious pain > point! > > Our production OpenStack (Wallaby) cloud spans two AZs (different > datacenters) right now, and before we?re done we may have 3 or 4 AZs. We > use non-overlapping VLAN-based networks in different datacenters. > Moreover, we have different kinds of server hardware in each datacenter > (often recycled HPC compute nodes). We also have different Cinder storage > backends in each datacenter. > > What we end up having is AZs in Nova, AZs in Neutron, AZs in Cinder, and > Image Flavors for Glance that are rather AZ dependent (in order to best fit > the node geometry). We instantiate these objects with explicit AZs when > they support it. In the case of Neutron networks, we can use scheduler > hints. > > The real crux of the problem is the end-user education though. Because > OpenStack clients (Horizon, OSC) are perfectly happy to let the end-user > try to build impossible combinations. We strongly feel that if the > individual services aren?t going to communicate AZ data to each other via > RPC, that the clients at least should do some kind of filtering on AZs or > scheduler hints about AZs. I'm not entirely sure how you'd solve this > problem on CLI easily without RPC communication, but Horizon could > (should?) dynamically filter the different menus as you select options: > e.g. the first selection screen ?Details? has you choose an AZ. Once the > user makes a selection there, the flavors, storage, and network options > should adjust accordingly. > > > Jonathan > > > > On Dec 18, 2021, at 9:44 PM, hai wu wrote: > > > > Not using vxlan, so there's no network span across availability zones. > > It seems there's no built-in way to configure Horizon to do so. The > > problem with this is that sometimes if user picks the wrong network > > and availability zone combo, the VM will end up in the wrong state. > > This is why it would be ideal to only show available networks for the > > availability zone that the user already picked. > > > > I am wondering how to modify the horizon source code to achieve this, > > but I am not familiar with Angular, and the workflow code is pretty > > complex for newbie .. > > > > On Sat, Dec 18, 2021 at 6:31 PM Sean Mooney wrote: > >> > >> On Sat, 2021-12-18 at 13:45 -0600, hai wu wrote: > >>> In Horizon "Launch Instance" workflow, in its 4th step of selecting > >>> networks for the new instance, it would always show all the available > >>> networks from all availability zones, regardless of which > >>> "Availability Zone" got picked in 1st step of "Details". > >>> > >>> I tried to update some DB field for availability zone hint for the > >>> relevant network, and that did not help. > >>> > >>> Any way to ensure in Horizon "Launch Instance" GUI workflow, after a > >>> user picking one availability zone in step 1, it would only show the > >>> related networks in that availability zone as available networks in > >>> step 4? > >> > >> networks do not have any affinity to Avaiablity Zones. > >> there is no mapping between neutron physnets and nova host aggreates > which are use to > >> model aviablityitz zones. > >> > >> when using tunneled networks like vxlan it is assuems that all hosts in > a cloud acorss all avaiablity zones can access > >> any tenant vxlan network. the same is also ture of vlan or flat network > the only excption being that htey have physnet > >> assocaited with them. physnets may not be aviable on all host but there > is corralation between phsynets and aviaablity zones. > >> unless you manually algin them. > >>> > >> > > > > > -- Thanks, --Adam -------------- next part -------------- An HTML attachment was scrubbed... URL: From senrique at redhat.com Wed Dec 22 13:55:39 2021 From: senrique at redhat.com (Sofia Enriquez) Date: Wed, 22 Dec 2021 10:55:39 -0300 Subject: [cinder] Bug deputy report for week of 12-22-2021 Message-ID: No meeting today. This is a bug report from 12-15-2021 to 12-22-2021. Agenda: https://etherpad.opendev.org/p/cinder-bug-squad-meeting ----------------------------------------------------------------------------------------- Medium - https://bugs.launchpad.net/os-brick/+bug/1955020 "multipath -l is failing". Assigned to Rajat Dhasmana. Low - https://bugs.launchpad.net/cinder/+bug/1955057 "NetApp driver fails to compare ONTAP version". Assigned to Nahim Alves de Souza. Wishlist - https://bugs.launchpad.net/cinder/+bug/1955360 "Cannot clone a volume from the image within the same ceph cluster". Assigned to Eric Xie. Cheers -- Sof?a Enriquez she/her Software Engineer Red Hat PnT IRC: @enriquetaso @RedHat Red Hat Red Hat -------------- next part -------------- An HTML attachment was scrubbed... URL: From noonedeadpunk at ya.ru Wed Dec 22 15:23:16 2021 From: noonedeadpunk at ya.ru (Dmitriy Rabotyagov) Date: Wed, 22 Dec 2021 17:23:16 +0200 Subject: [requirements][OpenStackClients][rally][tempest][watcher][nova][glance] Use prettytable in requirements Message-ID: <4141640184867@mail.yandex.ru> An HTML attachment was scrubbed... URL: From fungi at yuggoth.org Wed Dec 22 15:46:27 2021 From: fungi at yuggoth.org (Jeremy Stanley) Date: Wed, 22 Dec 2021 15:46:27 +0000 Subject: [requirements][OpenStackClients][rally][tempest][watcher][nova][glance] Use prettytable in requirements In-Reply-To: <4141640184867@mail.yandex.ru> References: <4141640184867@mail.yandex.ru> Message-ID: <20211222154626.mpmnvc5l4oeb6clx@yuggoth.org> On 2021-12-22 17:23:16 +0200 (+0200), Dmitriy Rabotyagov wrote: > Nowadays, PrettyTable package has been replaced with prettytable > in PyPi. In the meanwhile in global-requirements we still have old > name being mentioned[1]. This brings unpleasant inconsistency, > considering that in upper-constraints it's called as > prettytable[2] > > Based on that I propesed patch[3] that aims to add prettytable > package, so that affected projects could replace package in their > requirements.txt in a timely manner. [...] I'm confused. PrettyTable and prettytable are the same thing as far as PyPI/Warehouse is concerned. Python package names are case-insensitive, so both of these normalize to an identical string. In fact, if you try to go to https://pypi.org/project/PrettyTable/ you will simply end up at https://pypi.org/project/prettytable/ anyway, same if you try to `pip install PrettyTable` then `pip freeze` will tell you that you have prettytable installed. Can you elaborate on what problem the mixed-case version of the name is causing? -- Jeremy Stanley -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 963 bytes Desc: not available URL: From noonedeadpunk at ya.ru Wed Dec 22 16:09:54 2021 From: noonedeadpunk at ya.ru (Dmitriy Rabotyagov) Date: Wed, 22 Dec 2021 18:09:54 +0200 Subject: [adjutant][requirements] Adjutant Django version Message-ID: <11731640188844@mail.yandex.ru> An HTML attachment was scrubbed... URL: From miguel at mlavalle.com Wed Dec 22 16:19:42 2021 From: miguel at mlavalle.com (Miguel Lavalle) Date: Wed, 22 Dec 2021 10:19:42 -0600 Subject: Can neutron-fwaas project be revived? In-Reply-To: References: Message-ID: Hi Qin, I think that in principle the community will be delighted if you and your team can reactivate the project and maintain it. Probably the best next step is for you to attend the next Neutron drivers meeting ( https://wiki.openstack.org/wiki/Meetings/NeutronDrivers) so we can discuss the specifics of your proposal. This meeting takes place on Fridays at 1400 UTC over IRC in oftc.net, channel #openstack-neutron. Due to the end of year festivities in much of Europe and America, the next meeting will take place until January 7th. Is that a good next step for you? If yes, I'll add this topic to the meeting's agenda. Best regards On Tue, Dec 21, 2021 at 10:29 AM Dazhong Qin (???)-??????? < qinhaizhong01 at inspur.com> wrote: > Hi? > > The firewall project is a necessary function when the project is > delivered. The lack of firewall function after switching OVN is not > acceptable to customers. We intend to maintain this project and develop the > fwaas driver based on ovn. Whether the neutron-fwaas project can be > reactivate? What should I do ? > > > > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From noonedeadpunk at ya.ru Wed Dec 22 16:41:10 2021 From: noonedeadpunk at ya.ru (Dmitriy Rabotyagov) Date: Wed, 22 Dec 2021 18:41:10 +0200 Subject: [requirements][OpenStackClients][rally][tempest][watcher][nova][glance] Use prettytable in requirements In-Reply-To: <20211222154626.mpmnvc5l4oeb6clx@yuggoth.org> References: <4141640184867@mail.yandex.ru> <20211222154626.mpmnvc5l4oeb6clx@yuggoth.org> Message-ID: <11821640189539@mail.yandex.ru> An HTML attachment was scrubbed... URL: From smooney at redhat.com Wed Dec 22 18:31:28 2021 From: smooney at redhat.com (Sean Mooney) Date: Wed, 22 Dec 2021 18:31:28 +0000 Subject: Networks for different availability zones in Horizon In-Reply-To: <3863DFA9-9573-431F-B91A-51C4421926DD@me.com> References: <040f6fdaeca76a854ee70b238e6e63c553d290eb.camel@redhat.com> <3863DFA9-9573-431F-B91A-51C4421926DD@me.com> Message-ID: <2c1ec0019d05c363bd1065640cbfafb62ae5c5d8.camel@redhat.com> On Tue, 2021-12-21 at 10:58 -0500, Jonathan Mills wrote: > Thank you, Hai Wu, for raising the visibility of this issue. We at NASA Goddard want to echo this sentiment wholeheartedly. It?s a serious pain point! > > Our production OpenStack (Wallaby) cloud spans two AZs (different datacenters) right now, and before we?re done we may have 3 or 4 AZs. We use non-overlapping VLAN-based networks in different datacenters. Moreover, we have different kinds of server hardware in each datacenter (often recycled HPC compute nodes). We also have different Cinder storage backends in each datacenter. > > What we end up having is AZs in Nova, AZs in Neutron, AZs in Cinder, and Image Flavors for Glance that are rather AZ dependent (in order to best fit the node geometry). We instantiate these objects with explicit AZs when they support it. In the case of Neutron networks, we can use scheduler hints. > > The real crux of the problem is the end-user education though. Because OpenStack clients (Horizon, OSC) are perfectly happy to let the end-user try to build impossible combinations. We strongly feel that if the individual services aren?t going to communicate AZ data to each other via RPC, that the clients at least should do some kind of filtering on AZs or scheduler hints about AZs. I'm not entirely sure how you'd solve this problem on CLI easily without RPC communication, but Horizon could (should?) dynamically filter the different menus as you select options: e.g. the first selection screen ?Details? has you choose an AZ. Once the user makes a selection there, the flavors, storage, and network options should adjust accordingly. i think part of hte issue here is also that nova and neutorn do not modele netowrk affinity to avaiablity zone. by definiton todxday networks do not have a AZ affintiy of any kind. even when using vlan network the expection form a nova perspective is every network is aviable on every host. this is a long know design limiatation since the pysnet affinity is not models in placment or in the nova schduler filetrs. the only expction to this is really sriov where we can take pci devices as assocatied with a pshynet. so this i not realy a horizon or osc issue. you are trying to do something that is not supprote by nova or neutron. that does not mean that nova and neutron cant eb enahcned to map phsyenst to placement agregatres and aggreats to hosts and then that could be quired by horizon or osc to narrow the approrpiate networks. for l3 routed network this already partly exists but for normal vlan based networkign it does not. it is really a request for network aware sheduling and a a way to then query the avaiable networks/phsynets in horizon based on az. horizon cannot currently support the filtering as it does not have the data required to fileter on since at a data model level network are not related to AZs at all. > > > Jonathan > > > > On Dec 18, 2021, at 9:44 PM, hai wu wrote: > > > > Not using vxlan, so there's no network span across availability zones. > > It seems there's no built-in way to configure Horizon to do so. The > > problem with this is that sometimes if user picks the wrong network > > and availability zone combo, the VM will end up in the wrong state. > > This is why it would be ideal to only show available networks for the > > availability zone that the user already picked. > > > > I am wondering how to modify the horizon source code to achieve this, > > but I am not familiar with Angular, and the workflow code is pretty > > complex for newbie .. > > > > On Sat, Dec 18, 2021 at 6:31 PM Sean Mooney wrote: > > > > > > On Sat, 2021-12-18 at 13:45 -0600, hai wu wrote: > > > > In Horizon "Launch Instance" workflow, in its 4th step of selecting > > > > networks for the new instance, it would always show all the available > > > > networks from all availability zones, regardless of which > > > > "Availability Zone" got picked in 1st step of "Details". > > > > > > > > I tried to update some DB field for availability zone hint for the > > > > relevant network, and that did not help. > > > > > > > > Any way to ensure in Horizon "Launch Instance" GUI workflow, after a > > > > user picking one availability zone in step 1, it would only show the > > > > related networks in that availability zone as available networks in > > > > step 4? > > > > > > networks do not have any affinity to Avaiablity Zones. > > > there is no mapping between neutron physnets and nova host aggreates which are use to > > > model aviablityitz zones. > > > > > > when using tunneled networks like vxlan it is assuems that all hosts in a cloud acorss all avaiablity zones can access > > > any tenant vxlan network. the same is also ture of vlan or flat network the only excption being that htey have physnet > > > assocaited with them. physnets may not be aviable on all host but there is corralation between phsynets and aviaablity zones. > > > unless you manually algin them. > > > > > > > > > > > From smooney at redhat.com Wed Dec 22 18:53:26 2021 From: smooney at redhat.com (Sean Mooney) Date: Wed, 22 Dec 2021 18:53:26 +0000 Subject: Networks for different availability zones in Horizon In-Reply-To: References: <040f6fdaeca76a854ee70b238e6e63c553d290eb.camel@redhat.com> <3863DFA9-9573-431F-B91A-51C4421926DD@me.com> Message-ID: <743e3ef0e3cb64a5f7f97cbfa1520320929ca595.camel@redhat.com> On Wed, 2021-12-22 at 22:36 +0900, Adam Harwell wrote: > +1 from Yahoo. > We have a number of compute AZs in each of our clouds, and networks that > are specifically crafted for those AZs (mostly due to security zones). Our > users very commonly pick incompatible combinations. We have a user manual > with a table that shows valid combinations in an attempt to educate them, > but this is really something that would be better handled directly in the > tooling somehow. Maybe this kind of setup is more common than people think, > at least for large private cloud operators? it might be but you have to keep in mind that an AZ in openstack is just a metadata tag on a nova host aggreate. it is not a concept that really spans serivce and its not a top level resoruece in the api. yes cinder and neuton can havce sudo AZ by setting some data in there config but resouce exposed by cinder, neutron and other service dont really have a real az affintiy. the cinder backed can be labled as bing intend for use with a given AZ but a volume does not actully have an AZ assocatied with it. it may have a volume type assocaiated with it and that volume type may only be supported by a backedn in a specific AZ but if you made the same volume type avaibale in a backend in a different AZ both would be equally valide to create the voluem. by default nova allows cross az attach https://docs.openstack.org/nova/latest/configuration/config.html#cinder.cross_az_attach it will only be prevented if you explcitly set that to false. that is a nova feature not a cinder one. from a cinder point of view it only cares about volumes volume types and backends. the same is pretty much true of neutron it really does not know much about aviablity zones nor should it really as desigend today. to really be able to model what is being discussed we woudl need to have a neutorn extention to tag a netork with a az so that you could query the networks by az and map them to azs. without that there is not much you can do in horizon and you can get the correct behavior in nova. for network az affintiy to work nova woudl have to be enhacned to take the network az request into accoutn when schdulign an instance if you dont also spcirfy it manually on server create and we woudl also want to ensure we done allow instance to be create with networks form differnt AZ. if there is a grop of peole that want to work on this i would suggest propsoing a pair of nova and neutron specs starting with the neturon spec to add az affintiy. you would have to think carfully about if the affinity shoudl exsit at the network, segments or subents level and how it affect other resouces like subnet pools,qos policies and floating ips. once you have the rudemnts of a desgin for how to model this in the neutron api then an nova spec should created desciribing how this will impact the port resoucxe request and scheduling. with the neutron sepec in place the horizon and OSC comunities could also start working on enbiling the ui and cli to filter network resouces by avaiablity zone while the nova work was done to make the end to end usecase work for network avaiableity zone aware scheduling. a lot of the ground work is alreay avaiable by the way for this in placment but neutron would have to be enhanced to consume it. the same is true of cinder. for cinder case it would need to start modeling storage backedn as shareing resouce providers of disk_gb with the volume type model as a trait. if they plance that shareing resouce provider in nova host aggrate for a give AZ then that would cleanly model the affintiy of that backend to a given az and nova could use the volumes az as an input to schduling when one is not speficed to ensure it only considers hosts in that az. to make this work transparnetly is a lot of work and to date no one has really come forward to do it. > > ?Adam > > PS: I think routed networks may fix part of our configuration woes, but > certainly not all. I wonder if there is a sane way to link networks to AZs > like this? > > On Wed, Dec 22, 2021 at 1:08 Jonathan Mills wrote: > > > Thank you, Hai Wu, for raising the visibility of this issue. We at NASA > > Goddard want to echo this sentiment wholeheartedly. It?s a serious pain > > point! > > > > Our production OpenStack (Wallaby) cloud spans two AZs (different > > datacenters) right now, and before we?re done we may have 3 or 4 AZs. We > > use non-overlapping VLAN-based networks in different datacenters. > > Moreover, we have different kinds of server hardware in each datacenter > > (often recycled HPC compute nodes). We also have different Cinder storage > > backends in each datacenter. > > > > What we end up having is AZs in Nova, AZs in Neutron, AZs in Cinder, and > > Image Flavors for Glance that are rather AZ dependent (in order to best fit > > the node geometry). We instantiate these objects with explicit AZs when > > they support it. In the case of Neutron networks, we can use scheduler > > hints. > > > > The real crux of the problem is the end-user education though. Because > > OpenStack clients (Horizon, OSC) are perfectly happy to let the end-user > > try to build impossible combinations. We strongly feel that if the > > individual services aren?t going to communicate AZ data to each other via > > RPC, that the clients at least should do some kind of filtering on AZs or > > scheduler hints about AZs. I'm not entirely sure how you'd solve this > > problem on CLI easily without RPC communication, but Horizon could > > (should?) dynamically filter the different menus as you select options: > > e.g. the first selection screen ?Details? has you choose an AZ. Once the > > user makes a selection there, the flavors, storage, and network options > > should adjust accordingly. > > > > > > Jonathan > > > > > > > On Dec 18, 2021, at 9:44 PM, hai wu wrote: > > > > > > Not using vxlan, so there's no network span across availability zones. > > > It seems there's no built-in way to configure Horizon to do so. The > > > problem with this is that sometimes if user picks the wrong network > > > and availability zone combo, the VM will end up in the wrong state. > > > This is why it would be ideal to only show available networks for the > > > availability zone that the user already picked. > > > > > > I am wondering how to modify the horizon source code to achieve this, > > > but I am not familiar with Angular, and the workflow code is pretty > > > complex for newbie .. > > > > > > On Sat, Dec 18, 2021 at 6:31 PM Sean Mooney wrote: > > > > > > > > On Sat, 2021-12-18 at 13:45 -0600, hai wu wrote: > > > > > In Horizon "Launch Instance" workflow, in its 4th step of selecting > > > > > networks for the new instance, it would always show all the available > > > > > networks from all availability zones, regardless of which > > > > > "Availability Zone" got picked in 1st step of "Details". > > > > > > > > > > I tried to update some DB field for availability zone hint for the > > > > > relevant network, and that did not help. > > > > > > > > > > Any way to ensure in Horizon "Launch Instance" GUI workflow, after a > > > > > user picking one availability zone in step 1, it would only show the > > > > > related networks in that availability zone as available networks in > > > > > step 4? > > > > > > > > networks do not have any affinity to Avaiablity Zones. > > > > there is no mapping between neutron physnets and nova host aggreates > > which are use to > > > > model aviablityitz zones. > > > > > > > > when using tunneled networks like vxlan it is assuems that all hosts in > > a cloud acorss all avaiablity zones can access > > > > any tenant vxlan network. the same is also ture of vlan or flat network > > the only excption being that htey have physnet > > > > assocaited with them. physnets may not be aviable on all host but there > > is corralation between phsynets and aviaablity zones. > > > > unless you manually algin them. > > > > > > > > > > > > > > > > > > -- > Thanks, > --Adam From ozzzo at yahoo.com Wed Dec 22 19:14:13 2021 From: ozzzo at yahoo.com (Albert Braden) Date: Wed, 22 Dec 2021 19:14:13 +0000 (UTC) Subject: [adjutant][requirements] Adjutant Django version In-Reply-To: <11731640188844@mail.yandex.ru> References: <11731640188844@mail.yandex.ru> Message-ID: <1198016075.1409340.1640200453030@mail.yahoo.com> Nobody is working on Adjutant right now. Adrian was the PTL and he no longer works on Openstack: http://lists.openstack.org/pipermail/openstack-discuss/2021-October/025555.html I've asked my employer for permission to volunteer, and they are thinking about that very slowly. I would love to see someone more qualified take the PTL role. Do you want to volunteer? On Wednesday, December 22, 2021, 11:30:37 AM EST, Dmitriy Rabotyagov wrote: Hi!?Right now, upper-constraints [1] assumes that version of Django to be used for Yoga should be 3.2.*. In the meanwhile Adjutant obviously does not support Django 3, according to the project requirements [2]. This brings inconsistency which leads to the breakage of adjutant installation if respecting u-c because package have conflicting dependencies then.?Based on that, I'm wondering about the way forward. Should be upper-constraints be used for Adjutant installation as for any other project? Or there's an ongoing work to bring in support for Django 3.2 to Adjutant??[1] https://opendev.org/openstack/requirements/src/commit/8b5e97b6563f076206b0bfc7276ea9afd179e6b7/upper-constraints.txt#L452[2] https://opendev.org/openstack/adjutant/src/commit/b305d7285f3ea2e12210e592702c4d66a31d6646/requirements.txt#L3--? Kind Regards,Dmitriy Rabotyagov? -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Wed Dec 22 21:53:25 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Wed, 22 Dec 2021 15:53:25 -0600 Subject: [all][tc] Technical Committee next weekly meeting on Dec 23rd at 1500 UTC In-Reply-To: <17dda2f42bc.dbd5816756818.7976500778946876675@ghanshyammann.com> References: <17dda2f42bc.dbd5816756818.7976500778946876675@ghanshyammann.com> Message-ID: <17de4227cd8.e3af0e4c9862.6357815524716456641@ghanshyammann.com> Hello Everyone, Below is the agenda for Tomorrow's TC IRC meeting schedule at 1500 UTC. https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting == Agenda for tomorrow's TC meeting == * Roll call * Follow up on past action items * Gate health check ** Fixing Zuul config error in OpenStack *** https://etherpad.opendev.org/p/zuul-config-error-openstack * SIG i18n status check ** Xena translation missing *** http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026244.html ** Translation bug *** https://review.opendev.org/c/openstack/contributor-guide/+/821371 * Adjutant need PTLs and maintainers ** http://lists.openstack.org/pipermail/openstack-discuss/2021-October/025555.html * Open Reviews ** https://review.opendev.org/q/projects:openstack/governance+is:open -gmann ---- On Mon, 20 Dec 2021 17:31:10 -0600 Ghanshyam Mann wrote ---- > Hello Everyone, > > Technical Committee's next weekly meeting is scheduled for Dec 23rd at 1500 UTC. > > If you would like to add topics for discussion, please add them to the below wiki page by > Wednesday, Dec 22nd, at 2100 UTC. > > https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting > > -gmann > > From rosmaita.fossdev at gmail.com Thu Dec 23 03:42:23 2021 From: rosmaita.fossdev at gmail.com (Brian Rosmaita) Date: Wed, 22 Dec 2021 22:42:23 -0500 Subject: [cinder] proposing Simon Dodsley for cinder-specs-core Message-ID: <65562de2-0254-e231-d7e1-c9435cfac93b@gmail.com> As everyone is painfully aware, we have review bandwidth issues in the Cinder project. I propose to address this issue with respect to cinder specs, at least, by adding a contributor who is not currently a cinder-core as a member of the cinder-specs-core team. (Right now, the only members of cinder-specs-core are the cinder-core team.) Simon Dodsley (simondodsley on IRC) has been a cinder contributor for a long time, and has been especially active in PTGs and midcycle meetings over the past few cycles. He's done quality reviews, especially of specs, this cycle. As a vendor representative, he brings a useful perspective to reviews that affect drivers, as well as general deployment and usability issues. Additionally, given that the current core team has reviewing responsibilities spread over multiple repositories, having a person focused primarily on specs should help us to get more timely feedback for specs proposers. Finally, don't forget that it is thanks to Simon that the Cinder mascot finally has a name ("Argo"), which in turn has given the Cinder team the really cool name "The Argonauts". In the absence of objections, I'll add Simon to the cinder-specs-core team shortly after 1500 UTC on Friday 31 December. Please communicate any concerns to me before that time. cheers, brian From rdhasman at redhat.com Thu Dec 23 05:34:04 2021 From: rdhasman at redhat.com (Rajat Dhasmana) Date: Thu, 23 Dec 2021 11:04:04 +0530 Subject: [cinder] proposing Simon Dodsley for cinder-specs-core In-Reply-To: <65562de2-0254-e231-d7e1-c9435cfac93b@gmail.com> References: <65562de2-0254-e231-d7e1-c9435cfac93b@gmail.com> Message-ID: Simon would be a great addition to the specs core team. +1 from my side. On Thu, Dec 23, 2021 at 9:16 AM Brian Rosmaita wrote: > As everyone is painfully aware, we have review bandwidth issues in the > Cinder project. I propose to address this issue with respect to cinder > specs, at least, by adding a contributor who is not currently a > cinder-core as a member of the cinder-specs-core team. (Right now, the > only members of cinder-specs-core are the cinder-core team.) > > Simon Dodsley (simondodsley on IRC) has been a cinder contributor for a > long time, and has been especially active in PTGs and midcycle meetings > over the past few cycles. He's done quality reviews, especially of > specs, this cycle. As a vendor representative, he brings a useful > perspective to reviews that affect drivers, as well as general > deployment and usability issues. Additionally, given that the current > core team has reviewing responsibilities spread over multiple > repositories, having a person focused primarily on specs should help us > to get more timely feedback for specs proposers. > > Finally, don't forget that it is thanks to Simon that the Cinder mascot > finally has a name ("Argo"), which in turn has given the Cinder team the > really cool name "The Argonauts". > > In the absence of objections, I'll add Simon to the cinder-specs-core > team shortly after 1500 UTC on Friday 31 December. Please communicate > any concerns to me before that time. > > cheers, > brian > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From bkslash at poczta.onet.pl Thu Dec 23 08:25:18 2021 From: bkslash at poczta.onet.pl (Adam Tomas) Date: Thu, 23 Dec 2021 09:25:18 +0100 Subject: [kolla-ansible][cinder] MPIO - how to set "storage_interface" when using two NICs for storage connection Message-ID: <040F81D5-04EC-4562-BC76-D6EC51098D92@poczta.onet.pl> Hi again, I found some informations about MPIO here (https://docs.oracle.com/cd/E78305_01/E78304/html/setup-cinder.html) : "By default, the Cinder block storage service uses the iSCSI protocol to connect instances to volumes. The iscsid container runs on compute nodes and handles the iSCSI session using an iSCSI initiator name that is generated when you deploy Nova compute services. The Nova compute service supports iSCSI multipath for failover purposes and increased performance. Multipath is enabled by default (configured with the enable_multipathd property). When multipath is enabled, the iSCSI initiator (the Compute node) is able to obtain a list of addresses from the storage node that the initiator can use as multiple paths to the iSCSI LUN." But still - how should I set storage_interface, if I have two NICs for storage connection (let?s say eth4 and eth5)? storage_interface=eth4,eth5 ? Best regards Adam Tomas > Hi, > I?d like to use MPIO in connection to storage array, so I have two physical NIC?s in different VLANS on each node and also 4 NIC?s in the storage array (2 in each storage VLAN) so it should result in 4 MPIO paths from every node (2 paths in each VLAN). But how to set it in kolla-ansible? Can I point to more than one interface using storage_interface variable? Or maybe I should configure it in different way? -------------- next part -------------- An HTML attachment was scrubbed... URL: From radoslaw.piliszek at gmail.com Thu Dec 23 08:36:59 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Thu, 23 Dec 2021 09:36:59 +0100 Subject: [kolla-ansible][cinder] MPIO - how to set "storage_interface" when using two NICs for storage connection In-Reply-To: <040F81D5-04EC-4562-BC76-D6EC51098D92@poczta.onet.pl> References: <040F81D5-04EC-4562-BC76-D6EC51098D92@poczta.onet.pl> Message-ID: Hi Adam, The ``storage_interface`` variable is only used to set the default of ``swift_storage_interface`` and, in turn, ``swift_replication_interface`` by default as well, i.e., it controls where swift will offer its services (and possibly where the replication backbone will be). [1] It is irrelevant to block (volume) storage interfaces. Enabling multipathing (via enable_multipathd) should be sufficient in your case, as long as the targets advertise multiple paths. Please let us know how the cinder guide could be improved. [2] [1] https://docs.openstack.org/kolla-ansible/xena/admin/production-architecture-guide.html [2] https://docs.openstack.org/kolla-ansible/xena/reference/storage/cinder-guide.html Kind regards, -yoctozepto On Thu, 23 Dec 2021 at 09:26, Adam Tomas wrote: > > Hi again, > > I found some informations about MPIO here (https://docs.oracle.com/cd/E78305_01/E78304/html/setup-cinder.html) : > > "By default, the Cinder block storage service uses the iSCSI protocol to connect instances to volumes. The iscsid container runs on compute nodes and handles the iSCSI session using an iSCSI initiator name that is generated when you deploy Nova compute services. > > The Nova compute service supports iSCSI multipath for failover purposes and increased performance. Multipath is enabled by default (configured with the enable_multipathd property). When multipath is enabled, the iSCSI initiator (the Compute node) is able to obtain a list of addresses from the storage node that the initiator can use as multiple paths to the iSCSI LUN." > > But still - how should I set storage_interface, if I have two NICs for storage connection (let?s say eth4 and eth5)? storage_interface=eth4,eth5 ? > > Best regards > Adam Tomas > > Hi, > I?d like to use MPIO in connection to storage array, so I have two physical NIC?s in different VLANS on each node and also 4 NIC?s in the storage array (2 in each storage VLAN) so it should result in 4 MPIO paths from every node (2 paths in each VLAN). But how to set it in kolla-ansible? Can I point to more than one interface using storage_interface variable? Or maybe I should configure it in different way? From bkslash at poczta.onet.pl Thu Dec 23 09:31:00 2021 From: bkslash at poczta.onet.pl (Adam Tomas) Date: Thu, 23 Dec 2021 10:31:00 +0100 Subject: [kolla-ansible][cinder] MPIO - how to set "storage_interface" when using two NICs for storage connection In-Reply-To: References: <040F81D5-04EC-4562-BC76-D6EC51098D92@poczta.onet.pl> Message-ID: <5781C430-9D3A-4143-A5D5-A02717F5E5A1@poczta.onet.pl> Hi Rados?aw, thanks for the answer. So if I?m not going to use swift, then I shouldn?t set storage_interface, right? Ok then - if I use external iSCSI storage (like NetApp, Nexenta, etc) I have to configure proper cinder.volume.driver.iscsi.XXXX and specify storage host. But what if my storage host is reachable on multiple IP addresses? Should I set XXX_host=FIRST_IP XXX_host=SECOND_IP XXX_host=THIRD_IP ? ? I assume, that because each container sees all host NICs then it?s enough to have interfaces on host connected to the same network as external storage, the one specified in cinder.conf? If I use only external iSCSI storage, then only cinder container is responsible for connecting volumes on external storage with instances? Are iscsid & tgtd containers used only if I have external volumes mounted to compute host? By enabling multipathing you mean to set enable_multipathd in globals.yml? Best regards Adam Tomas > Wiadomo?? napisana przez Rados?aw Piliszek w dniu 23.12.2021, o godz. 09:36: > > Hi Adam, > > The ``storage_interface`` variable is only used to set the default of > ``swift_storage_interface`` and, in turn, > ``swift_replication_interface`` by default as well, i.e., it controls > where swift will offer its services (and possibly where the > replication backbone will be). [1] > It is irrelevant to block (volume) storage interfaces. > > Enabling multipathing (via enable_multipathd) should be sufficient in > your case, as long as the targets advertise multiple paths. > > Please let us know how the cinder guide could be improved. [2] > > [1] https://docs.openstack.org/kolla-ansible/xena/admin/production-architecture-guide.html > [2] https://docs.openstack.org/kolla-ansible/xena/reference/storage/cinder-guide.html > > Kind regards, > -yoctozepto > > On Thu, 23 Dec 2021 at 09:26, Adam Tomas wrote: >> >> Hi again, >> >> I found some informations about MPIO here (https://docs.oracle.com/cd/E78305_01/E78304/html/setup-cinder.html) : >> >> "By default, the Cinder block storage service uses the iSCSI protocol to connect instances to volumes. The iscsid container runs on compute nodes and handles the iSCSI session using an iSCSI initiator name that is generated when you deploy Nova compute services. >> >> The Nova compute service supports iSCSI multipath for failover purposes and increased performance. Multipath is enabled by default (configured with the enable_multipathd property). When multipath is enabled, the iSCSI initiator (the Compute node) is able to obtain a list of addresses from the storage node that the initiator can use as multiple paths to the iSCSI LUN." >> >> But still - how should I set storage_interface, if I have two NICs for storage connection (let?s say eth4 and eth5)? storage_interface=eth4,eth5 ? >> >> Best regards >> Adam Tomas >> >> Hi, >> I?d like to use MPIO in connection to storage array, so I have two physical NIC?s in different VLANS on each node and also 4 NIC?s in the storage array (2 in each storage VLAN) so it should result in 4 MPIO paths from every node (2 paths in each VLAN). But how to set it in kolla-ansible? Can I point to more than one interface using storage_interface variable? Or maybe I should configure it in different way? From flux.adam at gmail.com Thu Dec 23 11:16:50 2021 From: flux.adam at gmail.com (Adam Harwell) Date: Thu, 23 Dec 2021 20:16:50 +0900 Subject: Networks for different availability zones in Horizon In-Reply-To: <743e3ef0e3cb64a5f7f97cbfa1520320929ca595.camel@redhat.com> References: <040f6fdaeca76a854ee70b238e6e63c553d290eb.camel@redhat.com> <3863DFA9-9573-431F-B91A-51C4421926DD@me.com> <743e3ef0e3cb64a5f7f97cbfa1520320929ca595.camel@redhat.com> Message-ID: Yeah, that pretty much describes what I was thinking. We?ve done work internally previously to do something similar (or maybe that was at GoDaddy? who I?m pretty sure have similar requirements, though I?d have to sync up with folks there again to double check). It?s definitely something I?d like to bring up at the next major get together, be it Summit or PTG, I?ve lost track. Maybe some of us can compare notes and priorities? It?s definitely not going to be fast, so treating it as urgent would be a bit overzealous probably. ? On Thu, Dec 23, 2021 at 3:53 Sean Mooney wrote: > On Wed, 2021-12-22 at 22:36 +0900, Adam Harwell wrote: > > +1 from Yahoo. > > We have a number of compute AZs in each of our clouds, and networks that > > are specifically crafted for those AZs (mostly due to security zones). > Our > > users very commonly pick incompatible combinations. We have a user manual > > with a table that shows valid combinations in an attempt to educate them, > > but this is really something that would be better handled directly in the > > tooling somehow. Maybe this kind of setup is more common than people > think, > > at least for large private cloud operators? > it might be but you have to keep in mind that an AZ in openstack is just a > metadata tag on a nova host aggreate. > > it is not a concept that really spans serivce and its not a top level > resoruece in the api. > > yes cinder and neuton can havce sudo AZ by setting some data in there > config but resouce exposed by > cinder, neutron and other service dont really have a real az affintiy. > > the cinder backed can be labled as bing intend for use with a given AZ but > a volume does not actully have an AZ assocatied with it. > it may have a volume type assocaiated with it and that volume type may > only be supported by a backedn in a specific AZ > but if you made the same volume type avaibale in a backend in a different > AZ both would be equally valide to create the voluem. > > by default nova allows cross az attach > > https://docs.openstack.org/nova/latest/configuration/config.html#cinder.cross_az_attach > it will only be prevented if you explcitly set that to false. that is a > nova feature not a cinder one. > from a cinder point of view it only cares about volumes volume types and > backends. > > the same is pretty much true of neutron it really does not know much about > aviablity zones nor should it really as desigend today. > > to really be able to model what is being discussed we woudl need to have a > neutorn extention to tag a netork with a az so that you could query > the networks by az and map them to azs. without that there is not much you > can do in horizon and you can get the correct behavior in nova. > for network az affintiy to work nova woudl have to be enhacned to take the > network az request into accoutn when schdulign an instance if you dont > also spcirfy it manually on server create and we woudl also want to ensure > we done allow instance to be create with networks form differnt AZ. > > if there is a grop of peole that want to work on this i would suggest > propsoing a pair of nova and neutron specs starting with the neturon spec > to add az affintiy. you would have to think carfully about if the affinity > shoudl exsit at the network, segments or subents level and how it affect > other > resouces like subnet pools,qos policies and floating ips. once you have > the rudemnts of a desgin for how to model this in the neutron api then > an nova spec should created desciribing how this will impact the port > resoucxe request and scheduling. > > with the neutron sepec in place the horizon and OSC comunities could also > start working on enbiling the ui and cli to filter network resouces by > avaiablity zone > while the nova work was done to make the end to end usecase work for > network avaiableity zone aware scheduling. > > > a lot of the ground work is alreay avaiable by the way for this in > placment but neutron would have to be enhanced to consume it. > the same is true of cinder. for cinder case it would need to start > modeling storage backedn as shareing resouce providers of disk_gb with the > volume type model as a trait. > if they plance that shareing resouce provider in nova host aggrate for a > give AZ then that would cleanly model the affintiy of that backend to a > given az and nova > could use the volumes az as an input to schduling when one is not speficed > to ensure it only considers hosts in that az. > > to make this work transparnetly is a lot of work and to date no one has > really come forward to do it. > > > > ?Adam > > > > PS: I think routed networks may fix part of our configuration woes, but > > certainly not all. I wonder if there is a sane way to link networks to > AZs > > like this? > > > > On Wed, Dec 22, 2021 at 1:08 Jonathan Mills wrote: > > > > > Thank you, Hai Wu, for raising the visibility of this issue. We at > NASA > > > Goddard want to echo this sentiment wholeheartedly. It?s a serious > pain > > > point! > > > > > > Our production OpenStack (Wallaby) cloud spans two AZs (different > > > datacenters) right now, and before we?re done we may have 3 or 4 AZs. > We > > > use non-overlapping VLAN-based networks in different datacenters. > > > Moreover, we have different kinds of server hardware in each datacenter > > > (often recycled HPC compute nodes). We also have different Cinder > storage > > > backends in each datacenter. > > > > > > What we end up having is AZs in Nova, AZs in Neutron, AZs in Cinder, > and > > > Image Flavors for Glance that are rather AZ dependent (in order to > best fit > > > the node geometry). We instantiate these objects with explicit AZs > when > > > they support it. In the case of Neutron networks, we can use scheduler > > > hints. > > > > > > The real crux of the problem is the end-user education though. Because > > > OpenStack clients (Horizon, OSC) are perfectly happy to let the > end-user > > > try to build impossible combinations. We strongly feel that if the > > > individual services aren?t going to communicate AZ data to each other > via > > > RPC, that the clients at least should do some kind of filtering on AZs > or > > > scheduler hints about AZs. I'm not entirely sure how you'd solve this > > > problem on CLI easily without RPC communication, but Horizon could > > > (should?) dynamically filter the different menus as you select options: > > > e.g. the first selection screen ?Details? has you choose an AZ. Once > the > > > user makes a selection there, the flavors, storage, and network options > > > should adjust accordingly. > > > > > > > > > Jonathan > > > > > > > > > > On Dec 18, 2021, at 9:44 PM, hai wu wrote: > > > > > > > > Not using vxlan, so there's no network span across availability > zones. > > > > It seems there's no built-in way to configure Horizon to do so. The > > > > problem with this is that sometimes if user picks the wrong network > > > > and availability zone combo, the VM will end up in the wrong state. > > > > This is why it would be ideal to only show available networks for the > > > > availability zone that the user already picked. > > > > > > > > I am wondering how to modify the horizon source code to achieve this, > > > > but I am not familiar with Angular, and the workflow code is pretty > > > > complex for newbie .. > > > > > > > > On Sat, Dec 18, 2021 at 6:31 PM Sean Mooney > wrote: > > > > > > > > > > On Sat, 2021-12-18 at 13:45 -0600, hai wu wrote: > > > > > > In Horizon "Launch Instance" workflow, in its 4th step of > selecting > > > > > > networks for the new instance, it would always show all the > available > > > > > > networks from all availability zones, regardless of which > > > > > > "Availability Zone" got picked in 1st step of "Details". > > > > > > > > > > > > I tried to update some DB field for availability zone hint for > the > > > > > > relevant network, and that did not help. > > > > > > > > > > > > Any way to ensure in Horizon "Launch Instance" GUI workflow, > after a > > > > > > user picking one availability zone in step 1, it would only show > the > > > > > > related networks in that availability zone as available networks > in > > > > > > step 4? > > > > > > > > > > networks do not have any affinity to Avaiablity Zones. > > > > > there is no mapping between neutron physnets and nova host > aggreates > > > which are use to > > > > > model aviablityitz zones. > > > > > > > > > > when using tunneled networks like vxlan it is assuems that all > hosts in > > > a cloud acorss all avaiablity zones can access > > > > > any tenant vxlan network. the same is also ture of vlan or flat > network > > > the only excption being that htey have physnet > > > > > assocaited with them. physnets may not be aviable on all host but > there > > > is corralation between phsynets and aviaablity zones. > > > > > unless you manually algin them. > > > > > > > > > > > > > > > > > > > > > > > > -- > > Thanks, > > --Adam > > -- Thanks, --Adam -------------- next part -------------- An HTML attachment was scrubbed... URL: From radoslaw.piliszek at gmail.com Thu Dec 23 11:38:54 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Thu, 23 Dec 2021 12:38:54 +0100 Subject: [kolla-ansible][cinder] MPIO - how to set "storage_interface" when using two NICs for storage connection In-Reply-To: <5781C430-9D3A-4143-A5D5-A02717F5E5A1@poczta.onet.pl> References: <040F81D5-04EC-4562-BC76-D6EC51098D92@poczta.onet.pl> <5781C430-9D3A-4143-A5D5-A02717F5E5A1@poczta.onet.pl> Message-ID: On Thu, 23 Dec 2021 at 10:31, Adam Tomas wrote: > > Hi Rados?aw, thanks for the answer. > So if I?m not going to use swift, then I shouldn?t set storage_interface, right? That's right. And we will deprecate and remove this confusing variable. > Ok then - if I use external iSCSI storage (like NetApp, Nexenta, etc) I have to configure proper cinder.volume.driver.iscsi.XXXX and specify storage host. But what if my storage host is reachable on multiple IP addresses? Should I set > > XXX_host=FIRST_IP > XXX_host=SECOND_IP > XXX_host=THIRD_IP > ? > > ? No idea, I have not used it like this. Probably best to ask cinder folks specifically about this config. Kolla does not affect this. > I assume, that because each container sees all host NICs then it?s enough to have interfaces on host connected to the same network as external storage, the one specified in cinder.conf? That sounds sensible. > If I use only external iSCSI storage, then only cinder container is responsible for connecting volumes on external storage with instances? Are iscsid & tgtd containers used only if I have external volumes mounted to compute host? tgtd is only used if you are on Debuntu and use Linux LVM-backed iSCSI target, it's irrelevant for external iSCSI. iscsid, on the other hand, is used to setup the initiator (client), and thus it runs where cinder-volume and nova-compute run, to establish the connection to the target. Same for multipathd. > By enabling multipathing you mean to set enable_multipathd in globals.yml? Yes, that's precisely what I meant. :-) Kind regards, -yoctozepto > Best regards > > Adam Tomas > > > > Wiadomo?? napisana przez Rados?aw Piliszek w dniu 23.12.2021, o godz. 09:36: > > > > Hi Adam, > > > > The ``storage_interface`` variable is only used to set the default of > > ``swift_storage_interface`` and, in turn, > > ``swift_replication_interface`` by default as well, i.e., it controls > > where swift will offer its services (and possibly where the > > replication backbone will be). [1] > > It is irrelevant to block (volume) storage interfaces. > > > > Enabling multipathing (via enable_multipathd) should be sufficient in > > your case, as long as the targets advertise multiple paths. > > > > Please let us know how the cinder guide could be improved. [2] > > > > [1] https://docs.openstack.org/kolla-ansible/xena/admin/production-architecture-guide.html > > [2] https://docs.openstack.org/kolla-ansible/xena/reference/storage/cinder-guide.html > > > > Kind regards, > > -yoctozepto > > > > On Thu, 23 Dec 2021 at 09:26, Adam Tomas wrote: > >> > >> Hi again, > >> > >> I found some informations about MPIO here (https://docs.oracle.com/cd/E78305_01/E78304/html/setup-cinder.html) : > >> > >> "By default, the Cinder block storage service uses the iSCSI protocol to connect instances to volumes. The iscsid container runs on compute nodes and handles the iSCSI session using an iSCSI initiator name that is generated when you deploy Nova compute services. > >> > >> The Nova compute service supports iSCSI multipath for failover purposes and increased performance. Multipath is enabled by default (configured with the enable_multipathd property). When multipath is enabled, the iSCSI initiator (the Compute node) is able to obtain a list of addresses from the storage node that the initiator can use as multiple paths to the iSCSI LUN." > >> > >> But still - how should I set storage_interface, if I have two NICs for storage connection (let?s say eth4 and eth5)? storage_interface=eth4,eth5 ? > >> > >> Best regards > >> Adam Tomas > >> > >> Hi, > >> I?d like to use MPIO in connection to storage array, so I have two physical NIC?s in different VLANS on each node and also 4 NIC?s in the storage array (2 in each storage VLAN) so it should result in 4 MPIO paths from every node (2 paths in each VLAN). But how to set it in kolla-ansible? Can I point to more than one interface using storage_interface variable? Or maybe I should configure it in different way? > From bkslash at poczta.onet.pl Thu Dec 23 11:50:59 2021 From: bkslash at poczta.onet.pl (Adam Tomas) Date: Thu, 23 Dec 2021 12:50:59 +0100 Subject: [cinder] MPIO - how to set "storage_interface" when using two NICs for storage connection In-Reply-To: References: <040F81D5-04EC-4562-BC76-D6EC51098D92@poczta.onet.pl> <5781C430-9D3A-4143-A5D5-A02717F5E5A1@poczta.onet.pl> Message-ID: Thanks a lot :) So - anyone from cinder team - how to use cinder with external MPIO iSCSI storage? If I use external iSCSI storage (like NetApp, Nexenta, etc) I have to configure proper cinder.volume.driver.iscsi.XXXX and specify storage host. But what if my storage host is reachable on multiple IP addresses? Should I set XXX_host=FIRST_IP XXX_host=SECOND_IP XXX_host=THIRD_IP ? ? or one XXX_host=IP is enough to get additional paths from iSCSI target? Best regards Adam Tomas From qinhaizhong01 at inspur.com Thu Dec 23 01:50:46 2021 From: qinhaizhong01 at inspur.com (=?utf-8?B?RGF6aG9uZyBRaW4gKOenpua1t+S4rSkt5LqR5pWw5o2u5Lit5b+D6ZuG5Zui?=) Date: Thu, 23 Dec 2021 01:50:46 +0000 Subject: =?utf-8?B?562U5aSNOiBDYW4gbmV1dHJvbi1md2FhcyBwcm9qZWN0IGJlIHJldml2ZWQ/?= In-Reply-To: References: Message-ID: <6c9355fa8f45455e9c5e5c550b94113d@inspur.com> ???: Miguel Lavalle [mailto:miguel at mlavalle.com] ????: 2021?12?23? 0:20 ???: Dazhong Qin (???)-??????? ??: openstack-discuss at lists.openstack.org ??: Re: Can neutron-fwaas project be revived? Hi Qin, I think that in principle the community will be delighted if you and your team can reactivate the project and maintain it. Probably the best next step is for you to attend the next Neutron drivers meeting (https://wiki.openstack.org/wiki/Meetings/NeutronDrivers) so we can discuss the specifics of your proposal. This meeting takes place on Fridays at 1400 UTC over IRC in oftc.net , channel #openstack-neutron. Due to the end of year festivities in much of Europe and America, the next meeting will take place until January 7th. Is that a good next step for you? If yes, I'll add this topic to the meeting's agenda. Best regards On Tue, Dec 21, 2021 at 10:29 AM Dazhong Qin (???)-??????? > wrote: Hi? The firewall project is a necessary function when the project is delivered. The lack of firewall function after switching OVN is not acceptable to customers. We intend to maintain this project and develop the fwaas driver based on ovn. Whether the neutron-fwaas project can be reactivate? What should I do ? -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3615 bytes Desc: not available URL: From qinhaizhong01 at inspur.com Thu Dec 23 02:09:32 2021 From: qinhaizhong01 at inspur.com (=?utf-8?B?RGF6aG9uZyBRaW4gKOenpua1t+S4rSkt5LqR5pWw5o2u5Lit5b+D6ZuG5Zui?=) Date: Thu, 23 Dec 2021 02:09:32 +0000 Subject: =?utf-8?B?562U5aSNOiBDYW4gbmV1dHJvbi1md2FhcyBwcm9qZWN0IGJlIHJldml2ZWQ/?= In-Reply-To: References: Message-ID: <87f786a0f0714c0a89fb7097e965b1d6@inspur.com> Hi Miguel, I am glad to hear this news. How about our discussion on January 7th, this Friday is not convenient, what do I need to prepare before the discussion, do I need to submit rfe or other descriptions? ???: Miguel Lavalle [mailto:miguel at mlavalle.com] ????: 2021?12?23? 0:20 ???: Dazhong Qin (???)-??????? ??: openstack-discuss at lists.openstack.org ??: Re: Can neutron-fwaas project be revived? Hi Qin, I think that in principle the community will be delighted if you and your team can reactivate the project and maintain it. Probably the best next step is for you to attend the next Neutron drivers meeting (https://wiki.openstack.org/wiki/Meetings/NeutronDrivers) so we can discuss the specifics of your proposal. This meeting takes place on Fridays at 1400 UTC over IRC in oftc.net , channel #openstack-neutron. Due to the end of year festivities in much of Europe and America, the next meeting will take place until January 7th. Is that a good next step for you? If yes, I'll add this topic to the meeting's agenda. Best regards On Tue, Dec 21, 2021 at 10:29 AM Dazhong Qin (???)-??????? > wrote: Hi? The firewall project is a necessary function when the project is delivered. The lack of firewall function after switching OVN is not acceptable to customers. We intend to maintain this project and develop the fwaas driver based on ovn. Whether the neutron-fwaas project can be reactivate? What should I do ? -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3615 bytes Desc: not available URL: From ianyrchoi at gmail.com Thu Dec 23 15:35:40 2021 From: ianyrchoi at gmail.com (Ian Y. Choi) Date: Fri, 24 Dec 2021 00:35:40 +0900 Subject: [tc][sig][i18n] What's the status for Xena translation? In-Reply-To: References: Message-ID: Hello, While I18n SIG is really looking for volunteers, I can move forward. However, I am not sure of the balance between the importance of creating the Xena version on Zanata vs. so much manual stuff to create such versions in Zanata from admin(s). Appreciate previous PTL + AJaeger's such valuable work previously. Note that I am still on paternity leave in December but will be back and echo to tc & sig - thank you Akihiro. With many thanks, /Ian p.s., ping "ianychoi" on IRC.. this should be faster than e-mails to me, while I admit that I didn't share my leave days with OpenStack community. On Thu, Dec 9, 2021 at 5:43 PM Andreas Jaeger wrote: > [I'm not sure whether this mail goes through since I'm not subscribed > to the mailing lists] > > If anybody needs some guideance on what needs to be done, I'm happy to > bring anybody up to speed and guide through the process. Ian has done > this in the past as well, so he can do this himself or help others as > well. > > Andreas > > On Thu, Dec 9, 2021 at 3:14 AM Akihiro Motoki wrote: > > > > Hi, > > > > # I include TC and SIG tags to get broader attractions to this topic. > > # Also Cc'ed to Ian (i18n SIG char) and Andreas. > > > > I would like to ask the status of Xena translation in > > translate.openstack.org (Zanata). > > In past releases, stable-XXX branches were created on Zanata, > > but a branch for Xena translation has not been created yet. > > > > Ajaeger usually prepared these branches as a volunteer (I am really > > appreciated) > > but AFAIK he no longer works on OpenStack actively. > > > > Who takes care of the translatio preparation around the release now? > > I hope the i18n SIG is not dead. > > > > Thanks, > > Akihiro Motoki (amotoki) > > > > -- > Andreas Jaeger jaegerandi at gmail.com > -------------- next part -------------- An HTML attachment was scrubbed... URL: From miguel at mlavalle.com Thu Dec 23 16:43:09 2021 From: miguel at mlavalle.com (Miguel Lavalle) Date: Thu, 23 Dec 2021 10:43:09 -0600 Subject: Can neutron-fwaas project be revived? In-Reply-To: <87f786a0f0714c0a89fb7097e965b1d6@inspur.com> References: <87f786a0f0714c0a89fb7097e965b1d6@inspur.com> Message-ID: Hi Qin, In preparation for your meeting with the drivers team, I suggest we follow as a starting point the Neutron Stadium Governance rules and processes as outlined in the official documentation: https://docs.openstack.org/neutron/latest/contributor/stadium/governance.html. In the past, we have re-incorporated projects to the Stadium, like for example in the case of neutron-vpnaas. This document in the Neutron specs repo summarizes how we assessed the readiness of vpnaas for the stadium: https://specs.openstack.org/openstack/neutron-specs/specs/stadium/queens/neutron-vpnaas.html (https://review.opendev.org/c/openstack/neutron-specs/+/506012). I suggest you start a similar document for fwaas in the folder for the current cycle: https://specs.openstack.org/openstack/neutron-specs/specs/yoga/index.html. As soon as you can, please push it to gerrit, so we can start reviewing it. Did I understand correctly that you will attend the drivers meeting on January 7th? Best regards Miguel On Wed, Dec 22, 2021 at 8:09 PM Dazhong Qin (???)-??????? < qinhaizhong01 at inspur.com> wrote: > Hi Miguel, > > I am glad to hear this news. How about our discussion on January 7th, this > Friday is not convenient, what do I need to prepare before the discussion, > do I need to submit rfe or other descriptions? > > > > *???:* Miguel Lavalle [mailto:miguel at mlavalle.com] > *????:* 2021?12?23? 0:20 > *???:* Dazhong Qin (???)-??????? > *??:* openstack-discuss at lists.openstack.org > *??:* Re: Can neutron-fwaas project be revived? > > > > Hi Qin, > > > > I think that in principle the community will be delighted if you and your > team can reactivate the project and maintain it. Probably the best next > step is for you to attend the next Neutron drivers meeting ( > https://wiki.openstack.org/wiki/Meetings/NeutronDrivers) so we > can discuss the specifics of your proposal. This meeting takes place on > Fridays at 1400 UTC over IRC in oftc.net, channel #openstack-neutron. Due > to the end of year festivities in much of Europe and America, the next > meeting will take place until January 7th. Is that a good next step for > you? If yes, I'll add this topic to the meeting's agenda. > > > > Best regards > > > > On Tue, Dec 21, 2021 at 10:29 AM Dazhong Qin (???)-??????? < > qinhaizhong01 at inspur.com> wrote: > > Hi? > > The firewall project is a necessary function when the project is > delivered. The lack of firewall function after switching OVN is not > acceptable to customers. We intend to maintain this project and develop the > fwaas driver based on ovn. Whether the neutron-fwaas project can be > reactivate? What should I do ? > > > > > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From jungleboyj at gmail.com Thu Dec 23 18:51:24 2021 From: jungleboyj at gmail.com (Jay S. Bryant) Date: Thu, 23 Dec 2021 12:51:24 -0600 Subject: [cinder] proposing Simon Dodsley for cinder-specs-core In-Reply-To: <65562de2-0254-e231-d7e1-c9435cfac93b@gmail.com> References: <65562de2-0254-e231-d7e1-c9435cfac93b@gmail.com> Message-ID: <8dad48bf-6261-f15c-f945-dbebb9226653@gmail.com> +2 from me! Thank you for all your help recently Simon! Jay On 12/22/2021 9:42 PM, Brian Rosmaita wrote: > As everyone is painfully aware, we have review bandwidth issues in the > Cinder project.? I propose to address this issue with respect to > cinder specs, at least, by adding a contributor who is not currently a > cinder-core as a member of the cinder-specs-core team.? (Right now, > the only members of cinder-specs-core are the cinder-core team.) > > Simon Dodsley (simondodsley on IRC) has been a cinder contributor for > a long time, and has been especially active in PTGs and midcycle > meetings over the past few cycles.? He's done quality reviews, > especially of specs, this cycle.? As a vendor representative, he > brings a useful perspective to reviews that affect drivers, as well as > general deployment and usability issues.? Additionally, given that the > current core team has reviewing responsibilities spread over multiple > repositories, having a person focused primarily on specs should help > us to get more timely feedback for specs proposers. > > Finally, don't forget that it is thanks to Simon that the Cinder > mascot finally has a name ("Argo"), which in turn has given the Cinder > team the really cool name "The Argonauts". > > In the absence of objections, I'll add Simon to the cinder-specs-core > team shortly after 1500 UTC on Friday 31 December.? Please communicate > any concerns to me before that time. > > cheers, > brian > From gmann at ghanshyammann.com Fri Dec 24 00:10:36 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Thu, 23 Dec 2021 18:10:36 -0600 Subject: [gate][tempest][plugins][stable] Stable/train tempest plugins jobs are broken In-Reply-To: <17dd8e834d9.c0af5eee46883.3204950283589713289@ghanshyammann.com> References: <17dd8e834d9.c0af5eee46883.3204950283589713289@ghanshyammann.com> Message-ID: <17de9c67319.fda7e42066549.877845019887687317@ghanshyammann.com> ---- On Mon, 20 Dec 2021 11:33:56 -0600 Ghanshyam Mann wrote ---- > Hello Everyone, > > gthiemonge reported the issue for tempest plugin stable/train jobs are broken. > > I have explained the issue in https://bugs.launchpad.net/tempest/+bug/1955418 and > fix is up (it is under testing). This is applicable for any tempest plugin job on stable/train. > > Please hold the recheck until the below fix is merged: > > https://review.opendev.org/c/openstack/tempest/+/822339 I need more time to figure out which version of Tempest will work fine on stable/train. 28.0.0, 27.0.0 did not work due to oslo.utils constraints in stable/train and 26.1.0 not working due to ansible role run-tempest which I am not sure why and continue debugging it. Meanwhile, to unblock the gate, we have reverted the Tempest 28.0.0 pin on stable/train[1]. With that stable/train (stable/ussuri grenade ) jobs are passing and you can recheck you patch if failing for this issue. [1] https://review.opendev.org/c/openstack/devstack/+/822678 -gmann > > -gmann > > From abishop at redhat.com Fri Dec 24 00:42:10 2021 From: abishop at redhat.com (Alan Bishop) Date: Thu, 23 Dec 2021 16:42:10 -0800 Subject: [cinder] proposing Simon Dodsley for cinder-specs-core In-Reply-To: <8dad48bf-6261-f15c-f945-dbebb9226653@gmail.com> References: <65562de2-0254-e231-d7e1-c9435cfac93b@gmail.com> <8dad48bf-6261-f15c-f945-dbebb9226653@gmail.com> Message-ID: I'm not a core, but definitely +1 from me fwiw! On Thu, Dec 23, 2021 at 10:54 AM Jay S. Bryant wrote: > +2 from me! > > Thank you for all your help recently Simon! > > Jay > > On 12/22/2021 9:42 PM, Brian Rosmaita wrote: > > As everyone is painfully aware, we have review bandwidth issues in the > > Cinder project. I propose to address this issue with respect to > > cinder specs, at least, by adding a contributor who is not currently a > > cinder-core as a member of the cinder-specs-core team. (Right now, > > the only members of cinder-specs-core are the cinder-core team.) > > > > Simon Dodsley (simondodsley on IRC) has been a cinder contributor for > > a long time, and has been especially active in PTGs and midcycle > > meetings over the past few cycles. He's done quality reviews, > > especially of specs, this cycle. As a vendor representative, he > > brings a useful perspective to reviews that affect drivers, as well as > > general deployment and usability issues. Additionally, given that the > > current core team has reviewing responsibilities spread over multiple > > repositories, having a person focused primarily on specs should help > > us to get more timely feedback for specs proposers. > > > > Finally, don't forget that it is thanks to Simon that the Cinder > > mascot finally has a name ("Argo"), which in turn has given the Cinder > > team the really cool name "The Argonauts". > > > > In the absence of objections, I'll add Simon to the cinder-specs-core > > team shortly after 1500 UTC on Friday 31 December. Please communicate > > any concerns to me before that time. > > > > cheers, > > brian > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From miguel at mlavalle.com Fri Dec 24 16:18:25 2021 From: miguel at mlavalle.com (Miguel Lavalle) Date: Fri, 24 Dec 2021 10:18:25 -0600 Subject: Can neutron-fwaas project be revived? In-Reply-To: <035f05bc97204972bb975c962d47507c@inspur.com> References: <87f786a0f0714c0a89fb7097e965b1d6@inspur.com> <035f05bc97204972bb975c962d47507c@inspur.com> Message-ID: Hi Qin, I have added this topic to the drivers meeting agenda (see on demand agenda close to the bottom): https://wiki.openstack.org/wiki/Meetings/NeutronDrivers Cheers On Thu, Dec 23, 2021 at 7:42 PM Dazhong Qin (???)-??????? < qinhaizhong01 at inspur.com> wrote: > Hi Miguel, > > Thank you for your suggestion. My colleague HengZhou will submit relevant > documents as soon as possible in accordance with the official neutron rules. > > Yes?we will attend the neutron drivers meeting on January 7th. > > Merry Christmas! > > Best wish for you! > > > > *???:* Miguel Lavalle [mailto:miguel at mlavalle.com] > *????:* 2021?12?24? 0:43 > *???:* Dazhong Qin (???)-??????? > *??:* openstack-discuss at lists.openstack.org > *??:* Re: Can neutron-fwaas project be revived? > > > > Hi Qin, > > > > In preparation for your meeting with the drivers team, I suggest we follow > as a starting point the Neutron Stadium Governance rules and processes as > outlined in the official documentation: > https://docs.openstack.org/neutron/latest/contributor/stadium/governance.html. > In the past, we have re-incorporated projects to the Stadium, like for > example in the case of neutron-vpnaas. This document in the Neutron specs > repo summarizes how we assessed the readiness of vpnaas for the stadium: > https://specs.openstack.org/openstack/neutron-specs/specs/stadium/queens/neutron-vpnaas.html > (https://review.opendev.org/c/openstack/neutron-specs/+/506012). I > suggest you start a similar document for fwaas in the folder for the > current cycle: > https://specs.openstack.org/openstack/neutron-specs/specs/yoga/index.html. > As soon as you can, please push it to gerrit, so we can start reviewing it. > > > > Did I understand correctly that you will attend the drivers meeting on > January 7th? > > > > Best regards > > > > Miguel > > > > > > On Wed, Dec 22, 2021 at 8:09 PM Dazhong Qin (???)-??????? < > qinhaizhong01 at inspur.com> wrote: > > Hi Miguel, > > I am glad to hear this news. How about our discussion on January 7th, this > Friday is not convenient, what do I need to prepare before the discussion, > do I need to submit rfe or other descriptions? > > > > *???:* Miguel Lavalle [mailto:miguel at mlavalle.com] > *????:* 2021?12?23? 0:20 > *???:* Dazhong Qin (???)-??????? > *??:* openstack-discuss at lists.openstack.org > *??:* Re: Can neutron-fwaas project be revived? > > > > Hi Qin, > > > > I think that in principle the community will be delighted if you and your > team can reactivate the project and maintain it. Probably the best next > step is for you to attend the next Neutron drivers meeting ( > https://wiki.openstack.org/wiki/Meetings/NeutronDrivers) so we > can discuss the specifics of your proposal. This meeting takes place on > Fridays at 1400 UTC over IRC in oftc.net, channel #openstack-neutron. Due > to the end of year festivities in much of Europe and America, the next > meeting will take place until January 7th. Is that a good next step for > you? If yes, I'll add this topic to the meeting's agenda. > > > > Best regards > > > > On Tue, Dec 21, 2021 at 10:29 AM Dazhong Qin (???)-??????? < > qinhaizhong01 at inspur.com> wrote: > > Hi? > > The firewall project is a necessary function when the project is > delivered. The lack of firewall function after switching OVN is not > acceptable to customers. We intend to maintain this project and develop the > fwaas driver based on ovn. Whether the neutron-fwaas project can be > reactivate? What should I do ? > > > > > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From qinhaizhong01 at inspur.com Fri Dec 24 01:41:19 2021 From: qinhaizhong01 at inspur.com (=?gb2312?B?RGF6aG9uZyBRaW4gKMfYuqPW0Ckt1MbK/b7d1tDQxLyvzcU=?=) Date: Fri, 24 Dec 2021 01:41:19 +0000 Subject: =?gb2312?B?tPC4tDogQ2FuIG5ldXRyb24tZndhYXMgcHJvamVjdCBiZSByZXZpdmVkPw==?= In-Reply-To: References: <87f786a0f0714c0a89fb7097e965b1d6@inspur.com> Message-ID: <035f05bc97204972bb975c962d47507c@inspur.com> A non-text attachment was scrubbed... Name: smime.p7m Type: application/pkcs7-mime Size: 22416 bytes Desc: not available URL: From amonster369 at gmail.com Fri Dec 24 17:24:41 2021 From: amonster369 at gmail.com (A Monster) Date: Fri, 24 Dec 2021 18:24:41 +0100 Subject: No network access for Openstack instances Message-ID: I have used kolla ansible to deploy openstack on two machine servers, but after setting up the images and the virtual networks, the vm instances don't get the ip address that neutron provides for them, so I had to manually set the addresses inside the vm's, but even with that I can't seem to connect vm's to neither internet nor any other network. I'm on centos 7 and trying to deploy openstack train using kolla ansible 9.3.1 Ansible 2.9.25 And neutron open vSwitch. -------------- next part -------------- An HTML attachment was scrubbed... URL: From raubvogel at gmail.com Fri Dec 24 19:12:35 2021 From: raubvogel at gmail.com (Mauricio Tavares) Date: Fri, 24 Dec 2021 14:12:35 -0500 Subject: No network access for Openstack instances In-Reply-To: References: Message-ID: On Fri, Dec 24, 2021 at 12:29 PM A Monster wrote: > > I have used kolla ansible to deploy openstack on two machine servers, but after setting up the images and the virtual networks, the vm instances don't get the ip address that neutron provides for them, so I had to manually set the addresses inside the vm's, but even with that I can't seem to connect vm's to neither internet nor any other network. > I'm on centos 7 and trying to deploy openstack train using > kolla ansible 9.3.1 > Ansible 2.9.25 > And neutron open vSwitch. I am away from my computer, but this has happened to me before. As you know, neutron must use vswitch to create the gateway to connect the private vlan to the outside per your instructions. Sometimes it does not. A few times it was just that the vms would not get the DHCP-provided IP, but it seems you are past that and manually entered the internal IP and route. If that is the case, the next thing is to go to vswitch and see if you can follow the network to see what is missing. From gmann at ghanshyammann.com Fri Dec 24 19:59:33 2021 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Fri, 24 Dec 2021 13:59:33 -0600 Subject: [all][tc] What's happening in Technical Committee: summary 24th Dec, 21: Reading: 5 min Message-ID: <17dee06f4be.b611357894598.3807980798649420697@ghanshyammann.com> Hello Everyone, Here is this week's summary of the Technical Committee activities. 1. TC Meetings: ============ * We had a quick meeting yesterday. Most of the meeting discussions are summarized below (Completed or in-progress activities section). Meeting full logs are available @https://meetings.opendev.org/meetings/tc/2021/tc.2021-12-23-15.00.log.html * Next week's meeting on 30th Dec is canceled. We will continue our weekly meeting from Jan 6th, Thursday 15:00 UTC, feel free the topic on the agenda[1] by 5th Jan. 2. What we completed this week: ========================= * Nothing much in this week. 3. Activities In progress: ================== TC Tracker for Yoga cycle ------------------------------ * This etherpad includes the Yoga cycle targets/working items for TC[2]. Open Reviews ----------------- * 4 open reviews for ongoing activities[3]. SIG i18n status check -------------------------- Ian (SIG chair) responded on ML thread and will look into the things after he is back from leave next month[4]. OpenStack Pain points discussion ---------------------------------------- No updates in this week. Stay tuned on the ML thread[5] and Rico will inform you about the next meeting details. TC position on release cadence ------------------------------------- No updates on this. Discussion is in-progress in ML thread[6]. No other updates from TC on this and we will set up a separate call to continue the discussion sometime in Jan (after the Christmas holidays). Fixing Zuul config error ---------------------------- Requesting projects with zuul config error to look into those and fix them which should not take much time[7]. Skyline as official projects ------------------------------ We accepted the Skyline as an official project last week and next work on project renaming etc are also started. Adjutant need maintainers and PTLs ------------------------------------------- We are still waiting to hear from Braden, Albert on permission to work on this project[8]. Project updates ------------------- * Retire js-openstack-lib (waiting on Adjutant to have new PTL/maintainer) [9] 4. How to contact the TC: ==================== If you would like to discuss or give feedback to TC, you can reach out to us in multiple ways: 1. Email: you can send the email with tag [tc] on openstack-discuss ML[10]. 2. Weekly meeting: The Technical Committee conduct a weekly meeting every Thursday 15 UTC [11] 3. Ping us using 'tc-members' nickname on #openstack-tc IRC channel. Marry Chirsmas and Happy new year to everyone and stay safe! [1] https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting [2] https://etherpad.opendev.org/p/tc-yoga-tracker [3] https://review.opendev.org/q/projects:openstack/governance+status:open [4] http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026441.html [5] http://lists.openstack.org/pipermail/openstack-discuss/2021-December/026245.html [6] http://lists.openstack.org/pipermail/openstack-discuss/2021-November/025684.html [7] https://etherpad.opendev.org/p/zuul-config-error-openstack [8] http://lists.openstack.org/pipermail/openstack-discuss/2021-November/025786.html [9] https://review.opendev.org/c/openstack/governance/+/798540 [10] http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-discuss [11] http://eavesdrop.openstack.org/#Technical_Committee_Meeting -gmann From romain.chanu at univ-lyon1.fr Sun Dec 26 09:55:47 2021 From: romain.chanu at univ-lyon1.fr (CHANU ROMAIN) Date: Sun, 26 Dec 2021 09:55:47 +0000 Subject: No network access for Openstack instances In-Reply-To: References: Message-ID: Hello, If you a cloud image, cloud-init set the IP address from neutron-dhcp- server. You should check neutron dhcp server and metadata agent log and verify connectivity between your instance and DHCP server and try to contact metadata server: wget 169.254.169.254 Best regards, Romain On Fri, 2021-12-24 at 14:12 -0500, Mauricio Tavares wrote: > On Fri, Dec 24, 2021 at 12:29 PM A Monster > wrote: > > > > I have used kolla ansible to deploy openstack on two machine > > servers, but after setting up the images and the virtual networks, > > the vm instances don't get the ip address that neutron provides for > > them, so I had to manually set the addresses inside the vm's, but > > even with that I can't seem to connect vm's to neither internet nor > > any other network. > > I'm on centos 7 and trying to deploy openstack train using > > kolla ansible 9.3.1 > > Ansible 2.9.25 > > And neutron open vSwitch. > > I am away from my computer, but this has happened to me before. As > you > know, neutron must use vswitch to create the gateway to connect the > private vlan to the outside per your instructions. Sometimes it does > not. A few times it was just that the vms would not get the > DHCP-provided IP, but it seems you are past that and manually entered > the internal IP and route. If that is the case, the next thing is to > go to vswitch and see if you can follow the network to see what is > missing. > -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 3217 bytes Desc: not available URL: From yasufum.o at gmail.com Mon Dec 27 02:53:43 2021 From: yasufum.o at gmail.com (Yasufumi Ogawa) Date: Mon, 27 Dec 2021 11:53:43 +0900 Subject: [tacker] No weekly meeting until 2022 Message-ID: <72789e04-c7e3-cf3c-dec7-992a2183193c@gmail.com> Hi, As agreed in the previous meeting, we will skip IRC meeting this week and be on 11th Jan next year. Happy holidays! Cheers, yasufumi From mnasiadka at gmail.com Mon Dec 27 06:55:26 2021 From: mnasiadka at gmail.com (=?UTF-8?Q?Micha=C5=82_Nasiadka?=) Date: Mon, 27 Dec 2021 07:55:26 +0100 Subject: [kolla] No IRC meeting this week Message-ID: Hello Koalas, As agreed in the previous meeting, we will skip IRC meeting this week (planned for 29th Dec) and meet on 5th Jan. Happy new year! Michal -- Micha? Nasiadka mnasiadka at gmail.com -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Mon Dec 27 10:44:56 2021 From: zigo at debian.org (Thomas Goirand) Date: Mon, 27 Dec 2021 11:44:56 +0100 Subject: [nova] Getting away from cdrkit / genisoimage Message-ID: <6494f3ac-cfe3-d028-33d6-05ab9b625d67@debian.org> Hi, Please see: https://bugs.debian.org/982241 It looks like cdrkit / genisoimage wont be available in the next Debian 12 (and therefore probably it will also be dropped in Ubuntu). As per the bug report: "I'm told xorriso and libburnia are alternatives and are alive and doing well." Would it be possible to switch to that? Cheers, Thomas Goirand (zigo) From rafaelweingartner at gmail.com Mon Dec 27 13:35:19 2021 From: rafaelweingartner at gmail.com (=?UTF-8?Q?Rafael_Weing=C3=A4rtner?=) Date: Mon, 27 Dec 2021 10:35:19 -0300 Subject: [cloudkitty] No weekly meeting today Message-ID: Hello guys, As agreed in the previous meeting, we will skip IRC meeting this week and be on 10th Jan next year. I hope you all had a merry Christmans, and happy holidays! -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From bence.romsics at gmail.com Mon Dec 27 14:38:13 2021 From: bence.romsics at gmail.com (Bence Romsics) Date: Mon, 27 Dec 2021 15:38:13 +0100 Subject: [neutron] bug deputy report of week 2021-12-20 Message-ID: Hi Team, We only had a few new bugs last week, here they are: High: * https://bugs.launchpad.net/neutron/+bug/1955478 Scenario test failed due to metadata service not available slaweq investigated this gate failure, but did not find enough information in the logs. * https://bugs.launchpad.net/neutron/+bug/1955546 [stable only] Logging service plugin expects wrong arguments in the callback function fixes mostly merged: https://review.opendev.org/q/I0003bd566dc769436ad1342351ad058394bd52da * https://bugs.launchpad.net/neutron/+bug/1955578 OVN transaction could not be completed due to a race condition proposed fix: https://review.opendev.org/c/openstack/neutron/+/821401 * https://bugs.launchpad.net/neutron/+bug/1955775 Error when l3-agent get filter id for ip proposed fix: https://review.opendev.org/c/openstack/neutron/+/822987 Medium: * https://bugs.launchpad.net/neutron/+bug/1955639 Performance of mariadb's neutron.agents table unassigned Incomplete: * https://bugs.launchpad.net/neutron/+bug/1955503 [ovn] OVN agents showing as dead until neutron services restarted If ovn folks have an intuition what may go wrong with agent caching, then please look at it, but we don't have reproduction steps, so marked as Incomplete. Other: * https://bugs.launchpad.net/neutron/+bug/1955765 Devstack - Can no longer enable qos with neutron-qos Now that we un-deprecated the legacy plugin what do we want to do with the new plugin? Possible discussion point for the next meeting. Cheers, Bence From anyrude10 at gmail.com Mon Dec 27 13:01:36 2021 From: anyrude10 at gmail.com (Anirudh Gupta) Date: Mon, 27 Dec 2021 18:31:36 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines Message-ID: Hi Team, I am trying to deploy TripleO Train with 3 controller and 1 Compute. For overcloud images, I have a registry server at undercloud only. I executed the following command to deploy overcloud openstack overcloud deploy --templates \ -r /home/stack/templates/roles_data.yaml \ -e /home/stack/templates/node-info.yaml \ -e /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ -e /home/stack/containers-prepare-parameter.yaml The command ran for around 1.5 hrs and initially stack got successfully created and after that for 45 mins, ansible tasks were getting executed. It then gave following error in overcloud-controller-0 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug output for task: Start containers for step 2 | overcloud-novacompute-0 | result={ "changed": false, "failed_when_result": false, "start_containers_outputs.stdout_lines | default([]) | union(start_containers_outputs.stderr_lines | default([]))": [ "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", "", "Info: Loading facts", "Notice: Compiled catalog for overcloud-novacompute-0.localdomain in environment production in 0.05 seconds", "Info: Applying configuration version '1640604309'", "Notice: /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", "Info: Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: Unscheduling all events on Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", "Info: Creating state file /var/lib/puppet/state/state.yaml", "Notice: Applied catalog in 0.01 seconds", "Changes:", " Total: 1", "Events:", " Success: 1", "Resources:", " Changed: 1", " Out of sync: 1", " Skipped: 7", " Total: 8", "Time:", " File: 0.00", " Transaction evaluation: 0.01", " Catalog application: 0.01", " Config retrieval: 0.09", " Last run: 1640604309", " Total: 0.01", "Version:", " Config: 1640604309", " Puppet: 5.5.10", * "Error executing ['podman', 'container', 'exists', 'nova_compute_init_log']: returned 1", "Did not find container with \"['podman', 'ps', '-a', '--filter', 'label=container_name=nova_compute_init_log', '--filter', 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying without config_id", "Did not find container with \"['podman', 'ps', '-a', '--filter', 'label=container_name=nova_compute_init_log', '--format', '{{.Names}}']\"", "Error executing ['podman', 'container', 'exists', 'create_haproxy_wrapper']: returned 1", "Did not find container with \"['podman', 'ps', '-a', '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying without config_id", "Did not find container with \"['podman', 'ps', '-a', '--filter', 'label=container_name=create_haproxy_wrapper', '--format', '{{.Names}}']\""* ] } I am also attaching ansible.log file for more information. Note: On Centos 8, there is no docker, so I didn't pass docker-ha.yml Can someone please help in resolving my issue Regards Anirudh Gupta -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: ansible.log Type: application/octet-stream Size: 2137544 bytes Desc: not available URL: From ykarel at redhat.com Tue Dec 28 04:37:26 2021 From: ykarel at redhat.com (Yatin Karel) Date: Tue, 28 Dec 2021 10:07:26 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines In-Reply-To: References: Message-ID: Hi Anirudh, On Mon, Dec 27, 2021 at 9:39 PM Anirudh Gupta wrote: > > Hi Team, > > I am trying to deploy TripleO Train with 3 controller and 1 Compute. > For overcloud images, I have a registry server at undercloud only. > > I executed the following command to deploy overcloud > > openstack overcloud deploy --templates \ > -r /home/stack/templates/roles_data.yaml \ > -e /home/stack/templates/node-info.yaml \ > -e /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ > -e /home/stack/containers-prepare-parameter.yaml > > The command ran for around 1.5 hrs and initially stack got successfully created and after that for 45 mins, ansible tasks were getting executed. It then gave following error in overcloud-controller-0 > > 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug output for task: Start containers for step 2 | overcloud-novacompute-0 | result={ > "changed": false, > "failed_when_result": false, > "start_containers_outputs.stdout_lines | default([]) | union(start_containers_outputs.stderr_lines | default([]))": [ > "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", > "", > "Info: Loading facts", > "Notice: Compiled catalog for overcloud-novacompute-0.localdomain in environment production in 0.05 seconds", > "Info: Applying configuration version '1640604309'", > "Notice: /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", > "Info: Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: Unscheduling all events on Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", > "Info: Creating state file /var/lib/puppet/state/state.yaml", > "Notice: Applied catalog in 0.01 seconds", > "Changes:", > " Total: 1", > "Events:", > " Success: 1", > "Resources:", > " Changed: 1", > " Out of sync: 1", > " Skipped: 7", > " Total: 8", > "Time:", > " File: 0.00", > " Transaction evaluation: 0.01", > " Catalog application: 0.01", > " Config retrieval: 0.09", > " Last run: 1640604309", > " Total: 0.01", > "Version:", > " Config: 1640604309", > " Puppet: 5.5.10", > "Error executing ['podman', 'container', 'exists', 'nova_compute_init_log']: returned 1", > "Did not find container with \"['podman', 'ps', '-a', '--filter', 'label=container_name=nova_compute_init_log', '--filter', 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying without config_id", > "Did not find container with \"['podman', 'ps', '-a', '--filter', 'label=container_name=nova_compute_init_log', '--format', '{{.Names}}']\"", > "Error executing ['podman', 'container', 'exists', 'create_haproxy_wrapper']: returned 1", > "Did not find container with \"['podman', 'ps', '-a', '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying without config_id", > "Did not find container with \"['podman', 'ps', '-a', '--filter', 'label=container_name=create_haproxy_wrapper', '--format', '{{.Names}}']\"" > ] > } This is not the actual error, actual error is: puppet-user: Error: /Stage[main]/Tripleo::Profile::Base::Rabbitmq/Rabbitmq_policy[ha-all@/]: Could not evaluate: Command is still failing after 180 seconds expired!" > > I am also attaching ansible.log file for more information. > > Note: On Centos 8, there is no docker, so I didn't pass docker-ha.yml For enabling HA and with podman in Train on CentOS8, you need to pass both docker-ha.yaml and podman.yaml in order(*order is important here*, so -e /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml -e /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml), this way you will have deployment with HA and podman, i agree docker-ha name is confusing here with podman but that has to be passed here to get the required deployment. Also with Ussuri+ HA is turned on by default so those releases may work even without passing docker-ha.yaml but for Train at least it's needed. > > Can someone please help in resolving my issue > As per your requirement I would suggest running with the above config. > Regards > Anirudh Gupta Thanks and Regards Yatin Karel -------------- next part -------------- An HTML attachment was scrubbed... URL: From ykarel at redhat.com Tue Dec 28 04:55:31 2021 From: ykarel at redhat.com (Yatin Karel) Date: Tue, 28 Dec 2021 10:25:31 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines In-Reply-To: References: Message-ID: Hi Anirudh, Sorry which timer? Timer adjustment is not needed for the issue you are seeing, if you mean overcloud deploy timeout then overcloud deploy provides the option to do so using --timeout option. The best option for now is to try docker-ha and podman in order as suggested earlier. Thanks and Regards Yatin Karel On Tue, Dec 28, 2021 at 10:12 AM Anirudh Gupta wrote: > Thanks Yatin for your response. > > Please suggest how can this timer be increased or any other steps that > needs to be followed to rectify this? > > Regards > Anirudh Gupta > > On Tue, Dec 28, 2021 at 10:08 AM Yatin Karel wrote: > >> Hi Anirudh, >> >> >> On Mon, Dec 27, 2021 at 9:39 PM Anirudh Gupta >> wrote: >> > >> > Hi Team, >> > >> > I am trying to deploy TripleO Train with 3 controller and 1 Compute. >> > For overcloud images, I have a registry server at undercloud only. >> > >> > I executed the following command to deploy overcloud >> > >> > openstack overcloud deploy --templates \ >> > -r /home/stack/templates/roles_data.yaml \ >> > -e /home/stack/templates/node-info.yaml \ >> > -e >> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >> > -e /home/stack/containers-prepare-parameter.yaml >> > >> > The command ran for around 1.5 hrs and initially stack got successfully >> created and after that for 45 mins, ansible tasks were getting executed. It >> then gave following error in overcloud-controller-0 >> > >> > 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 >> 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug >> output for task: Start containers for step 2 | overcloud-novacompute-0 | >> result={ >> > "changed": false, >> > "failed_when_result": false, >> > "start_containers_outputs.stdout_lines | default([]) | >> union(start_containers_outputs.stderr_lines | default([]))": [ >> > >> "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", >> > "", >> > "Info: Loading facts", >> > "Notice: Compiled catalog for >> overcloud-novacompute-0.localdomain in environment production in 0.05 >> seconds", >> > "Info: Applying configuration version '1640604309'", >> > "Notice: >> /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: >> defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", >> > "Info: >> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: >> Unscheduling all events on >> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", >> > "Info: Creating state file /var/lib/puppet/state/state.yaml", >> > "Notice: Applied catalog in 0.01 seconds", >> > "Changes:", >> > " Total: 1", >> > "Events:", >> > " Success: 1", >> > "Resources:", >> > " Changed: 1", >> > " Out of sync: 1", >> > " Skipped: 7", >> > " Total: 8", >> > "Time:", >> > " File: 0.00", >> > " Transaction evaluation: 0.01", >> > " Catalog application: 0.01", >> > " Config retrieval: 0.09", >> > " Last run: 1640604309", >> > " Total: 0.01", >> > "Version:", >> > " Config: 1640604309", >> > " Puppet: 5.5.10", >> > "Error executing ['podman', 'container', 'exists', >> 'nova_compute_init_log']: returned 1", >> > "Did not find container with \"['podman', 'ps', '-a', >> '--filter', 'label=container_name=nova_compute_init_log', '--filter', >> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >> without config_id", >> > "Did not find container with \"['podman', 'ps', '-a', >> '--filter', 'label=container_name=nova_compute_init_log', '--format', >> '{{.Names}}']\"", >> > "Error executing ['podman', 'container', 'exists', >> 'create_haproxy_wrapper']: returned 1", >> > "Did not find container with \"['podman', 'ps', '-a', >> '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', >> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >> without config_id", >> > "Did not find container with \"['podman', 'ps', '-a', >> '--filter', 'label=container_name=create_haproxy_wrapper', '--format', >> '{{.Names}}']\"" >> > ] >> > } >> >> This is not the actual error, actual error is: puppet-user: Error: >> /Stage[main]/Tripleo::Profile::Base::Rabbitmq/Rabbitmq_policy[ha-all@/]: >> Could not evaluate: Command is still failing after 180 seconds expired!" >> >> > >> > I am also attaching ansible.log file for more information. >> > >> > Note: On Centos 8, there is no docker, so I didn't pass docker-ha.yml >> For enabling HA and with podman in Train on CentOS8, you need to pass >> both docker-ha.yaml and podman.yaml in order(*order is important here*, >> so -e >> /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml -e >> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml), this >> way you will have deployment with HA and podman, i agree docker-ha name is >> confusing here with podman but that has to be passed here to get the >> required deployment. Also with Ussuri+ HA is turned on by default so those >> releases may work even without passing docker-ha.yaml but for Train at >> least it's needed. >> > >> > Can someone please help in resolving my issue >> > >> As per your requirement I would suggest running with the above config. >> >> > Regards >> > Anirudh Gupta >> >> Thanks and Regards >> Yatin Karel >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From ykarel at redhat.com Tue Dec 28 05:42:34 2021 From: ykarel at redhat.com (Yatin Karel) Date: Tue, 28 Dec 2021 11:12:34 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines In-Reply-To: References: Message-ID: Hi Anirudh, As said order is important here, docker-ha.yaml should be followed by podman.yaml, the parameters in environment files override the parameters from previous environment files passed and that would make deployment to use podman instead of docker. Name of the parameter to which makes this switch is "ContainerCli". Thanks and regards Yatin Karel On Tue, Dec 28, 2021 at 10:59 AM Anirudh Gupta wrote: > If this is a docker-ha issue, then that has also been tried. > > Since this is Centos 8, there is no docker available. If I pass the > docker-ha.yml, then it gives the following error > > FATAL | Pull > undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo > image | overcloud-controller-1 | error={"changed": true, "cmd": "docker > pull > undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo", > "delta": "0:00:00.005932", "end": "2021-12-27 12:42:33.927484", "msg": > "non-zero return code", "rc": 127, "start": "2021-12-27 12:42:33.921552", > "stderr": "/bin/sh: docker: command not found", "*stderr_lines": > ["/bin/sh: docker: command not found"], "stdout": "", "stdout_lines": []}* > > Regards > Anirudh Gupta > > On Tue, Dec 28, 2021 at 10:26 AM Yatin Karel wrote: > >> Hi Anirudh, >> >> Sorry which timer? Timer adjustment is not needed for the issue you are >> seeing, if you mean overcloud deploy timeout then overcloud deploy provides >> the option to do so using --timeout option. The best option for now is to >> try docker-ha and podman in order as suggested earlier. >> >> >> Thanks and Regards >> Yatin Karel >> >> On Tue, Dec 28, 2021 at 10:12 AM Anirudh Gupta >> wrote: >> >>> Thanks Yatin for your response. >>> >>> Please suggest how can this timer be increased or any other steps that >>> needs to be followed to rectify this? >>> >>> Regards >>> Anirudh Gupta >>> >>> On Tue, Dec 28, 2021 at 10:08 AM Yatin Karel wrote: >>> >>>> Hi Anirudh, >>>> >>>> >>>> On Mon, Dec 27, 2021 at 9:39 PM Anirudh Gupta >>>> wrote: >>>> > >>>> > Hi Team, >>>> > >>>> > I am trying to deploy TripleO Train with 3 controller and 1 Compute. >>>> > For overcloud images, I have a registry server at undercloud only. >>>> > >>>> > I executed the following command to deploy overcloud >>>> > >>>> > openstack overcloud deploy --templates \ >>>> > -r /home/stack/templates/roles_data.yaml \ >>>> > -e /home/stack/templates/node-info.yaml \ >>>> > -e >>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >>>> > -e /home/stack/containers-prepare-parameter.yaml >>>> > >>>> > The command ran for around 1.5 hrs and initially stack got >>>> successfully created and after that for 45 mins, ansible tasks were getting >>>> executed. It then gave following error in overcloud-controller-0 >>>> > >>>> > 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 >>>> 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug >>>> output for task: Start containers for step 2 | overcloud-novacompute-0 | >>>> result={ >>>> > "changed": false, >>>> > "failed_when_result": false, >>>> > "start_containers_outputs.stdout_lines | default([]) | >>>> union(start_containers_outputs.stderr_lines | default([]))": [ >>>> > >>>> "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", >>>> > "", >>>> > "Info: Loading facts", >>>> > "Notice: Compiled catalog for >>>> overcloud-novacompute-0.localdomain in environment production in 0.05 >>>> seconds", >>>> > "Info: Applying configuration version '1640604309'", >>>> > "Notice: >>>> /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: >>>> defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", >>>> > "Info: >>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: >>>> Unscheduling all events on >>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", >>>> > "Info: Creating state file /var/lib/puppet/state/state.yaml", >>>> > "Notice: Applied catalog in 0.01 seconds", >>>> > "Changes:", >>>> > " Total: 1", >>>> > "Events:", >>>> > " Success: 1", >>>> > "Resources:", >>>> > " Changed: 1", >>>> > " Out of sync: 1", >>>> > " Skipped: 7", >>>> > " Total: 8", >>>> > "Time:", >>>> > " File: 0.00", >>>> > " Transaction evaluation: 0.01", >>>> > " Catalog application: 0.01", >>>> > " Config retrieval: 0.09", >>>> > " Last run: 1640604309", >>>> > " Total: 0.01", >>>> > "Version:", >>>> > " Config: 1640604309", >>>> > " Puppet: 5.5.10", >>>> > "Error executing ['podman', 'container', 'exists', >>>> 'nova_compute_init_log']: returned 1", >>>> > "Did not find container with \"['podman', 'ps', '-a', >>>> '--filter', 'label=container_name=nova_compute_init_log', '--filter', >>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>> without config_id", >>>> > "Did not find container with \"['podman', 'ps', '-a', >>>> '--filter', 'label=container_name=nova_compute_init_log', '--format', >>>> '{{.Names}}']\"", >>>> > "Error executing ['podman', 'container', 'exists', >>>> 'create_haproxy_wrapper']: returned 1", >>>> > "Did not find container with \"['podman', 'ps', '-a', >>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', >>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>> without config_id", >>>> > "Did not find container with \"['podman', 'ps', '-a', >>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--format', >>>> '{{.Names}}']\"" >>>> > ] >>>> > } >>>> >>>> This is not the actual error, actual error is: puppet-user: Error: >>>> /Stage[main]/Tripleo::Profile::Base::Rabbitmq/Rabbitmq_policy[ha-all@/]: >>>> Could not evaluate: Command is still failing after 180 seconds expired!" >>>> >>>> > >>>> > I am also attaching ansible.log file for more information. >>>> > >>>> > Note: On Centos 8, there is no docker, so I didn't pass docker-ha.yml >>>> For enabling HA and with podman in Train on CentOS8, you need to pass >>>> both docker-ha.yaml and podman.yaml in order(*order is important here*, >>>> so -e >>>> /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml -e >>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml), this >>>> way you will have deployment with HA and podman, i agree docker-ha name is >>>> confusing here with podman but that has to be passed here to get the >>>> required deployment. Also with Ussuri+ HA is turned on by default so those >>>> releases may work even without passing docker-ha.yaml but for Train at >>>> least it's needed. >>>> > >>>> > Can someone please help in resolving my issue >>>> > >>>> As per your requirement I would suggest running with the above config. >>>> >>>> > Regards >>>> > Anirudh Gupta >>>> >>>> Thanks and Regards >>>> Yatin Karel >>>> >>> -------------- next part -------------- An HTML attachment was scrubbed... URL: From ykarel at redhat.com Tue Dec 28 08:57:53 2021 From: ykarel at redhat.com (Yatin Karel) Date: Tue, 28 Dec 2021 14:27:53 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines In-Reply-To: References: Message-ID: Hi Anirudh, Not sure what can cause this issue, and also the shared log file is incomplete. So I believe you tried the command on the same overcloud deployment which was failing earlier(when docker-ha.yaml was not passed). If yes, to rule out if the issue is caused by an already deployed environment can delete the overcloud and then redeploy with correct environment files as used in the last run. One reason for the password expiration that i found could be the Time is not in Sync on the overcloud nodes. So it would be good to check that as well and fix(by using correct NTP sources) before attempting redeployment. Thanks and regards Yatin Karel On Tue, Dec 28, 2021 at 2:03 PM Anirudh Gupta wrote: > Hi Yatin & Team > > Thanks for your response. > > When I executed the command as below, the installation moved ahead and > encountered another error. > > openstack overcloud deploy --templates \ > -r /home/stack/templates/roles_data.yaml \ > -e /home/stack/templates/node-info.yaml \ > -e environment.yaml \ > -e > /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml \ > -e > /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ > -e /home/stack/containers-prepare-parameter.yaml > > *Issue:* > The error was: keystoneauth1.exceptions.http.Unauthorized: *The password > is expired and needs to be changed for user*: > 4f7d1dbf58574e64af9e359cb98ccbbc. (HTTP 401) (Request-ID: > req-b29aa655-e3ec-4d4b-8ada-397f9a132582) > > I am attaching the ansible.logs for your reference. It would be a great > help if you could suggest some pointers to resolve this issue. > > Regards > Anirudh Gupta > > On Tue, Dec 28, 2021 at 11:13 AM Yatin Karel wrote: > >> Hi Anirudh, >> >> As said order is important here, docker-ha.yaml should be followed by >> podman.yaml, the parameters in environment files override the parameters >> from previous environment files passed and that would make deployment to >> use podman instead of docker. Name of the parameter to which makes this >> switch is "ContainerCli". >> >> >> Thanks and regards >> Yatin Karel >> >> On Tue, Dec 28, 2021 at 10:59 AM Anirudh Gupta >> wrote: >> >>> If this is a docker-ha issue, then that has also been tried. >>> >>> Since this is Centos 8, there is no docker available. If I pass the >>> docker-ha.yml, then it gives the following error >>> >>> FATAL | Pull >>> undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo >>> image | overcloud-controller-1 | error={"changed": true, "cmd": "docker >>> pull >>> undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo", >>> "delta": "0:00:00.005932", "end": "2021-12-27 12:42:33.927484", "msg": >>> "non-zero return code", "rc": 127, "start": "2021-12-27 12:42:33.921552", >>> "stderr": "/bin/sh: docker: command not found", "*stderr_lines": >>> ["/bin/sh: docker: command not found"], "stdout": "", "stdout_lines": []}* >>> >>> Regards >>> Anirudh Gupta >>> >>> On Tue, Dec 28, 2021 at 10:26 AM Yatin Karel wrote: >>> >>>> Hi Anirudh, >>>> >>>> Sorry which timer? Timer adjustment is not needed for the issue you are >>>> seeing, if you mean overcloud deploy timeout then overcloud deploy provides >>>> the option to do so using --timeout option. The best option for now is to >>>> try docker-ha and podman in order as suggested earlier. >>>> >>>> >>>> Thanks and Regards >>>> Yatin Karel >>>> >>>> On Tue, Dec 28, 2021 at 10:12 AM Anirudh Gupta >>>> wrote: >>>> >>>>> Thanks Yatin for your response. >>>>> >>>>> Please suggest how can this timer be increased or any other steps that >>>>> needs to be followed to rectify this? >>>>> >>>>> Regards >>>>> Anirudh Gupta >>>>> >>>>> On Tue, Dec 28, 2021 at 10:08 AM Yatin Karel >>>>> wrote: >>>>> >>>>>> Hi Anirudh, >>>>>> >>>>>> >>>>>> On Mon, Dec 27, 2021 at 9:39 PM Anirudh Gupta >>>>>> wrote: >>>>>> > >>>>>> > Hi Team, >>>>>> > >>>>>> > I am trying to deploy TripleO Train with 3 controller and 1 Compute. >>>>>> > For overcloud images, I have a registry server at undercloud only. >>>>>> > >>>>>> > I executed the following command to deploy overcloud >>>>>> > >>>>>> > openstack overcloud deploy --templates \ >>>>>> > -r /home/stack/templates/roles_data.yaml \ >>>>>> > -e /home/stack/templates/node-info.yaml \ >>>>>> > -e >>>>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >>>>>> > -e /home/stack/containers-prepare-parameter.yaml >>>>>> > >>>>>> > The command ran for around 1.5 hrs and initially stack got >>>>>> successfully created and after that for 45 mins, ansible tasks were getting >>>>>> executed. It then gave following error in overcloud-controller-0 >>>>>> > >>>>>> > 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 >>>>>> 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug >>>>>> output for task: Start containers for step 2 | overcloud-novacompute-0 | >>>>>> result={ >>>>>> > "changed": false, >>>>>> > "failed_when_result": false, >>>>>> > "start_containers_outputs.stdout_lines | default([]) | >>>>>> union(start_containers_outputs.stderr_lines | default([]))": [ >>>>>> > >>>>>> "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", >>>>>> > "", >>>>>> > "Info: Loading facts", >>>>>> > "Notice: Compiled catalog for >>>>>> overcloud-novacompute-0.localdomain in environment production in 0.05 >>>>>> seconds", >>>>>> > "Info: Applying configuration version '1640604309'", >>>>>> > "Notice: >>>>>> /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: >>>>>> defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", >>>>>> > "Info: >>>>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: >>>>>> Unscheduling all events on >>>>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", >>>>>> > "Info: Creating state file >>>>>> /var/lib/puppet/state/state.yaml", >>>>>> > "Notice: Applied catalog in 0.01 seconds", >>>>>> > "Changes:", >>>>>> > " Total: 1", >>>>>> > "Events:", >>>>>> > " Success: 1", >>>>>> > "Resources:", >>>>>> > " Changed: 1", >>>>>> > " Out of sync: 1", >>>>>> > " Skipped: 7", >>>>>> > " Total: 8", >>>>>> > "Time:", >>>>>> > " File: 0.00", >>>>>> > " Transaction evaluation: 0.01", >>>>>> > " Catalog application: 0.01", >>>>>> > " Config retrieval: 0.09", >>>>>> > " Last run: 1640604309", >>>>>> > " Total: 0.01", >>>>>> > "Version:", >>>>>> > " Config: 1640604309", >>>>>> > " Puppet: 5.5.10", >>>>>> > "Error executing ['podman', 'container', 'exists', >>>>>> 'nova_compute_init_log']: returned 1", >>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>> '--filter', 'label=container_name=nova_compute_init_log', '--filter', >>>>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>>>> without config_id", >>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>> '--filter', 'label=container_name=nova_compute_init_log', '--format', >>>>>> '{{.Names}}']\"", >>>>>> > "Error executing ['podman', 'container', 'exists', >>>>>> 'create_haproxy_wrapper']: returned 1", >>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', >>>>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>>>> without config_id", >>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--format', >>>>>> '{{.Names}}']\"" >>>>>> > ] >>>>>> > } >>>>>> >>>>>> This is not the actual error, actual error is: puppet-user: Error: >>>>>> /Stage[main]/Tripleo::Profile::Base::Rabbitmq/Rabbitmq_policy[ha-all@/]: >>>>>> Could not evaluate: Command is still failing after 180 seconds expired!" >>>>>> >>>>>> > >>>>>> > I am also attaching ansible.log file for more information. >>>>>> > >>>>>> > Note: On Centos 8, there is no docker, so I didn't pass >>>>>> docker-ha.yml >>>>>> For enabling HA and with podman in Train on CentOS8, you need to pass >>>>>> both docker-ha.yaml and podman.yaml in order(*order is important >>>>>> here*, so -e >>>>>> /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml -e >>>>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml), this >>>>>> way you will have deployment with HA and podman, i agree docker-ha name is >>>>>> confusing here with podman but that has to be passed here to get the >>>>>> required deployment. Also with Ussuri+ HA is turned on by default so those >>>>>> releases may work even without passing docker-ha.yaml but for Train at >>>>>> least it's needed. >>>>>> > >>>>>> > Can someone please help in resolving my issue >>>>>> > >>>>>> As per your requirement I would suggest running with the above config. >>>>>> >>>>>> > Regards >>>>>> > Anirudh Gupta >>>>>> >>>>>> Thanks and Regards >>>>>> Yatin Karel >>>>>> >>>>> -------------- next part -------------- An HTML attachment was scrubbed... URL: From ykarel at redhat.com Tue Dec 28 11:27:52 2021 From: ykarel at redhat.com (Yatin Karel) Date: Tue, 28 Dec 2021 16:57:52 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines In-Reply-To: References: Message-ID: Hi Anirudh, On Tue, Dec 28, 2021 at 4:32 PM Anirudh Gupta wrote: > Hi Yatin, > > Thanks a lot for your help. I am deleting the stack and running the > overcloud deploy command as a process. > > Changing the NTP settings worked for me in proceeding ahead. > The current result looks good, it moved ahead a lot. As per the latest logs/error I feel you still have issues with NTP[1], you can check/confirm by running "chrony sources" command on overcloud nodes. The current issue should be clear once you have working NTP. Once you check/fix NTP thing you can rerun the same overcloud deploy command and it should move forward. [1] https://bugs.launchpad.net/tripleo/+bug/1955414 > > But it seems the issues are not ending here. > > I would require some more help from you in order to deploy this. > > *Issue:* > > FATAL | Check Keystone service status | undercloud | item=heat-cfn | > error={"ansible_job_id": "687227427425.307276", "ansible_loop_var": > "tripleo_keystone_resources_service_async_result_item", "attempts": 1, > "changed": false, "extra_data": {"data": null, "details": "The request you > have made requires authentication.", "response": > "{\"error\":{\"code\":401,\"message\":\"The request you have made requires > authentication.\",\"title\":\"Unauthorized\"}}\n"}, "finished": 1, "msg": > "Failed to list services: Client Error for url: > http://10.10.30.222:5000/v3/services, *The request you have made requires > authentication.",* > "tripleo_keystone_resources_service_async_result_item": {"ansible_job_id": > "687227427425.307276", "ansible_loop_var": > "tripleo_keystone_resources_data", "changed": true, "failed": false, > "finished": 0, "results_file": "/root/.ansible_async/687227427425.307276", > "started": 1, "tripleo_keystone_resources_data": {"key": "heat-cfn", > "value": {"endpoints": {"admin": "http://10.10.30.222:8000/v1", > "internal": "http://10.10.30.222:8000/v1", "public": " > http://10.10.30.222:8000/v1"}, "region": "regionOne", "service": > "cloudformation", "users": {"heat-cfn": {"password": > "3f3tHhxhna1CpRVPMjF7po49F"}}}}}} > > > PFA the ansible.log file. > > Thanks your help and Patience. > > Regards > Anirudh Gupta > > On Tue, Dec 28, 2021 at 2:28 PM Yatin Karel wrote: > >> Hi Anirudh, >> >> Not sure what can cause this issue, and also the shared log file is >> incomplete. So I believe you tried the command on the same overcloud >> deployment which was failing earlier(when docker-ha.yaml was not passed). >> If yes, to rule out if the issue is caused by an already deployed >> environment can delete the overcloud and then redeploy with correct >> environment files as used in the last run. >> >> One reason for the password expiration that i found could be the Time is >> not in Sync on the overcloud nodes. So it would be good to check that as >> well and fix(by using correct NTP sources) before attempting redeployment. >> >> Thanks and regards >> Yatin Karel >> >> >> >> On Tue, Dec 28, 2021 at 2:03 PM Anirudh Gupta >> wrote: >> >>> Hi Yatin & Team >>> >>> Thanks for your response. >>> >>> When I executed the command as below, the installation moved ahead and >>> encountered another error. >>> >>> openstack overcloud deploy --templates \ >>> -r /home/stack/templates/roles_data.yaml \ >>> -e /home/stack/templates/node-info.yaml \ >>> -e environment.yaml \ >>> -e >>> /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml \ >>> -e >>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >>> -e /home/stack/containers-prepare-parameter.yaml >>> >>> *Issue:* >>> The error was: keystoneauth1.exceptions.http.Unauthorized: *The >>> password is expired and needs to be changed for user*: >>> 4f7d1dbf58574e64af9e359cb98ccbbc. (HTTP 401) (Request-ID: >>> req-b29aa655-e3ec-4d4b-8ada-397f9a132582) >>> >>> I am attaching the ansible.logs for your reference. It would be a great >>> help if you could suggest some pointers to resolve this issue. >>> >>> Regards >>> Anirudh Gupta >>> >>> On Tue, Dec 28, 2021 at 11:13 AM Yatin Karel wrote: >>> >>>> Hi Anirudh, >>>> >>>> As said order is important here, docker-ha.yaml should be followed by >>>> podman.yaml, the parameters in environment files override the parameters >>>> from previous environment files passed and that would make deployment to >>>> use podman instead of docker. Name of the parameter to which makes this >>>> switch is "ContainerCli". >>>> >>>> >>>> Thanks and regards >>>> Yatin Karel >>>> >>>> On Tue, Dec 28, 2021 at 10:59 AM Anirudh Gupta >>>> wrote: >>>> >>>>> If this is a docker-ha issue, then that has also been tried. >>>>> >>>>> Since this is Centos 8, there is no docker available. If I pass the >>>>> docker-ha.yml, then it gives the following error >>>>> >>>>> FATAL | Pull >>>>> undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo >>>>> image | overcloud-controller-1 | error={"changed": true, "cmd": "docker >>>>> pull >>>>> undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo", >>>>> "delta": "0:00:00.005932", "end": "2021-12-27 12:42:33.927484", "msg": >>>>> "non-zero return code", "rc": 127, "start": "2021-12-27 12:42:33.921552", >>>>> "stderr": "/bin/sh: docker: command not found", "*stderr_lines": >>>>> ["/bin/sh: docker: command not found"], "stdout": "", "stdout_lines": []}* >>>>> >>>>> Regards >>>>> Anirudh Gupta >>>>> >>>>> On Tue, Dec 28, 2021 at 10:26 AM Yatin Karel >>>>> wrote: >>>>> >>>>>> Hi Anirudh, >>>>>> >>>>>> Sorry which timer? Timer adjustment is not needed for the issue you >>>>>> are seeing, if you mean overcloud deploy timeout then overcloud deploy >>>>>> provides the option to do so using --timeout option. The best option for >>>>>> now is to try docker-ha and podman in order as suggested earlier. >>>>>> >>>>>> >>>>>> Thanks and Regards >>>>>> Yatin Karel >>>>>> >>>>>> On Tue, Dec 28, 2021 at 10:12 AM Anirudh Gupta >>>>>> wrote: >>>>>> >>>>>>> Thanks Yatin for your response. >>>>>>> >>>>>>> Please suggest how can this timer be increased or any other steps >>>>>>> that needs to be followed to rectify this? >>>>>>> >>>>>>> Regards >>>>>>> Anirudh Gupta >>>>>>> >>>>>>> On Tue, Dec 28, 2021 at 10:08 AM Yatin Karel >>>>>>> wrote: >>>>>>> >>>>>>>> Hi Anirudh, >>>>>>>> >>>>>>>> >>>>>>>> On Mon, Dec 27, 2021 at 9:39 PM Anirudh Gupta >>>>>>>> wrote: >>>>>>>> > >>>>>>>> > Hi Team, >>>>>>>> > >>>>>>>> > I am trying to deploy TripleO Train with 3 controller and 1 >>>>>>>> Compute. >>>>>>>> > For overcloud images, I have a registry server at undercloud only. >>>>>>>> > >>>>>>>> > I executed the following command to deploy overcloud >>>>>>>> > >>>>>>>> > openstack overcloud deploy --templates \ >>>>>>>> > -r /home/stack/templates/roles_data.yaml \ >>>>>>>> > -e /home/stack/templates/node-info.yaml \ >>>>>>>> > -e >>>>>>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >>>>>>>> > -e /home/stack/containers-prepare-parameter.yaml >>>>>>>> > >>>>>>>> > The command ran for around 1.5 hrs and initially stack got >>>>>>>> successfully created and after that for 45 mins, ansible tasks were getting >>>>>>>> executed. It then gave following error in overcloud-controller-0 >>>>>>>> > >>>>>>>> > 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 >>>>>>>> 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug >>>>>>>> output for task: Start containers for step 2 | overcloud-novacompute-0 | >>>>>>>> result={ >>>>>>>> > "changed": false, >>>>>>>> > "failed_when_result": false, >>>>>>>> > "start_containers_outputs.stdout_lines | default([]) | >>>>>>>> union(start_containers_outputs.stderr_lines | default([]))": [ >>>>>>>> > >>>>>>>> "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", >>>>>>>> > "", >>>>>>>> > "Info: Loading facts", >>>>>>>> > "Notice: Compiled catalog for >>>>>>>> overcloud-novacompute-0.localdomain in environment production in 0.05 >>>>>>>> seconds", >>>>>>>> > "Info: Applying configuration version '1640604309'", >>>>>>>> > "Notice: >>>>>>>> /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: >>>>>>>> defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", >>>>>>>> > "Info: >>>>>>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: >>>>>>>> Unscheduling all events on >>>>>>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", >>>>>>>> > "Info: Creating state file >>>>>>>> /var/lib/puppet/state/state.yaml", >>>>>>>> > "Notice: Applied catalog in 0.01 seconds", >>>>>>>> > "Changes:", >>>>>>>> > " Total: 1", >>>>>>>> > "Events:", >>>>>>>> > " Success: 1", >>>>>>>> > "Resources:", >>>>>>>> > " Changed: 1", >>>>>>>> > " Out of sync: 1", >>>>>>>> > " Skipped: 7", >>>>>>>> > " Total: 8", >>>>>>>> > "Time:", >>>>>>>> > " File: 0.00", >>>>>>>> > " Transaction evaluation: 0.01", >>>>>>>> > " Catalog application: 0.01", >>>>>>>> > " Config retrieval: 0.09", >>>>>>>> > " Last run: 1640604309", >>>>>>>> > " Total: 0.01", >>>>>>>> > "Version:", >>>>>>>> > " Config: 1640604309", >>>>>>>> > " Puppet: 5.5.10", >>>>>>>> > "Error executing ['podman', 'container', 'exists', >>>>>>>> 'nova_compute_init_log']: returned 1", >>>>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>>>> '--filter', 'label=container_name=nova_compute_init_log', '--filter', >>>>>>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>>>>>> without config_id", >>>>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>>>> '--filter', 'label=container_name=nova_compute_init_log', '--format', >>>>>>>> '{{.Names}}']\"", >>>>>>>> > "Error executing ['podman', 'container', 'exists', >>>>>>>> 'create_haproxy_wrapper']: returned 1", >>>>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', >>>>>>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>>>>>> without config_id", >>>>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--format', >>>>>>>> '{{.Names}}']\"" >>>>>>>> > ] >>>>>>>> > } >>>>>>>> >>>>>>>> This is not the actual error, actual error is: puppet-user: Error: >>>>>>>> /Stage[main]/Tripleo::Profile::Base::Rabbitmq/Rabbitmq_policy[ha-all@/]: >>>>>>>> Could not evaluate: Command is still failing after 180 seconds expired!" >>>>>>>> >>>>>>>> > >>>>>>>> > I am also attaching ansible.log file for more information. >>>>>>>> > >>>>>>>> > Note: On Centos 8, there is no docker, so I didn't pass >>>>>>>> docker-ha.yml >>>>>>>> For enabling HA and with podman in Train on CentOS8, you need to >>>>>>>> pass both docker-ha.yaml and podman.yaml in order(*order is >>>>>>>> important here*, so -e >>>>>>>> /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml -e >>>>>>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml), this >>>>>>>> way you will have deployment with HA and podman, i agree docker-ha name is >>>>>>>> confusing here with podman but that has to be passed here to get the >>>>>>>> required deployment. Also with Ussuri+ HA is turned on by default so those >>>>>>>> releases may work even without passing docker-ha.yaml but for Train at >>>>>>>> least it's needed. >>>>>>>> > >>>>>>>> > Can someone please help in resolving my issue >>>>>>>> > >>>>>>>> As per your requirement I would suggest running with the above >>>>>>>> config. >>>>>>>> >>>>>>>> > Regards >>>>>>>> > Anirudh Gupta >>>>>>>> >>>>>>>> Thanks and Regards >>>>>>>> Yatin Karel >>>>>>>> >>>>>>> Thanks and Regards Yatin Karel -------------- next part -------------- An HTML attachment was scrubbed... URL: From syedammad83 at gmail.com Tue Dec 28 12:50:50 2021 From: syedammad83 at gmail.com (Ammad Syed) Date: Tue, 28 Dec 2021 17:50:50 +0500 Subject: [octavia] Management Network Issue Message-ID: Hi, I am using octavia 9.0. I have created a neutron vlan based management network of octavia. I am creating a loadbalancer with a subnet in the tenant network, its getting created successfully. Then added a listener and created a pool, both created successfully. Now a weird situation is happening. Now when I add a member in the pool, the management network interface get detached automatically from the AMPHORA instance and amphora keeps in PENDING_UPDATE state. I have created a dedicated service-project for octavia and management network and subnet resides there. Any advise how to fix it ? Ammad -------------- next part -------------- An HTML attachment was scrubbed... URL: From anyrude10 at gmail.com Tue Dec 28 04:41:49 2021 From: anyrude10 at gmail.com (Anirudh Gupta) Date: Tue, 28 Dec 2021 10:11:49 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines In-Reply-To: References: Message-ID: Thanks Yatin for your response. Please suggest how can this timer be increased or any other steps that needs to be followed to rectify this? Regards Anirudh Gupta On Tue, Dec 28, 2021 at 10:08 AM Yatin Karel wrote: > Hi Anirudh, > > > On Mon, Dec 27, 2021 at 9:39 PM Anirudh Gupta wrote: > > > > Hi Team, > > > > I am trying to deploy TripleO Train with 3 controller and 1 Compute. > > For overcloud images, I have a registry server at undercloud only. > > > > I executed the following command to deploy overcloud > > > > openstack overcloud deploy --templates \ > > -r /home/stack/templates/roles_data.yaml \ > > -e /home/stack/templates/node-info.yaml \ > > -e > /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ > > -e /home/stack/containers-prepare-parameter.yaml > > > > The command ran for around 1.5 hrs and initially stack got successfully > created and after that for 45 mins, ansible tasks were getting executed. It > then gave following error in overcloud-controller-0 > > > > 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 > 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug > output for task: Start containers for step 2 | overcloud-novacompute-0 | > result={ > > "changed": false, > > "failed_when_result": false, > > "start_containers_outputs.stdout_lines | default([]) | > union(start_containers_outputs.stderr_lines | default([]))": [ > > > "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", > > "", > > "Info: Loading facts", > > "Notice: Compiled catalog for > overcloud-novacompute-0.localdomain in environment production in 0.05 > seconds", > > "Info: Applying configuration version '1640604309'", > > "Notice: > /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: > defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", > > "Info: > Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: > Unscheduling all events on > Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", > > "Info: Creating state file /var/lib/puppet/state/state.yaml", > > "Notice: Applied catalog in 0.01 seconds", > > "Changes:", > > " Total: 1", > > "Events:", > > " Success: 1", > > "Resources:", > > " Changed: 1", > > " Out of sync: 1", > > " Skipped: 7", > > " Total: 8", > > "Time:", > > " File: 0.00", > > " Transaction evaluation: 0.01", > > " Catalog application: 0.01", > > " Config retrieval: 0.09", > > " Last run: 1640604309", > > " Total: 0.01", > > "Version:", > > " Config: 1640604309", > > " Puppet: 5.5.10", > > "Error executing ['podman', 'container', 'exists', > 'nova_compute_init_log']: returned 1", > > "Did not find container with \"['podman', 'ps', '-a', > '--filter', 'label=container_name=nova_compute_init_log', '--filter', > 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying > without config_id", > > "Did not find container with \"['podman', 'ps', '-a', > '--filter', 'label=container_name=nova_compute_init_log', '--format', > '{{.Names}}']\"", > > "Error executing ['podman', 'container', 'exists', > 'create_haproxy_wrapper']: returned 1", > > "Did not find container with \"['podman', 'ps', '-a', > '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', > 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying > without config_id", > > "Did not find container with \"['podman', 'ps', '-a', > '--filter', 'label=container_name=create_haproxy_wrapper', '--format', > '{{.Names}}']\"" > > ] > > } > > This is not the actual error, actual error is: puppet-user: Error: > /Stage[main]/Tripleo::Profile::Base::Rabbitmq/Rabbitmq_policy[ha-all@/]: > Could not evaluate: Command is still failing after 180 seconds expired!" > > > > > I am also attaching ansible.log file for more information. > > > > Note: On Centos 8, there is no docker, so I didn't pass docker-ha.yml > For enabling HA and with podman in Train on CentOS8, you need to pass both > docker-ha.yaml and podman.yaml in order(*order is important here*, so -e > /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml -e > /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml), this > way you will have deployment with HA and podman, i agree docker-ha name is > confusing here with podman but that has to be passed here to get the > required deployment. Also with Ussuri+ HA is turned on by default so those > releases may work even without passing docker-ha.yaml but for Train at > least it's needed. > > > > Can someone please help in resolving my issue > > > As per your requirement I would suggest running with the above config. > > > Regards > > Anirudh Gupta > > Thanks and Regards > Yatin Karel > -------------- next part -------------- An HTML attachment was scrubbed... URL: From anyrude10 at gmail.com Tue Dec 28 05:28:45 2021 From: anyrude10 at gmail.com (Anirudh Gupta) Date: Tue, 28 Dec 2021 10:58:45 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines In-Reply-To: References: Message-ID: If this is a docker-ha issue, then that has also been tried. Since this is Centos 8, there is no docker available. If I pass the docker-ha.yml, then it gives the following error FATAL | Pull undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo image | overcloud-controller-1 | error={"changed": true, "cmd": "docker pull undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo", "delta": "0:00:00.005932", "end": "2021-12-27 12:42:33.927484", "msg": "non-zero return code", "rc": 127, "start": "2021-12-27 12:42:33.921552", "stderr": "/bin/sh: docker: command not found", "*stderr_lines": ["/bin/sh: docker: command not found"], "stdout": "", "stdout_lines": []}* Regards Anirudh Gupta On Tue, Dec 28, 2021 at 10:26 AM Yatin Karel wrote: > Hi Anirudh, > > Sorry which timer? Timer adjustment is not needed for the issue you are > seeing, if you mean overcloud deploy timeout then overcloud deploy provides > the option to do so using --timeout option. The best option for now is to > try docker-ha and podman in order as suggested earlier. > > > Thanks and Regards > Yatin Karel > > On Tue, Dec 28, 2021 at 10:12 AM Anirudh Gupta > wrote: > >> Thanks Yatin for your response. >> >> Please suggest how can this timer be increased or any other steps that >> needs to be followed to rectify this? >> >> Regards >> Anirudh Gupta >> >> On Tue, Dec 28, 2021 at 10:08 AM Yatin Karel wrote: >> >>> Hi Anirudh, >>> >>> >>> On Mon, Dec 27, 2021 at 9:39 PM Anirudh Gupta >>> wrote: >>> > >>> > Hi Team, >>> > >>> > I am trying to deploy TripleO Train with 3 controller and 1 Compute. >>> > For overcloud images, I have a registry server at undercloud only. >>> > >>> > I executed the following command to deploy overcloud >>> > >>> > openstack overcloud deploy --templates \ >>> > -r /home/stack/templates/roles_data.yaml \ >>> > -e /home/stack/templates/node-info.yaml \ >>> > -e >>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >>> > -e /home/stack/containers-prepare-parameter.yaml >>> > >>> > The command ran for around 1.5 hrs and initially stack got >>> successfully created and after that for 45 mins, ansible tasks were getting >>> executed. It then gave following error in overcloud-controller-0 >>> > >>> > 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 >>> 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug >>> output for task: Start containers for step 2 | overcloud-novacompute-0 | >>> result={ >>> > "changed": false, >>> > "failed_when_result": false, >>> > "start_containers_outputs.stdout_lines | default([]) | >>> union(start_containers_outputs.stderr_lines | default([]))": [ >>> > >>> "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", >>> > "", >>> > "Info: Loading facts", >>> > "Notice: Compiled catalog for >>> overcloud-novacompute-0.localdomain in environment production in 0.05 >>> seconds", >>> > "Info: Applying configuration version '1640604309'", >>> > "Notice: >>> /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: >>> defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", >>> > "Info: >>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: >>> Unscheduling all events on >>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", >>> > "Info: Creating state file /var/lib/puppet/state/state.yaml", >>> > "Notice: Applied catalog in 0.01 seconds", >>> > "Changes:", >>> > " Total: 1", >>> > "Events:", >>> > " Success: 1", >>> > "Resources:", >>> > " Changed: 1", >>> > " Out of sync: 1", >>> > " Skipped: 7", >>> > " Total: 8", >>> > "Time:", >>> > " File: 0.00", >>> > " Transaction evaluation: 0.01", >>> > " Catalog application: 0.01", >>> > " Config retrieval: 0.09", >>> > " Last run: 1640604309", >>> > " Total: 0.01", >>> > "Version:", >>> > " Config: 1640604309", >>> > " Puppet: 5.5.10", >>> > "Error executing ['podman', 'container', 'exists', >>> 'nova_compute_init_log']: returned 1", >>> > "Did not find container with \"['podman', 'ps', '-a', >>> '--filter', 'label=container_name=nova_compute_init_log', '--filter', >>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>> without config_id", >>> > "Did not find container with \"['podman', 'ps', '-a', >>> '--filter', 'label=container_name=nova_compute_init_log', '--format', >>> '{{.Names}}']\"", >>> > "Error executing ['podman', 'container', 'exists', >>> 'create_haproxy_wrapper']: returned 1", >>> > "Did not find container with \"['podman', 'ps', '-a', >>> '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', >>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>> without config_id", >>> > "Did not find container with \"['podman', 'ps', '-a', >>> '--filter', 'label=container_name=create_haproxy_wrapper', '--format', >>> '{{.Names}}']\"" >>> > ] >>> > } >>> >>> This is not the actual error, actual error is: puppet-user: Error: >>> /Stage[main]/Tripleo::Profile::Base::Rabbitmq/Rabbitmq_policy[ha-all@/]: >>> Could not evaluate: Command is still failing after 180 seconds expired!" >>> >>> > >>> > I am also attaching ansible.log file for more information. >>> > >>> > Note: On Centos 8, there is no docker, so I didn't pass docker-ha.yml >>> For enabling HA and with podman in Train on CentOS8, you need to pass >>> both docker-ha.yaml and podman.yaml in order(*order is important here*, >>> so -e >>> /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml -e >>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml), this >>> way you will have deployment with HA and podman, i agree docker-ha name is >>> confusing here with podman but that has to be passed here to get the >>> required deployment. Also with Ussuri+ HA is turned on by default so those >>> releases may work even without passing docker-ha.yaml but for Train at >>> least it's needed. >>> > >>> > Can someone please help in resolving my issue >>> > >>> As per your requirement I would suggest running with the above config. >>> >>> > Regards >>> > Anirudh Gupta >>> >>> Thanks and Regards >>> Yatin Karel >>> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: From anyrude10 at gmail.com Tue Dec 28 08:33:37 2021 From: anyrude10 at gmail.com (Anirudh Gupta) Date: Tue, 28 Dec 2021 14:03:37 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines In-Reply-To: References: Message-ID: Hi Yatin & Team Thanks for your response. When I executed the command as below, the installation moved ahead and encountered another error. openstack overcloud deploy --templates \ -r /home/stack/templates/roles_data.yaml \ -e /home/stack/templates/node-info.yaml \ -e environment.yaml \ -e /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml \ -e /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ -e /home/stack/containers-prepare-parameter.yaml *Issue:* The error was: keystoneauth1.exceptions.http.Unauthorized: *The password is expired and needs to be changed for user*: 4f7d1dbf58574e64af9e359cb98ccbbc. (HTTP 401) (Request-ID: req-b29aa655-e3ec-4d4b-8ada-397f9a132582) I am attaching the ansible.logs for your reference. It would be a great help if you could suggest some pointers to resolve this issue. Regards Anirudh Gupta On Tue, Dec 28, 2021 at 11:13 AM Yatin Karel wrote: > Hi Anirudh, > > As said order is important here, docker-ha.yaml should be followed by > podman.yaml, the parameters in environment files override the parameters > from previous environment files passed and that would make deployment to > use podman instead of docker. Name of the parameter to which makes this > switch is "ContainerCli". > > > Thanks and regards > Yatin Karel > > On Tue, Dec 28, 2021 at 10:59 AM Anirudh Gupta > wrote: > >> If this is a docker-ha issue, then that has also been tried. >> >> Since this is Centos 8, there is no docker available. If I pass the >> docker-ha.yml, then it gives the following error >> >> FATAL | Pull >> undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo >> image | overcloud-controller-1 | error={"changed": true, "cmd": "docker >> pull >> undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo", >> "delta": "0:00:00.005932", "end": "2021-12-27 12:42:33.927484", "msg": >> "non-zero return code", "rc": 127, "start": "2021-12-27 12:42:33.921552", >> "stderr": "/bin/sh: docker: command not found", "*stderr_lines": >> ["/bin/sh: docker: command not found"], "stdout": "", "stdout_lines": []}* >> >> Regards >> Anirudh Gupta >> >> On Tue, Dec 28, 2021 at 10:26 AM Yatin Karel wrote: >> >>> Hi Anirudh, >>> >>> Sorry which timer? Timer adjustment is not needed for the issue you are >>> seeing, if you mean overcloud deploy timeout then overcloud deploy provides >>> the option to do so using --timeout option. The best option for now is to >>> try docker-ha and podman in order as suggested earlier. >>> >>> >>> Thanks and Regards >>> Yatin Karel >>> >>> On Tue, Dec 28, 2021 at 10:12 AM Anirudh Gupta >>> wrote: >>> >>>> Thanks Yatin for your response. >>>> >>>> Please suggest how can this timer be increased or any other steps that >>>> needs to be followed to rectify this? >>>> >>>> Regards >>>> Anirudh Gupta >>>> >>>> On Tue, Dec 28, 2021 at 10:08 AM Yatin Karel wrote: >>>> >>>>> Hi Anirudh, >>>>> >>>>> >>>>> On Mon, Dec 27, 2021 at 9:39 PM Anirudh Gupta >>>>> wrote: >>>>> > >>>>> > Hi Team, >>>>> > >>>>> > I am trying to deploy TripleO Train with 3 controller and 1 Compute. >>>>> > For overcloud images, I have a registry server at undercloud only. >>>>> > >>>>> > I executed the following command to deploy overcloud >>>>> > >>>>> > openstack overcloud deploy --templates \ >>>>> > -r /home/stack/templates/roles_data.yaml \ >>>>> > -e /home/stack/templates/node-info.yaml \ >>>>> > -e >>>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >>>>> > -e /home/stack/containers-prepare-parameter.yaml >>>>> > >>>>> > The command ran for around 1.5 hrs and initially stack got >>>>> successfully created and after that for 45 mins, ansible tasks were getting >>>>> executed. It then gave following error in overcloud-controller-0 >>>>> > >>>>> > 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 >>>>> 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug >>>>> output for task: Start containers for step 2 | overcloud-novacompute-0 | >>>>> result={ >>>>> > "changed": false, >>>>> > "failed_when_result": false, >>>>> > "start_containers_outputs.stdout_lines | default([]) | >>>>> union(start_containers_outputs.stderr_lines | default([]))": [ >>>>> > >>>>> "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", >>>>> > "", >>>>> > "Info: Loading facts", >>>>> > "Notice: Compiled catalog for >>>>> overcloud-novacompute-0.localdomain in environment production in 0.05 >>>>> seconds", >>>>> > "Info: Applying configuration version '1640604309'", >>>>> > "Notice: >>>>> /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: >>>>> defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", >>>>> > "Info: >>>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: >>>>> Unscheduling all events on >>>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", >>>>> > "Info: Creating state file /var/lib/puppet/state/state.yaml", >>>>> > "Notice: Applied catalog in 0.01 seconds", >>>>> > "Changes:", >>>>> > " Total: 1", >>>>> > "Events:", >>>>> > " Success: 1", >>>>> > "Resources:", >>>>> > " Changed: 1", >>>>> > " Out of sync: 1", >>>>> > " Skipped: 7", >>>>> > " Total: 8", >>>>> > "Time:", >>>>> > " File: 0.00", >>>>> > " Transaction evaluation: 0.01", >>>>> > " Catalog application: 0.01", >>>>> > " Config retrieval: 0.09", >>>>> > " Last run: 1640604309", >>>>> > " Total: 0.01", >>>>> > "Version:", >>>>> > " Config: 1640604309", >>>>> > " Puppet: 5.5.10", >>>>> > "Error executing ['podman', 'container', 'exists', >>>>> 'nova_compute_init_log']: returned 1", >>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>> '--filter', 'label=container_name=nova_compute_init_log', '--filter', >>>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>>> without config_id", >>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>> '--filter', 'label=container_name=nova_compute_init_log', '--format', >>>>> '{{.Names}}']\"", >>>>> > "Error executing ['podman', 'container', 'exists', >>>>> 'create_haproxy_wrapper']: returned 1", >>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', >>>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>>> without config_id", >>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--format', >>>>> '{{.Names}}']\"" >>>>> > ] >>>>> > } >>>>> >>>>> This is not the actual error, actual error is: puppet-user: Error: >>>>> /Stage[main]/Tripleo::Profile::Base::Rabbitmq/Rabbitmq_policy[ha-all@/]: >>>>> Could not evaluate: Command is still failing after 180 seconds expired!" >>>>> >>>>> > >>>>> > I am also attaching ansible.log file for more information. >>>>> > >>>>> > Note: On Centos 8, there is no docker, so I didn't pass docker-ha.yml >>>>> For enabling HA and with podman in Train on CentOS8, you need to pass >>>>> both docker-ha.yaml and podman.yaml in order(*order is important here*, >>>>> so -e >>>>> /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml -e >>>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml), this >>>>> way you will have deployment with HA and podman, i agree docker-ha name is >>>>> confusing here with podman but that has to be passed here to get the >>>>> required deployment. Also with Ussuri+ HA is turned on by default so those >>>>> releases may work even without passing docker-ha.yaml but for Train at >>>>> least it's needed. >>>>> > >>>>> > Can someone please help in resolving my issue >>>>> > >>>>> As per your requirement I would suggest running with the above config. >>>>> >>>>> > Regards >>>>> > Anirudh Gupta >>>>> >>>>> Thanks and Regards >>>>> Yatin Karel >>>>> >>>> -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: ansible.log Type: application/octet-stream Size: 65536 bytes Desc: not available URL: From anyrude10 at gmail.com Tue Dec 28 11:02:06 2021 From: anyrude10 at gmail.com (Anirudh Gupta) Date: Tue, 28 Dec 2021 16:32:06 +0530 Subject: [TripleO] Unable to deploy Overcloud Machines In-Reply-To: References: Message-ID: Hi Yatin, Thanks a lot for your help. I am deleting the stack and running the overcloud deploy command as a process. Changing the NTP settings worked for me in proceeding ahead. But it seems the issues are not ending here. I would require some more help from you in order to deploy this. *Issue:* FATAL | Check Keystone service status | undercloud | item=heat-cfn | error={"ansible_job_id": "687227427425.307276", "ansible_loop_var": "tripleo_keystone_resources_service_async_result_item", "attempts": 1, "changed": false, "extra_data": {"data": null, "details": "The request you have made requires authentication.", "response": "{\"error\":{\"code\":401,\"message\":\"The request you have made requires authentication.\",\"title\":\"Unauthorized\"}}\n"}, "finished": 1, "msg": "Failed to list services: Client Error for url: http://10.10.30.222:5000/v3/services, *The request you have made requires authentication.",* "tripleo_keystone_resources_service_async_result_item": {"ansible_job_id": "687227427425.307276", "ansible_loop_var": "tripleo_keystone_resources_data", "changed": true, "failed": false, "finished": 0, "results_file": "/root/.ansible_async/687227427425.307276", "started": 1, "tripleo_keystone_resources_data": {"key": "heat-cfn", "value": {"endpoints": {"admin": "http://10.10.30.222:8000/v1", "internal": "http://10.10.30.222:8000/v1", "public": "http://10.10.30.222:8000/v1"}, "region": "regionOne", "service": "cloudformation", "users": {"heat-cfn": {"password": "3f3tHhxhna1CpRVPMjF7po49F"}}}}}} PFA the ansible.log file. Thanks your help and Patience. Regards Anirudh Gupta On Tue, Dec 28, 2021 at 2:28 PM Yatin Karel wrote: > Hi Anirudh, > > Not sure what can cause this issue, and also the shared log file is > incomplete. So I believe you tried the command on the same overcloud > deployment which was failing earlier(when docker-ha.yaml was not passed). > If yes, to rule out if the issue is caused by an already deployed > environment can delete the overcloud and then redeploy with correct > environment files as used in the last run. > > One reason for the password expiration that i found could be the Time is > not in Sync on the overcloud nodes. So it would be good to check that as > well and fix(by using correct NTP sources) before attempting redeployment. > > Thanks and regards > Yatin Karel > > > > On Tue, Dec 28, 2021 at 2:03 PM Anirudh Gupta wrote: > >> Hi Yatin & Team >> >> Thanks for your response. >> >> When I executed the command as below, the installation moved ahead and >> encountered another error. >> >> openstack overcloud deploy --templates \ >> -r /home/stack/templates/roles_data.yaml \ >> -e /home/stack/templates/node-info.yaml \ >> -e environment.yaml \ >> -e >> /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml \ >> -e >> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >> -e /home/stack/containers-prepare-parameter.yaml >> >> *Issue:* >> The error was: keystoneauth1.exceptions.http.Unauthorized: *The password >> is expired and needs to be changed for user*: >> 4f7d1dbf58574e64af9e359cb98ccbbc. (HTTP 401) (Request-ID: >> req-b29aa655-e3ec-4d4b-8ada-397f9a132582) >> >> I am attaching the ansible.logs for your reference. It would be a great >> help if you could suggest some pointers to resolve this issue. >> >> Regards >> Anirudh Gupta >> >> On Tue, Dec 28, 2021 at 11:13 AM Yatin Karel wrote: >> >>> Hi Anirudh, >>> >>> As said order is important here, docker-ha.yaml should be followed by >>> podman.yaml, the parameters in environment files override the parameters >>> from previous environment files passed and that would make deployment to >>> use podman instead of docker. Name of the parameter to which makes this >>> switch is "ContainerCli". >>> >>> >>> Thanks and regards >>> Yatin Karel >>> >>> On Tue, Dec 28, 2021 at 10:59 AM Anirudh Gupta >>> wrote: >>> >>>> If this is a docker-ha issue, then that has also been tried. >>>> >>>> Since this is Centos 8, there is no docker available. If I pass the >>>> docker-ha.yml, then it gives the following error >>>> >>>> FATAL | Pull >>>> undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo >>>> image | overcloud-controller-1 | error={"changed": true, "cmd": "docker >>>> pull >>>> undercloud.ctlplane.localdomain:8787/tripleotraincentos8/centos-binary-cinder-volume:current-tripleo", >>>> "delta": "0:00:00.005932", "end": "2021-12-27 12:42:33.927484", "msg": >>>> "non-zero return code", "rc": 127, "start": "2021-12-27 12:42:33.921552", >>>> "stderr": "/bin/sh: docker: command not found", "*stderr_lines": >>>> ["/bin/sh: docker: command not found"], "stdout": "", "stdout_lines": []}* >>>> >>>> Regards >>>> Anirudh Gupta >>>> >>>> On Tue, Dec 28, 2021 at 10:26 AM Yatin Karel wrote: >>>> >>>>> Hi Anirudh, >>>>> >>>>> Sorry which timer? Timer adjustment is not needed for the issue you >>>>> are seeing, if you mean overcloud deploy timeout then overcloud deploy >>>>> provides the option to do so using --timeout option. The best option for >>>>> now is to try docker-ha and podman in order as suggested earlier. >>>>> >>>>> >>>>> Thanks and Regards >>>>> Yatin Karel >>>>> >>>>> On Tue, Dec 28, 2021 at 10:12 AM Anirudh Gupta >>>>> wrote: >>>>> >>>>>> Thanks Yatin for your response. >>>>>> >>>>>> Please suggest how can this timer be increased or any other steps >>>>>> that needs to be followed to rectify this? >>>>>> >>>>>> Regards >>>>>> Anirudh Gupta >>>>>> >>>>>> On Tue, Dec 28, 2021 at 10:08 AM Yatin Karel >>>>>> wrote: >>>>>> >>>>>>> Hi Anirudh, >>>>>>> >>>>>>> >>>>>>> On Mon, Dec 27, 2021 at 9:39 PM Anirudh Gupta >>>>>>> wrote: >>>>>>> > >>>>>>> > Hi Team, >>>>>>> > >>>>>>> > I am trying to deploy TripleO Train with 3 controller and 1 >>>>>>> Compute. >>>>>>> > For overcloud images, I have a registry server at undercloud only. >>>>>>> > >>>>>>> > I executed the following command to deploy overcloud >>>>>>> > >>>>>>> > openstack overcloud deploy --templates \ >>>>>>> > -r /home/stack/templates/roles_data.yaml \ >>>>>>> > -e /home/stack/templates/node-info.yaml \ >>>>>>> > -e >>>>>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >>>>>>> > -e /home/stack/containers-prepare-parameter.yaml >>>>>>> > >>>>>>> > The command ran for around 1.5 hrs and initially stack got >>>>>>> successfully created and after that for 45 mins, ansible tasks were getting >>>>>>> executed. It then gave following error in overcloud-controller-0 >>>>>>> > >>>>>>> > 2021-12-27 11:12:27,507 p=181 u=mistral n=ansible | 2021-12-27 >>>>>>> 11:12:27.506838 | 525400b1-b522-2a06-ea9d-00000000356f | OK | Debug >>>>>>> output for task: Start containers for step 2 | overcloud-novacompute-0 | >>>>>>> result={ >>>>>>> > "changed": false, >>>>>>> > "failed_when_result": false, >>>>>>> > "start_containers_outputs.stdout_lines | default([]) | >>>>>>> union(start_containers_outputs.stderr_lines | default([]))": [ >>>>>>> > >>>>>>> "f206c31a781641313aa4a0499c62475efc335de6faea785cd4e855dc32ebb571", >>>>>>> > "", >>>>>>> > "Info: Loading facts", >>>>>>> > "Notice: Compiled catalog for >>>>>>> overcloud-novacompute-0.localdomain in environment production in 0.05 >>>>>>> seconds", >>>>>>> > "Info: Applying configuration version '1640604309'", >>>>>>> > "Notice: >>>>>>> /Stage[main]/Tripleo::Profile::Base::Neutron::Ovn_metadata_agent_wrappers/Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]/File[/var/lib/neutron/ovn_metadata_haproxy_wrapper]/ensure: >>>>>>> defined content as '{md5}5bb050ca70c01981975efad9d8f81f2d'", >>>>>>> > "Info: >>>>>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]: >>>>>>> Unscheduling all events on >>>>>>> Tripleo::Profile::Base::Neutron::Wrappers::Haproxy[ovn_metadata_haproxy_process_wrapper]", >>>>>>> > "Info: Creating state file >>>>>>> /var/lib/puppet/state/state.yaml", >>>>>>> > "Notice: Applied catalog in 0.01 seconds", >>>>>>> > "Changes:", >>>>>>> > " Total: 1", >>>>>>> > "Events:", >>>>>>> > " Success: 1", >>>>>>> > "Resources:", >>>>>>> > " Changed: 1", >>>>>>> > " Out of sync: 1", >>>>>>> > " Skipped: 7", >>>>>>> > " Total: 8", >>>>>>> > "Time:", >>>>>>> > " File: 0.00", >>>>>>> > " Transaction evaluation: 0.01", >>>>>>> > " Catalog application: 0.01", >>>>>>> > " Config retrieval: 0.09", >>>>>>> > " Last run: 1640604309", >>>>>>> > " Total: 0.01", >>>>>>> > "Version:", >>>>>>> > " Config: 1640604309", >>>>>>> > " Puppet: 5.5.10", >>>>>>> > "Error executing ['podman', 'container', 'exists', >>>>>>> 'nova_compute_init_log']: returned 1", >>>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>>> '--filter', 'label=container_name=nova_compute_init_log', '--filter', >>>>>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>>>>> without config_id", >>>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>>> '--filter', 'label=container_name=nova_compute_init_log', '--format', >>>>>>> '{{.Names}}']\"", >>>>>>> > "Error executing ['podman', 'container', 'exists', >>>>>>> 'create_haproxy_wrapper']: returned 1", >>>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--filter', >>>>>>> 'label=config_id=tripleo_step2', '--format', '{{.Names}}']\" - retrying >>>>>>> without config_id", >>>>>>> > "Did not find container with \"['podman', 'ps', '-a', >>>>>>> '--filter', 'label=container_name=create_haproxy_wrapper', '--format', >>>>>>> '{{.Names}}']\"" >>>>>>> > ] >>>>>>> > } >>>>>>> >>>>>>> This is not the actual error, actual error is: puppet-user: Error: >>>>>>> /Stage[main]/Tripleo::Profile::Base::Rabbitmq/Rabbitmq_policy[ha-all@/]: >>>>>>> Could not evaluate: Command is still failing after 180 seconds expired!" >>>>>>> >>>>>>> > >>>>>>> > I am also attaching ansible.log file for more information. >>>>>>> > >>>>>>> > Note: On Centos 8, there is no docker, so I didn't pass >>>>>>> docker-ha.yml >>>>>>> For enabling HA and with podman in Train on CentOS8, you need to >>>>>>> pass both docker-ha.yaml and podman.yaml in order(*order is >>>>>>> important here*, so -e >>>>>>> /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml -e >>>>>>> /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml), this >>>>>>> way you will have deployment with HA and podman, i agree docker-ha name is >>>>>>> confusing here with podman but that has to be passed here to get the >>>>>>> required deployment. Also with Ussuri+ HA is turned on by default so those >>>>>>> releases may work even without passing docker-ha.yaml but for Train at >>>>>>> least it's needed. >>>>>>> > >>>>>>> > Can someone please help in resolving my issue >>>>>>> > >>>>>>> As per your requirement I would suggest running with the above >>>>>>> config. >>>>>>> >>>>>>> > Regards >>>>>>> > Anirudh Gupta >>>>>>> >>>>>>> Thanks and Regards >>>>>>> Yatin Karel >>>>>>> >>>>>> -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: ansible.log Type: application/octet-stream Size: 2312730 bytes Desc: not available URL: From hanguangyu2 at gmail.com Wed Dec 29 09:23:52 2021 From: hanguangyu2 at gmail.com (=?UTF-8?B?6Z+p5YWJ5a6H?=) Date: Wed, 29 Dec 2021 17:23:52 +0800 Subject: [neutron] Do I need to replace NetworkManager with network.service in centos8? Message-ID: Dear all, I am deploing OpenStack Victoria. I want to ask a question. Do I need to replace NetworkManager with network-scripts in centos8? I saw that "OpenStack Networking does not work on systems that have the Network Manager service enabled" in past redhat documentation and some other places. But I don't see this in latest documentation of OpenStack community? https://docs.openstack.org/install-guide/environment-packages-rdo.html?. And If NetworkManager is confit will OpenStack network, Do anyone know where I can see the reason of confit. I would appreciate any kind of guidance. Thanks, Han Guangyu. Past redhat documentation: https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/11/html/manual_installation_procedures/chap-prerequisites#Disabling_Network_Manager -------------- next part -------------- An HTML attachment was scrubbed... URL: From hanguangyu2 at gmail.com Wed Dec 29 09:31:32 2021 From: hanguangyu2 at gmail.com (=?UTF-8?B?6Z+p5YWJ5a6H?=) Date: Wed, 29 Dec 2021 17:31:32 +0800 Subject: [Deploy] Can three all-in-one nodes form a OpenStack cluster? Message-ID: Dear all, I am deploing OpenStack Victoria. Can I make a cluster with three all-in-one nodes? Will this have any problems or points to be aware of? Sorry to bother, Thank you very much. Han Guangyu. From radoslaw.piliszek at gmail.com Wed Dec 29 13:33:50 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Wed, 29 Dec 2021 14:33:50 +0100 Subject: [Deploy] Can three all-in-one nodes form a OpenStack cluster? In-Reply-To: References: Message-ID: Dear Han, This is possible and might be sensible depending on your use case. What deployment method are you using? Kind regards, -yoctozepto On Wed, 29 Dec 2021 at 10:32, ??? wrote: > > Dear all, > > I am deploing OpenStack Victoria. Can I make a cluster with three > all-in-one nodes? > > Will this have any problems or points to be aware of? > > Sorry to bother, Thank you very much. > > Han Guangyu. > From hanguangyu2 at gmail.com Wed Dec 29 13:46:27 2021 From: hanguangyu2 at gmail.com (=?UTF-8?B?6Z+p5YWJ5a6H?=) Date: Wed, 29 Dec 2021 21:46:27 +0800 Subject: [Deploy] Can three all-in-one nodes form a OpenStack cluster? In-Reply-To: References: Message-ID: Dear Radoslaw, Thank you for your reply. I'm manually building OpenStack Victoria in centos8 for learning cluster and high-availability. I refer to https://docs.openstack.org/ha-guide/ and some blogs. If you have some advices or you know some good document, you can tell me. I will very glad. Sincerely, Han Guangyu Rados?aw Piliszek ?2021?12?29??? 21:34??? > > Dear Han, > > This is possible and might be sensible depending on your use case. > > What deployment method are you using? > > Kind regards, > -yoctozepto > > On Wed, 29 Dec 2021 at 10:32, ??? wrote: > > > > Dear all, > > > > I am deploing OpenStack Victoria. Can I make a cluster with three > > all-in-one nodes? > > > > Will this have any problems or points to be aware of? > > > > Sorry to bother, Thank you very much. > > > > Han Guangyu. > > From radoslaw.piliszek at gmail.com Wed Dec 29 14:33:11 2021 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Wed, 29 Dec 2021 15:33:11 +0100 Subject: [Deploy] Can three all-in-one nodes form a OpenStack cluster? In-Reply-To: References: Message-ID: Dear Han, My answer will be opinionated but I really like the solution we provide with Kolla Ansible. It also supports Victoria still. [1] Of course, it might not be the best for learning "just OpenStack" as it includes a large portion of Docker and Ansible but I believe it is simple enough that you might be positively inspired. [1] https://docs.openstack.org/kolla-ansible/victoria/ Kind regards, -yoctozepto On Wed, 29 Dec 2021 at 14:46, ??? wrote: > > Dear Radoslaw, > > Thank you for your reply. > > I'm manually building OpenStack Victoria in centos8 for learning > cluster and high-availability. I refer to > https://docs.openstack.org/ha-guide/ and some blogs. > > If you have some advices or you know some good document, you can tell > me. I will very glad. > > Sincerely, > Han Guangyu > > Rados?aw Piliszek ?2021?12?29??? 21:34??? > > > > Dear Han, > > > > This is possible and might be sensible depending on your use case. > > > > What deployment method are you using? > > > > Kind regards, > > -yoctozepto > > > > On Wed, 29 Dec 2021 at 10:32, ??? wrote: > > > > > > Dear all, > > > > > > I am deploing OpenStack Victoria. Can I make a cluster with three > > > all-in-one nodes? > > > > > > Will this have any problems or points to be aware of? > > > > > > Sorry to bother, Thank you very much. > > > > > > Han Guangyu. > > > From senrique at redhat.com Wed Dec 29 15:03:37 2021 From: senrique at redhat.com (Sofia Enriquez) Date: Wed, 29 Dec 2021 12:03:37 -0300 Subject: [cinder] Bug deputy report for week of 12-29-2021 Message-ID: No meeting today. This is a bug report from 12-22-2021 to 12-29-2021. Agenda: https://etherpad.opendev.org/p/cinder-bug-squad-meeting ----------------------------------------------------------------------------------------- Medium - https://bugs.launchpad.net/cinder/+bug/1955670 "Unable to delete a group volume after transfer". Unassigned. Cheers, -- Sof?a Enriquez she/her Software Engineer Red Hat PnT IRC: @enriquetaso @RedHat Red Hat Red Hat -------------- next part -------------- An HTML attachment was scrubbed... URL: From tjoen at dds.nl Wed Dec 29 17:37:57 2021 From: tjoen at dds.nl (tjoen) Date: Wed, 29 Dec 2021 18:37:57 +0100 Subject: [neutron] Do I need to replace NetworkManager with network.service in centos8? In-Reply-To: References: Message-ID: On 12/29/21 10:23, ??? wrote: ... > I saw that "OpenStack Networking does not work on systems that have the > Network Manager service enabled" in past redhat documentation and some > other places. But I don't see this in latest documentation of OpenStack > community? ... > Past redhat documentation: > https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/11/html/manual_installation_procedures/chap-prerequisites#Disabling_Network_Manager I'll test it if Yoda is released From ozzzo at yahoo.com Wed Dec 29 22:41:41 2021 From: ozzzo at yahoo.com (Albert Braden) Date: Wed, 29 Dec 2021 22:41:41 +0000 (UTC) Subject: PTR records in Designate References: <1749759069.65668.1640817701283.ref@mail.yahoo.com> Message-ID: <1749759069.65668.1640817701283@mail.yahoo.com> We have hacked Designate to add PTR records and it is a mess. Every upgrade turns into a bug hunt. I'd like to advocate for going back to vanilla kolla-ansible but then we wouldn't have PTR records. How are other Designate users dealing with the PTR issue? From frickler at offenerstapel.de Thu Dec 30 09:28:33 2021 From: frickler at offenerstapel.de (Jens Harbott (frickler)) Date: Thu, 30 Dec 2021 10:28:33 +0100 Subject: PTR records in Designate In-Reply-To: <1749759069.65668.1640817701283@mail.yahoo.com> References: <1749759069.65668.1640817701283.ref@mail.yahoo.com> <1749759069.65668.1640817701283@mail.yahoo.com> Message-ID: On Wed, Dec 29, 2021 at 10:41:41PM +0000, Albert Braden wrote: > We have hacked Designate to add PTR records and it is a mess. Every upgrade turns into a bug hunt. I'd like to advocate for going back to vanilla kolla-ansible but then we wouldn't have PTR records. How are other Designate users dealing with the PTR issue? What exactly did you hack in Designate? For me it seems that using the neutron integration is doing just fine. Yours Jens -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 833 bytes Desc: not available URL: From ozzzo at yahoo.com Thu Dec 30 13:13:58 2021 From: ozzzo at yahoo.com (Albert Braden) Date: Thu, 30 Dec 2021 13:13:58 +0000 (UTC) Subject: PTR records in Designate In-Reply-To: References: <1749759069.65668.1640817701283.ref@mail.yahoo.com> <1749759069.65668.1640817701283@mail.yahoo.com> Message-ID: <1643952194.176605.1640870038796@mail.yahoo.com> It was done years ago before I worked here. What I'm hearing is that Designate didn't do PTR records, so we wrote custom code to add that feature. Were PTR records added at some point? We're running kolla-ansible and recently upgraded from Queens to Train. On Thursday, December 30, 2021, 05:30:41 AM EST, Jens Harbott (frickler) wrote: On Wed, Dec 29, 2021 at 10:41:41PM +0000, Albert Braden wrote: > We have hacked Designate to add PTR records and it is a mess. Every upgrade turns into a bug hunt. I'd like to advocate for going back to vanilla kolla-ansible but then we wouldn't have PTR records. How are other Designate users dealing with the PTR issue? What exactly did you hack in Designate? For me it seems that using the neutron integration is doing just fine. Yours Jens -------------- next part -------------- An HTML attachment was scrubbed... URL: From rosmaita.fossdev at gmail.com Fri Dec 31 15:13:28 2021 From: rosmaita.fossdev at gmail.com (Brian Rosmaita) Date: Fri, 31 Dec 2021 10:13:28 -0500 Subject: [cinder] proposing Simon Dodsley for cinder-specs-core In-Reply-To: <65562de2-0254-e231-d7e1-c9435cfac93b@gmail.com> References: <65562de2-0254-e231-d7e1-c9435cfac93b@gmail.com> Message-ID: On 12/22/21 10:42 PM, Brian Rosmaita wrote: > As everyone is painfully aware, we have review bandwidth issues in the > Cinder project.? I propose to address this issue with respect to cinder > specs, at least, by adding a contributor who is not currently a > cinder-core as a member of the cinder-specs-core team.? (Right now, the > only members of cinder-specs-core are the cinder-core team.) > > Simon Dodsley (simondodsley on IRC) has been a cinder contributor for a > long time, and has been especially active in PTGs and midcycle meetings > over the past few cycles.? He's done quality reviews, especially of > specs, this cycle.? As a vendor representative, he brings a useful > perspective to reviews that affect drivers, as well as general > deployment and usability issues.? Additionally, given that the current > core team has reviewing responsibilities spread over multiple > repositories, having a person focused primarily on specs should help us > to get more timely feedback for specs proposers. > > Finally, don't forget that it is thanks to Simon that the Cinder mascot > finally has a name ("Argo"), which in turn has given the Cinder team the > really cool name "The Argonauts". > > In the absence of objections, I'll add Simon to the cinder-specs-core > team shortly after 1500 UTC on Friday 31 December.? Please communicate > any concerns to me before that time. Having heard only positive responses, I've added Simon to the cinder-specs-core team, with all the privileges and responsibilities pertaining thereto. Congratulations, Simon! > > cheers, > brian From elod.illes at est.tech Fri Dec 31 16:46:29 2021 From: elod.illes at est.tech (=?UTF-8?B?RWzFkWQgSWxsw6lz?=) Date: Fri, 31 Dec 2021 17:46:29 +0100 Subject: [release] Release countdown for week R-12, Jan 03-07 Message-ID: <7952e2d6-c592-814a-883c-945b8787fc30@est.tech> Development Focus ----------------- Happy new year! The Yoga-2 milestone is next week, on January6th, 2022. Yoga-related specs should now be finalized so that teams can move to implementation ASAP. Some teams observe specific deadlines on the second milestone (mostly spec freezes): please refer to https://releases.openstack.org/ yoga /schedule.html for details. General Information ------------------- Libraries need to be released at least once per milestone period. Next week, the release team will propose releases for any library that has not been otherwise released since milestone 1. PTL's and release liaisons, please watch for these and give a +1 to acknowledge them. If there is some reason to hold off on a release, let us know that as well. A +1 would be appreciated, but if we do not hear anything at all by the end of the week, we will assume things are OK to proceed. Remember that non-library deliverables that follow the cycle-with-intermediary release model should have an intermediary release before milestone-2. Those who haven't will be proposed to switch to the cycle-with-rc model, which is more suited to deliverables that are released only once per cycle. Next week is also the deadline to freeze the contents of the final release. All new 'Yoga' deliverables need to have a deliverable file in https://opendev.org/openstack/releases/src/branch/master/deliverables and need to have done a release by milestone-2. Changes proposing missingdeliverables for inclusion in Yogahave been posted, so if you have one in your project team please update them with an actual release request before the milestone-2 deadline if you plan on including that deliverable in Yoga, or -1 if you need one more cycle to be ready. Upcoming Deadlines & Dates -------------------------- Yoga-2 Milestone:January6th, 2022 El?d Ill?s irc: elodilles -------------- next part -------------- An HTML attachment was scrubbed... URL: