From gmann at ghanshyammann.com Tue Mar 1 00:22:51 2022 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Mon, 28 Feb 2022 18:22:51 -0600 Subject: [all][tc] Technical Committee next weekly meeting on Mar 3rd (not 4th) at 1500 UTC In-Reply-To: <17f413dc13b.1112c1076280164.9078145143439661308@ghanshyammann.com> References: <17f413dc13b.1112c1076280164.9078145143439661308@ghanshyammann.com> Message-ID: <17f42dbbe35.e38a85b0295421.3903320940743592043@ghanshyammann.com> Sorry I mentioned Mar 4th as the next meeting but it is on Mar 3rd Thursday 15:00 UTC -gmann ---- On Mon, 28 Feb 2022 10:50:40 -0600 Ghanshyam Mann wrote ---- > Hello Everyone, > > Technical Committee's next weekly meeting is scheduled for Mar 4th at 1500 UTC. > > If you would like to add topics for discussion, please add them to the below wiki page by > Wednesday, Mar 3rd, at 2100 UTC. > > https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting > > -gmann > > From johnsomor at gmail.com Tue Mar 1 00:38:53 2022 From: johnsomor at gmail.com (Michael Johnson) Date: Mon, 28 Feb 2022 16:38:53 -0800 Subject: [Neutron][Octavia][ovn-octavia-provider] proposing Luis Tomas Bolivar for ovn-octavia-provider core reviewer In-Reply-To: <96c49d33-fc30-2380-fe82-8247fffc264f@gmail.com> References: <96c49d33-fc30-2380-fe82-8247fffc264f@gmail.com> Message-ID: +1 from me Michael On Mon, Feb 28, 2022 at 2:19 PM Brian Haley wrote: > > +1 from me, Luis has been great at finding and fixing issues in this > repo for a while. > > -Brian > > On 2/28/22 09:38, Lajos Katona wrote: > > Hi > > > > I would like to propose Luis Tomas Bolivar (ltomasbo) as a core reviewer > > to the ovn-octavia-provider project. > > Luis is very active in the project (see [1] and [2]) and has experience > > with Kuryr side also which uses ovn-octavia-provider. > > > > As ovn-octavia-provider is a link between Neutron and Octavia I ask both > > Neutron and Octavia cores to vote by answering to this thread, to have a > > final decision. > > Thanks for your consideration > > > > [1]: > > https://review.opendev.org/q/owner:ltomasbo%2540redhat.com+branch:master > > > > [2]: > > https://www.stackalytics.io/report/contribution?module=neutron-group&project_type=openstack&days=60 > > > > > > Cheers > > Lajos > From laurentfdumont at gmail.com Tue Mar 1 01:54:32 2022 From: laurentfdumont at gmail.com (Laurent Dumont) Date: Mon, 28 Feb 2022 20:54:32 -0500 Subject: Fluentd error with Kolla ansible deployment In-Reply-To: References: Message-ID: Not super familiar with the kolla docker part, but the code seems pretty straightforward. What is the IP of your registry? I dont think you should use localhost as the tasks are running on the computes/controllers? Can you use the registry IP in the same subnet/reachable subnet? Tracing the code - name: Ensure fluentd image is present for label check vars: service_name: "fluentd" service: "{{ common_services[service_name] }}" become: true kolla_docker: action: "ensure_image" common_options: "{{ docker_common_options }}" image: "{{ service.image }}" when: - fluentd_version is not defined or fluentd_binary is not defined - enable_fluentd | bool This calls def ensure_image(self): if not self.check_image(): self.pull_image() This then calls (if the image is not there) def pull_image(self): if self.params.get('auth_username'): self.dc.login( username=self.params.get('auth_username'), password=self.params.get('auth_password'), registry=self.params.get('auth_registry'), email=self.params.get('auth_email') ) image, tag = self.parse_image() old_image_id = self.get_image_id() statuses = [ json.loads(line.strip().decode('utf-8')) for line in self.dc.pull( repository=image, tag=tag, stream=True ) ] for status in reversed(statuses): if 'error' in status: if status['error'].endswith('not found'): self.module.fail_json( msg="The requested image does not exist: {}:{}".format( image, tag), failed=True ) else: self.module.fail_json( msg="Unknown error message: {}".format( status['error']), failed=True ) new_image_id = self.get_image_id() self.changed = old_image_id != new_image_id On Mon, Feb 28, 2022 at 6:26 PM Laurent Dumont wrote: > Can you pull from that registry manually? If that doesn't work, Kolla > won't work either. > > On Mon, Feb 28, 2022 at 6:47 AM A Monster wrote: > >> I'm getting the following error while trying to deploy openstack using >> kolla-ansible. >> I've set up a local docker registry and enabled insecure registries in >> daemon.json file and I even tried to pull fluentd docker image but I still >> get the following error >> >> TASK [common : Ensure fluentd image is present for label check] >> fatal: [localhost]: FAILED! => {"changed": true, "msg": "'Traceback (most >> recent call last):\\n File >> \"/usr/local/lib/python3.6/site-packages/docker/api/client.py\", line 268, >> in _raise_for_status\\n response.raise_for_status()\\n File >> \"/usr/lib/python3.6/site-packages/requests/models.py\", line 940, in >> raise_for_status\\n raise HTTPError(http_error_msg, >> response=self)\\nrequests.exceptions.HTTPError: 500 Server Error: Internal >> Server Error for url: >> http+docker://localhost/v1.41/images/create?tag=xena&fromImage=quay.io%2Fkolla%2Fcentos-source-fluentd\\n\\nDuring >> handling of the above exception, another exception occurred:\\n\\nTraceback >> (most recent call last):\\n File >> \"/tmp/ansible_kolla_docker_payload_xizndx9s/ansible_kolla_docker_payload.zip/ansible/modules/kolla_docker.py\", >> line 1241, in main\\n File >> \"/tmp/ansible_kolla_docker_payload_xizndx9s/ansible_kolla_docker_payload.zip/ansible/modules/kolla_docker.py\", >> line 1116, in ensure_image\\n File >> \"/tmp/ansible_kolla_docker_payload_xizndx9s/ansible_kolla_docker_payload.zip/ansible/modules/kolla_docker.py\", >> line 692, in pull_image\\n File >> \"/usr/local/lib/python3.6/site-packages/docker/api/image.py\", line 430, >> in pull\\n self._raise_for_status(response)\\n File >> \"/usr/local/lib/python3.6/site-packages/docker/api/client.py\", line 270, >> in _raise_for_status\\n raise create_api_error_from_http_exception(e)\\n >> File \"/usr/local/lib/python3.6/site-packages/docker/errors.py\", line 31, >> in create_api_error_from_http_exception\\n raise cls(e, >> response=response, explanation=explanation)\\ndocker.errors.APIError: 500 >> Server Error for >> http+docker://localhost/v1.41/images/create?tag=xena&fromImage=quay.io%2Fkolla%2Fcentos-source-fluentd: >> Internal Server Error (\"unauthorized: access to the requested resource is >> not authorized\")\\n'"} >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From tkajinam at redhat.com Tue Mar 1 02:11:56 2022 From: tkajinam at redhat.com (Takashi Kajinami) Date: Tue, 1 Mar 2022 11:11:56 +0900 Subject: [puppet] Gate blocker: CentOS 9 integration jobs are broken Message-ID: Hi, Please be aware that currently CentOS 9 integration jobs are broken because of the two problems caused by recent updates in c9s repos. https://bugs.launchpad.net/puppet-openstack-integration/+bug/1962506 https://bugs.launchpad.net/puppet-openstack-integration/+bug/1962507 I'm currently working on implementing workaround[1], but in case it would take some time, I'll make c9s integration jobs non-voting again to unblock our gate. [1] https://review.opendev.org/c/openstack/puppet-openstack-integration/+/831197 Please avoid approving changes which require integration jobs. Thank you, Takashi -------------- next part -------------- An HTML attachment was scrubbed... URL: From gagehugo at gmail.com Tue Mar 1 04:05:57 2022 From: gagehugo at gmail.com (Gage Hugo) Date: Mon, 28 Feb 2022 22:05:57 -0600 Subject: [openstack-helm] No Meeting Tomorrow Message-ID: Hey team, Since there's nothing on the agenda for tomorrow's meeting, it has been cancelled. We will meet again next week. -------------- next part -------------- An HTML attachment was scrubbed... URL: From johnsomor at gmail.com Tue Mar 1 04:48:51 2022 From: johnsomor at gmail.com (Michael Johnson) Date: Mon, 28 Feb 2022 20:48:51 -0800 Subject: [nova][octavia][masakari][designate][oslo][requirements] oslo.context>=4.0.0 seems to need updated tests In-Reply-To: References: <20220228165609.zzz7ur2mktfrqmue@mthode.org> <60238d97d4f9d44413441ec75e0b893e47bb9f6f.camel@redhat.com> Message-ID: Yep, thanks Takashi. Patches for both Octavia and Designate have now merged. I fired off a recheck to confirm. Michael On Mon, Feb 28, 2022 at 3:28 PM Takashi Kajinami wrote: > > I hope the issue with Octavia was resolved by https://review.opendev.org/c/openstack/octavia/+/801860 . > (Thank you Michael for merging this !) > > AFAIK Mistral is still dependent on the old arguments and we need > https://review.opendev.org/c/openstack/mistral/+/801853 > to fix it. It'd be nice if somebody from Mistral team can merge this > (and the series of clean up patches) > > On Tue, Mar 1, 2022 at 3:47 AM Sean Mooney wrote: >> >> On Mon, 2022-02-28 at 18:37 +0000, Sean Mooney wrote: >> > On Mon, 2022-02-28 at 10:09 -0800, Michael Johnson wrote: >> > > This is a legit API change in oslo.context (Thus the major version >> > > bump). It's been a long time coming too. >> > we are currently pass m3 in the rc period. >> > i dont know if we can do a breaking change like this in yoga at this point depending on how invaise it is. >> > we might need to hold this until after RC1 is out ideally this type of change would happen very erarly in the cycle before the ptg or m1 >> > not at the very end on the nova side there is only one failing unit test >> > https://6b96a36d22f450141095-9e4eb9f05513870f0e5456f29474f6b2.ssl.cf5.rackcdn.com/829599/1/check/cross-nova-py38/60113ea/testr_results.html >> > but i need to also run the func tests to see if there is other impact >> > tempest-full-py3 passed so nova iteslf is likely fine. >> > >> > ill try running it locally and see if that is the only impact or not. >> >> gmann beat me too it. >> https://review.opendev.org/c/openstack/nova/+/831244 >> so ya it is apprently just the one unit test so i dont think this should be an issue for nova >> >> > >> > > >> > > I will get started on the required code changes to Designate. Maybe >> > > Octavia as well if someone else doesn't get to it before I do. >> > > >> > > Michael >> > > >> > > On Mon, Feb 28, 2022 at 9:04 AM Matthew Thode wrote: >> > > > >> > > > The update to see what's breaking is here >> > > > https://review.opendev.org/829599 >> > > > >> > > > Feel free to make that patch depend on your fix to test (with rechecks) >> > > > >> > > > -- >> > > > Matthew Thode >> > > >> > >> >> From lucasagomes at gmail.com Tue Mar 1 08:48:40 2022 From: lucasagomes at gmail.com (Lucas Alvares Gomes) Date: Tue, 1 Mar 2022 08:48:40 +0000 Subject: [Neutron][Octavia][ovn-octavia-provider] proposing Luis Tomas Bolivar for ovn-octavia-provider core reviewer In-Reply-To: References: Message-ID: +1 from me! On Mon, Feb 28, 2022 at 2:42 PM Lajos Katona wrote: > > Hi > > I would like to propose Luis Tomas Bolivar (ltomasbo) as a core reviewer to the ovn-octavia-provider project. > Luis is very active in the project (see [1] and [2]) and has experience with Kuryr side also which uses ovn-octavia-provider. > > As ovn-octavia-provider is a link between Neutron and Octavia I ask both Neutron and Octavia cores to vote by answering to this thread, to have a final decision. > Thanks for your consideration > > [1]: https://review.opendev.org/q/owner:ltomasbo%2540redhat.com+branch:master > [2]: https://www.stackalytics.io/report/contribution?module=neutron-group&project_type=openstack&days=60 > > Cheers > Lajos From skaplons at redhat.com Tue Mar 1 09:53:33 2022 From: skaplons at redhat.com (Slawek Kaplonski) Date: Tue, 01 Mar 2022 10:53:33 +0100 Subject: [neutron] CI meeting - agenda for 1.03.2022 Message-ID: <2176895.iZASKD2KPV@p1> Hi, Agenda for today's Neutron CI meeting is at [1]. Please remember that we will meet on the video call today [2]. See You all there at 1500 UTC [1] https://etherpad.opendev.org/p/neutron-ci-meetings [2] https://meetpad.opendev.org/neutron-ci-meetings -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From tkajinam at redhat.com Tue Mar 1 13:47:41 2022 From: tkajinam at redhat.com (Takashi Kajinami) Date: Tue, 1 Mar 2022 22:47:41 +0900 Subject: [puppet] Gate blocker: CentOS 9 integration jobs are broken In-Reply-To: References: Message-ID: Unfortunately I've not yet managed to implement a workaround and it's likely I need some more time to fix it, I'm merging the patch to make c9s jobs non-voting. https://review.opendev.org/c/openstack/puppet-openstack-integration/+/831183 I'll revert that change once we can implement workaround/solution. On Tue, Mar 1, 2022 at 11:11 AM Takashi Kajinami wrote: > Hi, > > Please be aware that currently CentOS 9 integration jobs are broken > because of the two problems caused by recent updates in c9s repos. > https://bugs.launchpad.net/puppet-openstack-integration/+bug/1962506 > https://bugs.launchpad.net/puppet-openstack-integration/+bug/1962507 > > I'm currently working on implementing workaround[1], but in case it would > take > some time, I'll make c9s integration jobs non-voting again to unblock our > gate. > [1] > https://review.opendev.org/c/openstack/puppet-openstack-integration/+/831197 > > Please avoid approving changes which require integration jobs. > > Thank you, > Takashi > -------------- next part -------------- An HTML attachment was scrubbed... URL: From tkajinam at redhat.com Tue Mar 1 13:58:40 2022 From: tkajinam at redhat.com (Takashi Kajinami) Date: Tue, 1 Mar 2022 22:58:40 +0900 Subject: [puppet][tripleo][neutron][ops] Deprecating support for networking-bigswitch Message-ID: Hi, Today I noticed networking-bigswitch has been unmaintained recently. The repos have not been updated for 2 years and no release has been made since stable/train. https://opendev.org/x/networking-bigswitch Based on the above observation, I assume no users would be using the plugin with recent OpenStack releases and I've proposed deprecating support for the plugin in puppet-neutron. https://review.opendev.org/c/openstack/puppet-neutron/+/831376 In case if anyone has concern with this, please post your thoughts in the above review. Once this deprecation is approved/merged, I'll remove support for the plugin from TripleO because implementation in TripleO depends on puppet-neutron. Thank you, Takashi -------------- next part -------------- An HTML attachment was scrubbed... URL: From juliaashleykreger at gmail.com Tue Mar 1 14:06:01 2022 From: juliaashleykreger at gmail.com (Julia Kreger) Date: Tue, 1 Mar 2022 06:06:01 -0800 Subject: [Ironic] No suitable device was found for deployment In-Reply-To: <3209d9fc-8db1-6dca-5736-402c6f082fcc@cern.ch> References: <04af20bb-143b-9b56-6787-d5ac6ab0d38c@cern.ch> <3209d9fc-8db1-6dca-5736-402c6f082fcc@cern.ch> Message-ID: On Mon, Feb 28, 2022 at 1:12 AM Arne Wiebalck wrote: > > Hi Guangyu, > > I am not aware of anything in the Ironic Python Agent that > would remove disks from the system in a way that they would > not be visible after a reboot (apart from, as mentioned before, > the clean up of a hardware RAID in a way the IPA is not able > to see any devices after). > > How about trying to access and configure the hardware RAID with > the corresponding tool from the RAM disk you booted into from the > USB stick? Install the tool and see if it detects the controller. > > The very first step before doing anything with Ironic is to > get the disks back or understand why they are not visible. > Did cleaning fail at any given point with these machines? If you have physical access, try disconnecting all of the drives, and then powering up the machine and see if you can get into the firmware configuration screen with control-h. If you can, remove all of the prior configuration or disk volumes. They will look like they are in error states most likely. If your unable to get into this screen, I would be worried about your disk controller card. If your able to clear everything out of the controller, power off, try re-inserting drives, and see what happens. See if the controller can view/interact with the drives. If it sees no drives, then my next paragraph is likely the case. The disks sound like they might be in security locked state which will likely require a desktop SATA disk controller to remedy by attaching and manually removing from a security locked state. Megaraid controllers can't recognize security locked devices (most controllers and especially ones labeled "raid controllers" can't handle it) when in pass-through mode, but I've never heard of security lock commands actually getting through to the device with those controllers in pass-through mode. If the card was in raid mode to begin with, then it likely never did anything involving secure erase as the controller should not be offering that as a feature of provided disks to the OS. > Cheers, > Arne > > On 28.02.22 09:28, ??? wrote: > > Hi Arne, > > > > I didn't find hardware RAID config option during the initial boot > > sequence. Ctrl+H is unresponsive in this computer. I just saw "Press > > Del to enter firmware configuration, press F3 to enter boot menu, and > > press F12 to enter network boot". And I press 'Del' to enter the BIOS. > > But I didn't find RAID config menu in BIOS. Sorry that I have poor > > knowledge about this. > > > > And now, even though I installed the operating system manually using a > > USB stick, I still couldn't find the hard drive. Is there anything > > that ironic-agent did in the clean phase that would have caused this > > problem? > > > > I wonder if this is a thinking pointto solve the problem. Now, my idea > > is to first find a way to manually configure RAID. Do you agree with > > this? And than, whether RAID configurations are still cleared in the > > Clean phase if clean phase will do this? > > > > Sorry that I have so much confuse. > > > > love you, > > Guangyu > > > > Arne Wiebalck ?2022?2?14??? 15:59??? > >> > >> Hi Guangyu, > >> > >> It seems like Julia had the right idea and the disks > >> are not visible since the RAID controller does not > >> expose anything to the operating system. This seems > >> to be confirmed by you booting into the CentOS7 image. > >> > >> What I would suggest to try next is to look for the > >> hardware RAID config option during the initial boot > >> sequence to enter the RAID config menu (there should be > >> a message quite early on, and maybe Ctrl-H is needed > >> to enter the menu). > >> > >> Once there, manually configure the disks as JBODs or > >> create a RAID device. Upon reboot this should be visible > >> and accessible as a device. Maybe check from your CentOS7 > >> image again. If the devices are there, Ironic should > >> also be able to deploy on them (for this you can remove > >> the RAID config you added). > >> > >> It depends a little on what your goal is, but I would > >> try this first to see if you can make a device visible > >> and if the Ironic deploy bit works, before trying to > >> configure the hardware RAID via Ironic. > >> > >> Cheers, > >> Arne > >> > >> On 14.02.22 03:20, ??? wrote: > >>> Hi Arne and Julia, > >>> > >>> You make me feel so warm. Best wishes to you. > >>> > >>> I have tried to boot the node into a CentOS7, but it still coundnot to > >>> find disk. And sorry that I didn't notice the RAID card. > >>> > >>> # lspci -v > >>> ... > >>> 23:00.0 RAID bus controller: Broadcom / LSI MegaRAID SAS-3 3108 > >>> [Invader] (rev 02) > >>> Subsystem: Broadcom / LSI MegaRAID SAS 9361-8i > >>> Flags: bus master, fast devsel, latency 0, IRQ -2147483648, NUMA node 1 > >>> I/O ports at 3000 [size=256] > >>> Memory at e9900000 (64-bit, non-prefetchable) [size=64K] > >>> Memory at e9700000 (64-bit, non-prefetchable) [size=1M] > >>> Expansion ROM at e9800000 [disabled] [size=1M] > >>> Capabilities: [50] Power Management version 3 > >>> Capabilities: [68] Express Endpoint, MSI 00 > >>> Capabilities: [d0] Vital Product Data > >>> Capabilities: [a8] MSI: Enable- Count=1/1 Maskable+ 64bit+ > >>> Capabilities: [c0] MSI-X: Enable+ Count=97 Masked- > >>> Capabilities: [100] Advanced Error Reporting > >>> Capabilities: [1e0] #19 > >>> Capabilities: [1c0] Power Budgeting > >>> Capabilities: [148] Alternative Routing-ID Interpretation (ARI) > >>> Kernel driver in use: megaraid_sas > >>> Kernel modules: megaraid_sas > >>> ... > >>> > >>> I try to config raid fallowing > >>> https://docs.openstack.org/ironic/latest/admin/raid.html > >>> by `baremetal node set $NODE_UUID --target-raid-config raid.json`. The > >>> server have three same disk(Western Digital DC HA210 2TB SATA 6GB/s) > >>> # cat raid.json > >>> { > >>> "logical_disks": [ > >>> { > >>> "size_gb": "MAX", > >>> "raid_level": "0", > >>> "is_root_volume": true > >>> } > >>> ] > >>> } > >>> > >>> But Ironic still coundn't see disk. I still got > >>> ``` > >>> ## In deploy images > >>> # journalctl -fxeu ironic-python-agent > >>> Feb 14 02:17:22 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 > >>> 02:17:22.863 2329 WARNING root [-] Path /dev/disk/by-path is > >>> inaccessible, /dev/disk/by-path/* version of block device name is > >>> unavailable Cause: [Errno 2] No such file or directory: > >>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or > >>> directory: '/dev/disk/by-path' > >>> Feb 14 02:17:44 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 > >>> 02:17:44.391 2329 ERROR root [-] Unexpected error dispatching > >>> get_os_install_device to manager > >>> >>> 0x7efbf4da2208>: Error finding the disk or partition device to deploy > >>> the image onto: No suitable device was found for deployment - root > >>> device hints were not provided and all found block devices are smaller > >>> than 4294967296B.: ironic_python_agent.errors.DeviceNotFound: Error > >>> finding the disk or partition device to deploy the image onto: No > >>> suitable device was found for deployment - root device hints were not > >>> provided and all found block devices are smaller than 4294967296B. > >>> ``` > >>> > >>> I don't know if it's a lack of a RAID card driver or a lack of a disk > >>> driver or a lack of RAID configuration. Could you have some idea about > >>> this question? > >>> > >>> love you, > >>> Han Guangyu > >>> > >>> > >>> Julia Kreger ?2022?2?10??? 23:11??? > >>> > >>>> > >>>> If the disk controllers *are* enumerated in the kernel log, which is > >>>> something to also look for, then the disks themselves may be in some > >>>> weird state like security locked. Generally this shows up as the > >>>> operating system kind of sees the disk and the SATA port connected but > >>>> can't really access it. This is also an exceptionally rare state to > >>>> find one's self in. > >>>> > >>>> More common, especially in enterprise grade hardware: If the disk > >>>> controller is actually a raid controller, and there are no raid > >>>> volumes configured, then the operating system likely cannot see the > >>>> underlying disks and turn that into a usable block device. I've seen a > >>>> couple drivers over the years which expose hints of disks in the > >>>> kernel log and without raid configuration in the cards, the drivers > >>>> can't present usable block devices to the operating system system. > >>>> > >>>> -Julia > >>>> > >>>> On Thu, Feb 10, 2022 at 3:17 AM Arne Wiebalck wrote: > >>>>> > >>>>> Hi Guangyu, > >>>>> > >>>>> No worries about asking questions, this is what the mailing > >>>>> list is for :) > >>>>> > >>>>> Just to clarify, you do not have to set root device hints, > >>>>> it also works without (with the algorithm I mentioned). > >>>>> However, hints help to define the exact device and/or make > >>>>> deployment more predictable/repeatable. > >>>>> > >>>>> If it is really a driver problem, it is an issue with the > >>>>> operating system of the image you use, i.e. CentOS8. Some > >>>>> drivers were removed from 7 to 8, and we have seen issues > >>>>> with specific drive models as well. > >>>>> > >>>>> You can try to build your own IPA images as described in > >>>>> [1], e.g. to add your ssh key to be able to log into the > >>>>> IPA to debug further, and to eventually include drivers > >>>>> (if you can identify them and they are available for CentOS8). > >>>>> > >>>>> Another option may be to add another (newer) disk model to > >>>>> the server, just to confirm it is the disk model/driver which > >>>>> is the cause. > >>>>> > >>>>> You could also try to boot the node into a CentOS7 (and then > >>>>> a CentOS8) live image to confirm it can see the disks at all. > >>>>> > >>>>> Hope this helps! > >>>>> Arne > >>>>> > >>>>> [1] > >>>>> https://docs.openstack.org/ironic-python-agent-builder/latest/admin/dib.html > >>>>> > >>>>> > >>>>> On 10.02.22 11:15, ??? wrote: > >>>>>> Hi Arne, > >>>>>> > >>>>>> Thank you very much for your response. Love you. You take away a lot > >>>>>> of my confusion. > >>>>>> > >>>>>> You are right, I didn't set 'root device'. And Ironic also can not see > >>>>>> disk, the content of the 'lsblk' file in the deploy los is emply. > >>>>>> I tried to set 'root device', but because ironic can't find any disk, > >>>>>> the deploy still filed. > >>>>>> > >>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 > >>>>>> 09:51:55.045 2324 WARNING root [-] Path /dev/disk/by-path is > >>>>>> inaccessible, /dev/disk/by-path/* version of block device name is > >>>>>> unavailable Cause: [Errno 2] No such file or directory: > >>>>>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or > >>>>>> directory: '/dev/disk/by-path' > >>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 > >>>>>> 09:51:55.056 2324 WARNING ironic_lib.utils [-] No device found that > >>>>>> matches the root device hints {'wwn': '0x50014EE2691D724C'}: > >>>>>> StopIteration > >>>>>> > >>>>>> Sorry to bother you, I'm a newcomer of Ironic and I didn't find > >>>>>> information about it on google. > >>>>>> > >>>>>> The bare metal node have three same disk(Western Digital DC HA210 2TB > >>>>>> SATA 6GB/s). Where I can confirm whether ironic-python-agent supports > >>>>>> this disk? > >>>>>> > >>>>>> And If Ironic cannot find disk since the corresponding drivers in the > >>>>>> IPA image are missing, do you know how to resolve it? I have used the > >>>>>> latest deploy images in > >>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/ > >>>>>> . Do I need to find and manually add driver in the source code or > >>>>>> ramdisk(That was difficult tome)? > >>>>>> > >>>>>> Love you. > >>>>>> > >>>>>> Cheers, > >>>>>> Guangyu > >>>>>> > >>>>>> Arne Wiebalck ?2022?2?10??? 15:51??? > >>>>>>> > >>>>>>> Hi Guangyu, > >>>>>>> > >>>>>>> The error indicates that Ironic was not able to find > >>>>>>> a device where it could deploy the image to. > >>>>>>> > >>>>>>> To find a device, Ironic will use 'root device' > >>>>>>> hints [1], usually set by the admin on a node. If that > >>>>>>> does not yield anything, Ironic will loop over all > >>>>>>> block devices and pick the smallest which is larger > >>>>>>> than 4GB (and order them alphabetically). > >>>>>>> > >>>>>>> If you have disks in your server which are larger than > >>>>>>> 4GB, one potential explanation is that Ironic cannot see them, > >>>>>>> e.g. since the corresponding drivers in the IPA image are missing. > >>>>>>> The logs you posted seem to confirm something along those > >>>>>>> lines. > >>>>>>> > >>>>>>> Check the content of the 'lsblk' file in the deploy logs which > >>>>>>> you can find in the tar archive in /var/log/ironic/deploy/ > >>>>>>> on the controller for your deployment attempt to see what > >>>>>>> devices Ironic has access to. > >>>>>>> > >>>>>>> Cheers, > >>>>>>> Arne > >>>>>>> > >>>>>>> > >>>>>>> [1] https://docs.openstack.org/ironic/latest/install/advanced.html#root-device-hints > >>>>>>> > >>>>>>> On 10.02.22 02:50, ??? wrote: > >>>>>>>> Dear all, > >>>>>>>> > >>>>>>>> I have a OpenStack Victoria environment, and tried to use ironic > >>>>>>>> manage bare metal. But I got "- root device hints were not provided > >>>>>>>> and all found block devices are smaller than 4294967296B." in deploy > >>>>>>>> stage. > >>>>>>>> > >>>>>>>> 2022-02-09 17:57:56.492 3908982 ERROR > >>>>>>>> ironic.drivers.modules.agent_base [-] Agent returned error for deploy > >>>>>>>> step {'step': 'write_image', 'priority': 80, 'argsinfo': None, > >>>>>>>> 'interface': 'deploy'} on node cc68c450-ce54-4e1c-be04-8b0a6169ef92 : > >>>>>>>> No suitable device was found for deployment - root device hints were > >>>>>>>> not provided and all found block devices are smaller than > >>>>>>>> 4294967296B.. > >>>>>>>> > >>>>>>>> I used "openstack server create --flavor my-baremetal-flavor --nic > >>>>>>>> net-id=$net_id --image $image testing" to deploy bare metal node. I > >>>>>>>> download deploy images(ipa-centos8-master.kernel and > >>>>>>>> ipa-centos8-master.initramfs) in > >>>>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/. > >>>>>>>> > >>>>>>>> The baremetal node info and flavor info as following: > >>>>>>>> https://paste.opendev.org/show/bV7lgO6RkNQY6ZGPbT2e/ > >>>>>>>> Ironic configure file as following: > >>>>>>>> https://paste.opendev.org/show/bTgY9Kpn7KWqwQl73aEa/ > >>>>>>>> Ironic-conductor log: https://paste.opendev.org/show/bFKZYlXmccxNxU8lEogk/ > >>>>>>>> The log of ironic-python-agent in bare metal node: > >>>>>>>> https://paste.opendev.org/show/btAuaMuV2IutV2Pa7YIa/ > >>>>>>>> > >>>>>>>> I see some old discussion about this, such as: > >>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1312187. But those > >>>>>>>> discussions took place a long time ago, not version V, and no solution > >>>>>>>> was seen. > >>>>>>>> > >>>>>>>> Does anyone know how to solve this problem? I would appreciate any > >>>>>>>> kind of guidance or help. > >>>>>>>> > >>>>>>>> Thank you, > >>>>>>>> Han Guangyu > >>>>>>>> > >>>>> From jesse at odyssey4.me Tue Mar 1 14:57:48 2022 From: jesse at odyssey4.me (Jesse Pretorius) Date: Tue, 1 Mar 2022 14:57:48 +0000 Subject: [TripleO] centos9 jobs only for master, centos 8 & 9 for wallaby In-Reply-To: References: Message-ID: <32D116CA-5034-4557-BEDF-6D7E8C7FDD48@odyssey4.me> Hi folks, response in-line. > On 24 Feb 2022, at 15:11, Marios Andreou wrote: > > Hello TripleO o/ > > During the last period the tripleo-ci team has been working on getting > tripleo CI jobs to run centos-9. > > On master branches across all TripleO repos we now run C9 exclusively > (since [1]). > > On Wallaby branches, after the work at [2] (in particular after > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 merges) we > will run both C8 and C9 jobs. I started an etherpad at [3] thinking > this would be a PTG topic, but I think we'll want to have this > conversation way before then. There is a tripleo-heat-templates > example in the etherpad that shows what we can expect from the job > runs (note the exact jobs that run will vary per repo and even between > changes depending on the files touched - but generally there will be > duplication between 8 and 9 jobs). > > The current proposal is that we keep a minimal subset on C8 Wallaby, > with C9 wallaby having the full set of jobs. > > Two factors that will affect our decisions are (there may be more?) > > i) Upgrades requirements - are we supporting upgrading to Wallaby on > 8? For example, the coming undercloud-upgrade-ffu job will be train 8 > to wallaby 8. In which case we definitely need to keep at least some > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby > completely). I think trying to keep up with the moving target of 8-stream would be counterproductive given that we are pinned downstream to RHEL 8.4 for Train. We do have a pending patch to merge a job for 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be nice to keep but if it becomes troublesome then I?d suggest we remove it. This job adds some value to us (and has added quite a bit already in preparing it and making it pass), but if 8-stream starts making it fail, or it starts failing due to database migrations then we?ll have to shift that job downstream and deal with the consequences of only ever finding upgrade bugs post-merge. It is well known that a Fast-Forward Upgrade (FFU) of OpenStack services is not supported upstream, so we can not expect any upstream jobs to reliability support this process. If/when there is broader support for enabling FFU support officially in OpenStack services, then we can reconsider this position. > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is 8 > but this will soon switch. > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla standalone > (no scenarios? some subset of them?), undercloud-ffu, minor update. > > This is just to start the conversation so please reply if you have > thoughts or comments about any of the above. > > We are planning to discuss this in the coming tripleo-ci community > call this coming Tuesday at 1330 UTC - meeting link at [4] so please > join us if you can and would like to participate, > > regards, marios > > [1] https://review.opendev.org/q/topic:c8_teardown_master > [2] https://review.opendev.org/q/topic:c9_wallaby_gates > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 > [4] https://meet.google.com/bqx-xwht-wky > > From mthode at mthode.org Tue Mar 1 17:12:17 2022 From: mthode at mthode.org (Matthew Thode) Date: Tue, 1 Mar 2022 11:12:17 -0600 Subject: [masakari][oslo][requirements] oslo.context>=4.0.0 seems to need updated tests In-Reply-To: <20220228165609.zzz7ur2mktfrqmue@mthode.org> References: <20220228165609.zzz7ur2mktfrqmue@mthode.org> Message-ID: <20220301171217.ixkrog3yrxefqntz@mthode.org> On 22-02-28 10:56:09, Matthew Thode wrote: > The update to see what's breaking is here > https://review.opendev.org/829599 > > Feel free to make that patch depend on your fix to test (with rechecks) > Looks like it's just masakari now, not too many commits though :| -- Matthew Thode From elod.illes at est.tech Tue Mar 1 20:43:35 2022 From: elod.illes at est.tech (=?UTF-8?B?RWzFkWQgSWxsw6lz?=) Date: Tue, 1 Mar 2022 21:43:35 +0100 Subject: [release][requirements][oslo][kuryr][keystone] RFE request Message-ID: <116ca5d1-c8e3-d51b-dd98-3b2c444a06b7@est.tech> Hi, It turned out that kuryr [1] and keystoneauth [2] deliverables were missed at non-client libraries final release period. So we need to release them now and request Requirement Freeze Exception for them. (PTLs / release liaisons of kuryr and keystoneauth: please review the patches [1][2] as soon as possible) Also I want to mention here oslo.context as well, to formally include it to RFE. (It seems versions >=4.0.0 break things, see thread [3] and [4].) [1] https://review.opendev.org/c/openstack/releases/+/831025 [2] https://review.opendev.org/c/openstack/releases/+/831024 [3] http://lists.openstack.org/pipermail/openstack-discuss/2022-February/027455.html [4] http://lists.openstack.org/pipermail/openstack-discuss/2022-March/027480.html El?d From sunny at openstack.org Tue Mar 1 20:29:48 2022 From: sunny at openstack.org (Sunny Cai) Date: Tue, 1 Mar 2022 12:29:48 -0800 Subject: How OpenStack is Used in Academia - OpenInfra Live References: <59f2a75e-8a74-4e04-899b-05e46f7496ab@Spark> Message-ID: Hi everyone, This week on OpenInfra Live episode, we are featuring topics on how OpenStack is used in Academic space Open source cloud computing has gained a lot of adoption momentum due to advantages including low cost, flexibility, scalability, and more. Since the early days of OpenStack with NASA?s direct involvement, OpenStack has been deployed in various notable academic institutions all over the world for their infrastructure needs. For this episode, we have invited experts from Oregon State University, Monash University, T?l?com Paris, and OpenStack Scientific SIG to talk about how they use OpenStack in academia. Episode: How OpenStack is Used in Academia Date and time: Thursday, March 3 at 9am CT (1500 UTC) You can watch us live on YouTube [1], LinkedIn [2] and Facebook[3]. Host:?Kendall Nelson, Senior Upstream Developer Advocate at the OpenInfra Foundation Speakers: ? Stig Telfer, Scientific SIG Chair and CTO at StackHPC Ltd ? Steve Quenette, Deputy Director at Monash eResearch Centre, Monash University ? R?mi Sharrock, Researcher and Lecturer at T?l?com Paris & Marc Jeanmougin, Research Engineer at T?l?com Paris ? Lance Albertson, Director at Oregon State University Open Source Lab Have an idea for a future episode? Share it now at ideas.openinfra.live [4]. [1] https://www.youtube.com/watch?v=xqGvMn0B7jc&ab_channel=OpenInfrastructureFoundation [2]?https://www.linkedin.com/video/event/urn:li:ugcPost:6902672962731016192/ [3] https://www.facebook.com/104139126308032/posts/4883571768364720/ [4] https://openinfrafoundation.formstack.com/forms/openinfralive Thanks, Sunny Cai OpenInfra Foundation Marketing & Community -------------- next part -------------- An HTML attachment was scrubbed... URL: From mthode at mthode.org Tue Mar 1 20:53:46 2022 From: mthode at mthode.org (Matthew Thode) Date: Tue, 1 Mar 2022 14:53:46 -0600 Subject: [release][requirements][oslo][kuryr][keystone] RFE request In-Reply-To: <116ca5d1-c8e3-d51b-dd98-3b2c444a06b7@est.tech> References: <116ca5d1-c8e3-d51b-dd98-3b2c444a06b7@est.tech> Message-ID: <20220301205346.nbdx5yi6fbcifeks@mthode.org> On 22-03-01 21:43:35, El?d Ill?s wrote: > Hi, > > It turned out that kuryr [1] and keystoneauth [2] deliverables were missed > at non-client libraries final release period. So we need to release them now > and request Requirement Freeze Exception for them. (PTLs / release liaisons > of kuryr and keystoneauth: please review the patches [1][2] as soon as > possible) > > Also I want to mention here oslo.context as well, to formally include it to > RFE. (It seems versions >=4.0.0 break things, see thread [3] and [4].) > > > [1] https://review.opendev.org/c/openstack/releases/+/831025 > [2] https://review.opendev.org/c/openstack/releases/+/831024 > [3] http://lists.openstack.org/pipermail/openstack-discuss/2022-February/027455.html > [4] > http://lists.openstack.org/pipermail/openstack-discuss/2022-March/027480.html > > > El?d > Requirements is ok with these and approves. Thanks for the formal email :D -- Matthew Thode -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 833 bytes Desc: not available URL: From laurentfdumont at gmail.com Wed Mar 2 03:31:07 2022 From: laurentfdumont at gmail.com (Laurent Dumont) Date: Tue, 1 Mar 2022 22:31:07 -0500 Subject: [horizon][customization] In-Reply-To: <7b01ef62-8442-3ab8-8c40-5935547bffae@ovhcloud.com> References: <3aa4a3df-fab3-ccc0-5bc3-821cdb80ce67@ovhcloud.com> <7b01ef62-8442-3ab8-8c40-5935547bffae@ovhcloud.com> Message-ID: Hi! That is possible. Which version of the doc we're you looking at? I think those are able to be modified through merge requests. On Tue, Mar 1, 2022 at 9:07 AM Mathilde Hermet wrote: > Hello, > > > I manage to find that the problem in your documentation is that you forgot > to tell that command "$python manage.py compress --force" must be run after > collecting statics. > > I run it and then my custom themes worked. > > This appears to be required in earlier version of horizon that is why > documentation is may be not up to date. > > > Thanks > > > Mathilde > On 3/1/22 00:25, Laurent Dumont wrote: > > Salut! > > A 500 from Horizon is a bit strange. What do the Horizon logs say? > > If you remove your customization and restart Horizon, does it recover? > > On Mon, Feb 28, 2022 at 11:17 AM Mathilde Hermet < > mathilde.hermet at ovhcloud.com> wrote: > >> Hello Openstack team, >> >> >> I'm currently working on customizing the Horizon (xena) dashboard to use >> the OVHcloud assets: logo.svg and logo-splash.svg. >> >> I've been following this documentation >> https://docs.openstack.org/horizon/latest/configuration/themes.html >> >> I added the ovhcloud theme in >> horizon/openstack_dashboard/local/local_settings.py in AVAILABLE_THEMES, >> >> I created the subdir ovhcloud in horizon/openstack_dashboard/themes and >> I created a ovhcloud/img with logo.svg and logo-splash.svg in it. >> >> I put a _variables.scss and a _styles.scss in /themes/ovhcloud/ >> >> Then I executed python manage.py collectstatic >> >> It put all the files in horizon-sources/horizon.../static/themes/ovhcloud >> >> >> But then when trying to connect with the session cookie >> theme="ovhcloud", I have a 500 error. >> >> i don't think it comes from images because i've been puting it in >> /horizon-souurces/horizon.../static/dashbaord.img and they were >> displayed correctly. >> >> >> Do you have any idea about the misconfiguration I've done ? >> >> >> Best regards, >> >> >> Mathilde Hermet >> >> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: From marios at redhat.com Wed Mar 2 06:57:55 2022 From: marios at redhat.com (Marios Andreou) Date: Wed, 2 Mar 2022 08:57:55 +0200 Subject: [TripleO] centos9 jobs only for master, centos 8 & 9 for wallaby In-Reply-To: <32D116CA-5034-4557-BEDF-6D7E8C7FDD48@odyssey4.me> References: <32D116CA-5034-4557-BEDF-6D7E8C7FDD48@odyssey4.me> Message-ID: On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius wrote: > > Hi folks, response in-line. > > > On 24 Feb 2022, at 15:11, Marios Andreou wrote: > > > > Hello TripleO o/ > > > > During the last period the tripleo-ci team has been working on getting > > tripleo CI jobs to run centos-9. > > > > On master branches across all TripleO repos we now run C9 exclusively > > (since [1]). > > > > On Wallaby branches, after the work at [2] (in particular after > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 merges) we > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking > > this would be a PTG topic, but I think we'll want to have this > > conversation way before then. There is a tripleo-heat-templates > > example in the etherpad that shows what we can expect from the job > > runs (note the exact jobs that run will vary per repo and even between > > changes depending on the files touched - but generally there will be > > duplication between 8 and 9 jobs). > > > > The current proposal is that we keep a minimal subset on C8 Wallaby, > > with C9 wallaby having the full set of jobs. > > > > Two factors that will affect our decisions are (there may be more?) > > > > i) Upgrades requirements - are we supporting upgrading to Wallaby on > > 8? For example, the coming undercloud-upgrade-ffu job will be train 8 > > to wallaby 8. In which case we definitely need to keep at least some > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby > > completely). > > I think trying to keep up with the moving target of 8-stream would be counterproductive given that we are pinned downstream to RHEL 8.4 for Train. We do have a pending patch to merge a job for 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be nice to keep but if it becomes troublesome then I?d suggest we remove it. This job adds some value to us (and has added quite a bit already in preparing it and making it pass), but if 8-stream starts making it fail, or it starts failing due to database migrations then we?ll have to shift that job downstream and deal with the consequences of only ever finding upgrade bugs post-merge. > > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack services is not supported upstream, so we can not expect any upstream jobs to reliability support this process. If/when there is broader support for enabling FFU support officially in OpenStack services, then we can reconsider this position. > Hi Jesse, thanks for replying here too. As shaped by the discussions yesterday and for the benefit of others - our current plan is to keep the c8 undercloud-ffu job and in fact that will be the *only* job we will keep on C8 (if standalone ffu upgrade becomes a thing then that will also be 8 only). At least, we'll make a best effort to keep this and we can re-evaluate once we hit blocking issues. In order to maintain this C8 gate we'll also need to maintain the equivalent periodic version for the integration line. So we'll have a C8 integration line running the undercloud-ffu job and possibly one ovb job (TBD, likely featureset 1). To be clear, our plan is to remove *all* centos8 check/gate jobs, except the undercloud-ffu. We are not planning to keep any 'base' set of c8 jobs unless someone can make a case for their value. The tripleo-ci team will start implementing the removal of c8/wallaby jobs once we switch to use c9 for the imports which should happen within the next week or so. regards, marios > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is 8 > > but this will soon switch. > > > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla standalone > > (no scenarios? some subset of them?), undercloud-ffu, minor update. > > > > This is just to start the conversation so please reply if you have > > thoughts or comments about any of the above. > > > > We are planning to discuss this in the coming tripleo-ci community > > call this coming Tuesday at 1330 UTC - meeting link at [4] so please > > join us if you can and would like to participate, > > > > regards, marios > > > > [1] https://review.opendev.org/q/topic:c8_teardown_master > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 > > [4] https://meet.google.com/bqx-xwht-wky > > > > > From jpodivin at redhat.com Wed Mar 2 07:43:59 2022 From: jpodivin at redhat.com (Jiri Podivin) Date: Wed, 2 Mar 2022 08:43:59 +0100 Subject: [TripleO] centos9 jobs only for master, centos 8 & 9 for wallaby In-Reply-To: References: <32D116CA-5034-4557-BEDF-6D7E8C7FDD48@odyssey4.me> Message-ID: Hi, Sorry I'm replying this late, but this line raised some questions for me: > To be clear, our plan is to remove *all* centos8 check/gate jobs, It seems to imply that all centos8 jobs in both pipelines will be removed, across stable branches. Does this mean up to and including train? It was my understanding that everything older than wallaby will remain as it is. On Wed, Mar 2, 2022 at 8:04 AM Marios Andreou wrote: > On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius wrote: > > > > Hi folks, response in-line. > > > > > On 24 Feb 2022, at 15:11, Marios Andreou wrote: > > > > > > Hello TripleO o/ > > > > > > During the last period the tripleo-ci team has been working on getting > > > tripleo CI jobs to run centos-9. > > > > > > On master branches across all TripleO repos we now run C9 exclusively > > > (since [1]). > > > > > > On Wallaby branches, after the work at [2] (in particular after > > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 merges) we > > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking > > > this would be a PTG topic, but I think we'll want to have this > > > conversation way before then. There is a tripleo-heat-templates > > > example in the etherpad that shows what we can expect from the job > > > runs (note the exact jobs that run will vary per repo and even between > > > changes depending on the files touched - but generally there will be > > > duplication between 8 and 9 jobs). > > > > > > The current proposal is that we keep a minimal subset on C8 Wallaby, > > > with C9 wallaby having the full set of jobs. > > > > > > Two factors that will affect our decisions are (there may be more?) > > > > > > i) Upgrades requirements - are we supporting upgrading to Wallaby on > > > 8? For example, the coming undercloud-upgrade-ffu job will be train 8 > > > to wallaby 8. In which case we definitely need to keep at least some > > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby > > > completely). > > > > I think trying to keep up with the moving target of 8-stream would be > counterproductive given that we are pinned downstream to RHEL 8.4 for > Train. We do have a pending patch to merge a job for > 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be > nice to keep but if it becomes troublesome then I?d suggest we remove it. > This job adds some value to us (and has added quite a bit already in > preparing it and making it pass), but if 8-stream starts making it fail, or > it starts failing due to database migrations then we?ll have to shift that > job downstream and deal with the consequences of only ever finding upgrade > bugs post-merge. > > > > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack services > is not supported upstream, so we can not expect any upstream jobs to > reliability support this process. If/when there is broader support for > enabling FFU support officially in OpenStack services, then we can > reconsider this position. > > > > Hi Jesse, > > thanks for replying here too. > > As shaped by the discussions yesterday and for the benefit of others - > our current plan is to keep the c8 undercloud-ffu job and in fact that > will be the *only* job we will keep on C8 (if standalone ffu upgrade > becomes a thing then that will also be 8 only). At least, we'll make a > best effort to keep this and we can re-evaluate once we hit blocking > issues. > > In order to maintain this C8 gate we'll also need to maintain the > equivalent periodic version for the integration line. So we'll have a > C8 integration line running the undercloud-ffu job and possibly one > ovb job (TBD, likely featureset 1). > > To be clear, our plan is to remove *all* centos8 check/gate jobs, > except the undercloud-ffu. We are not planning to keep any 'base' set > of c8 jobs unless someone can make a case for their value. > > The tripleo-ci team will start implementing the removal of c8/wallaby > jobs once we switch to use c9 for the imports which should happen > within the next week or so. > > regards, marios > > > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is 8 > > > but this will soon switch. > > > > > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla standalone > > > (no scenarios? some subset of them?), undercloud-ffu, minor update. > > > > > > This is just to start the conversation so please reply if you have > > > thoughts or comments about any of the above. > > > > > > We are planning to discuss this in the coming tripleo-ci community > > > call this coming Tuesday at 1330 UTC - meeting link at [4] so please > > > join us if you can and would like to participate, > > > > > > regards, marios > > > > > > [1] https://review.opendev.org/q/topic:c8_teardown_master > > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates > > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 > > > [4] https://meet.google.com/bqx-xwht-wky > > > > > > > > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From marios at redhat.com Wed Mar 2 08:05:18 2022 From: marios at redhat.com (Marios Andreou) Date: Wed, 2 Mar 2022 10:05:18 +0200 Subject: [TripleO] centos9 jobs only for master, centos 8 & 9 for wallaby In-Reply-To: References: <32D116CA-5034-4557-BEDF-6D7E8C7FDD48@odyssey4.me> Message-ID: On Wed, Mar 2, 2022 at 9:44 AM Jiri Podivin wrote: > > Hi, > Sorry I'm replying this late, but this line raised some questions for me: > > To be clear, our plan is to remove *all* centos8 check/gate jobs, > > It seems to imply that all centos8 jobs in both pipelines will be removed, across stable branches. > Does this mean up to and including train? > It was my understanding that everything older than wallaby will remain as it is. > This discussion is exclusively about wallaby - older stable branches will be unaffected. > > On Wed, Mar 2, 2022 at 8:04 AM Marios Andreou wrote: >> >> On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius wrote: >> > >> > Hi folks, response in-line. >> > >> > > On 24 Feb 2022, at 15:11, Marios Andreou wrote: >> > > >> > > Hello TripleO o/ >> > > >> > > During the last period the tripleo-ci team has been working on getting >> > > tripleo CI jobs to run centos-9. >> > > >> > > On master branches across all TripleO repos we now run C9 exclusively >> > > (since [1]). >> > > >> > > On Wallaby branches, after the work at [2] (in particular after >> > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 merges) we >> > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking >> > > this would be a PTG topic, but I think we'll want to have this >> > > conversation way before then. There is a tripleo-heat-templates >> > > example in the etherpad that shows what we can expect from the job >> > > runs (note the exact jobs that run will vary per repo and even between >> > > changes depending on the files touched - but generally there will be >> > > duplication between 8 and 9 jobs). >> > > >> > > The current proposal is that we keep a minimal subset on C8 Wallaby, >> > > with C9 wallaby having the full set of jobs. >> > > >> > > Two factors that will affect our decisions are (there may be more?) >> > > >> > > i) Upgrades requirements - are we supporting upgrading to Wallaby on >> > > 8? For example, the coming undercloud-upgrade-ffu job will be train 8 >> > > to wallaby 8. In which case we definitely need to keep at least some >> > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby >> > > completely). >> > >> > I think trying to keep up with the moving target of 8-stream would be counterproductive given that we are pinned downstream to RHEL 8.4 for Train. We do have a pending patch to merge a job for 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be nice to keep but if it becomes troublesome then I?d suggest we remove it. This job adds some value to us (and has added quite a bit already in preparing it and making it pass), but if 8-stream starts making it fail, or it starts failing due to database migrations then we?ll have to shift that job downstream and deal with the consequences of only ever finding upgrade bugs post-merge. >> > >> > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack services is not supported upstream, so we can not expect any upstream jobs to reliability support this process. If/when there is broader support for enabling FFU support officially in OpenStack services, then we can reconsider this position. >> > >> >> Hi Jesse, >> >> thanks for replying here too. >> >> As shaped by the discussions yesterday and for the benefit of others - >> our current plan is to keep the c8 undercloud-ffu job and in fact that >> will be the *only* job we will keep on C8 (if standalone ffu upgrade >> becomes a thing then that will also be 8 only). At least, we'll make a >> best effort to keep this and we can re-evaluate once we hit blocking >> issues. >> >> In order to maintain this C8 gate we'll also need to maintain the >> equivalent periodic version for the integration line. So we'll have a >> C8 integration line running the undercloud-ffu job and possibly one >> ovb job (TBD, likely featureset 1). >> >> To be clear, our plan is to remove *all* centos8 check/gate jobs, >> except the undercloud-ffu. We are not planning to keep any 'base' set >> of c8 jobs unless someone can make a case for their value. >> >> The tripleo-ci team will start implementing the removal of c8/wallaby >> jobs once we switch to use c9 for the imports which should happen >> within the next week or so. >> >> regards, marios >> >> > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is 8 >> > > but this will soon switch. >> > > >> > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla standalone >> > > (no scenarios? some subset of them?), undercloud-ffu, minor update. >> > > >> > > This is just to start the conversation so please reply if you have >> > > thoughts or comments about any of the above. >> > > >> > > We are planning to discuss this in the coming tripleo-ci community >> > > call this coming Tuesday at 1330 UTC - meeting link at [4] so please >> > > join us if you can and would like to participate, >> > > >> > > regards, marios >> > > >> > > [1] https://review.opendev.org/q/topic:c8_teardown_master >> > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates >> > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 >> > > [4] https://meet.google.com/bqx-xwht-wky >> > > >> > > >> > >> >> On Wed, Mar 2, 2022 at 9:44 AM Jiri Podivin wrote: > > Hi, > Sorry I'm replying this late, but this line raised some questions for me: > > To be clear, our plan is to remove *all* centos8 check/gate jobs, > > It seems to imply that all centos8 jobs in both pipelines will be removed, across stable branches. > Does this mean up to and including train? > It was my understanding that everything older than wallaby will remain as it is. > > > On Wed, Mar 2, 2022 at 8:04 AM Marios Andreou wrote: >> >> On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius wrote: >> > >> > Hi folks, response in-line. >> > >> > > On 24 Feb 2022, at 15:11, Marios Andreou wrote: >> > > >> > > Hello TripleO o/ >> > > >> > > During the last period the tripleo-ci team has been working on getting >> > > tripleo CI jobs to run centos-9. >> > > >> > > On master branches across all TripleO repos we now run C9 exclusively >> > > (since [1]). >> > > >> > > On Wallaby branches, after the work at [2] (in particular after >> > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 merges) we >> > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking >> > > this would be a PTG topic, but I think we'll want to have this >> > > conversation way before then. There is a tripleo-heat-templates >> > > example in the etherpad that shows what we can expect from the job >> > > runs (note the exact jobs that run will vary per repo and even between >> > > changes depending on the files touched - but generally there will be >> > > duplication between 8 and 9 jobs). >> > > >> > > The current proposal is that we keep a minimal subset on C8 Wallaby, >> > > with C9 wallaby having the full set of jobs. >> > > >> > > Two factors that will affect our decisions are (there may be more?) >> > > >> > > i) Upgrades requirements - are we supporting upgrading to Wallaby on >> > > 8? For example, the coming undercloud-upgrade-ffu job will be train 8 >> > > to wallaby 8. In which case we definitely need to keep at least some >> > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby >> > > completely). >> > >> > I think trying to keep up with the moving target of 8-stream would be counterproductive given that we are pinned downstream to RHEL 8.4 for Train. We do have a pending patch to merge a job for 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be nice to keep but if it becomes troublesome then I?d suggest we remove it. This job adds some value to us (and has added quite a bit already in preparing it and making it pass), but if 8-stream starts making it fail, or it starts failing due to database migrations then we?ll have to shift that job downstream and deal with the consequences of only ever finding upgrade bugs post-merge. >> > >> > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack services is not supported upstream, so we can not expect any upstream jobs to reliability support this process. If/when there is broader support for enabling FFU support officially in OpenStack services, then we can reconsider this position. >> > >> >> Hi Jesse, >> >> thanks for replying here too. >> >> As shaped by the discussions yesterday and for the benefit of others - >> our current plan is to keep the c8 undercloud-ffu job and in fact that >> will be the *only* job we will keep on C8 (if standalone ffu upgrade >> becomes a thing then that will also be 8 only). At least, we'll make a >> best effort to keep this and we can re-evaluate once we hit blocking >> issues. >> >> In order to maintain this C8 gate we'll also need to maintain the >> equivalent periodic version for the integration line. So we'll have a >> C8 integration line running the undercloud-ffu job and possibly one >> ovb job (TBD, likely featureset 1). >> >> To be clear, our plan is to remove *all* centos8 check/gate jobs, >> except the undercloud-ffu. We are not planning to keep any 'base' set >> of c8 jobs unless someone can make a case for their value. >> >> The tripleo-ci team will start implementing the removal of c8/wallaby >> jobs once we switch to use c9 for the imports which should happen >> within the next week or so. >> >> regards, marios >> >> > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is 8 >> > > but this will soon switch. >> > > >> > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla standalone >> > > (no scenarios? some subset of them?), undercloud-ffu, minor update. >> > > >> > > This is just to start the conversation so please reply if you have >> > > thoughts or comments about any of the above. >> > > >> > > We are planning to discuss this in the coming tripleo-ci community >> > > call this coming Tuesday at 1330 UTC - meeting link at [4] so please >> > > join us if you can and would like to participate, >> > > >> > > regards, marios >> > > >> > > [1] https://review.opendev.org/q/topic:c8_teardown_master >> > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates >> > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 >> > > [4] https://meet.google.com/bqx-xwht-wky >> > > >> > > >> > >> >> From jpodivin at redhat.com Wed Mar 2 08:07:33 2022 From: jpodivin at redhat.com (Jiri Podivin) Date: Wed, 2 Mar 2022 09:07:33 +0100 Subject: [TripleO] centos9 jobs only for master, centos 8 & 9 for wallaby In-Reply-To: References: <32D116CA-5034-4557-BEDF-6D7E8C7FDD48@odyssey4.me> Message-ID: Great, Thanks for the clarification, it had me worried for a bit. On Wed, Mar 2, 2022 at 9:05 AM Marios Andreou wrote: > On Wed, Mar 2, 2022 at 9:44 AM Jiri Podivin wrote: > > > > Hi, > > Sorry I'm replying this late, but this line raised some questions for me: > > > To be clear, our plan is to remove *all* centos8 check/gate jobs, > > > > It seems to imply that all centos8 jobs in both pipelines will be > removed, across stable branches. > > Does this mean up to and including train? > > It was my understanding that everything older than wallaby will remain > as it is. > > > > This discussion is exclusively about wallaby - older stable branches > will be unaffected. > > > > > > > > > On Wed, Mar 2, 2022 at 8:04 AM Marios Andreou wrote: > >> > >> On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius > wrote: > >> > > >> > Hi folks, response in-line. > >> > > >> > > On 24 Feb 2022, at 15:11, Marios Andreou wrote: > >> > > > >> > > Hello TripleO o/ > >> > > > >> > > During the last period the tripleo-ci team has been working on > getting > >> > > tripleo CI jobs to run centos-9. > >> > > > >> > > On master branches across all TripleO repos we now run C9 > exclusively > >> > > (since [1]). > >> > > > >> > > On Wallaby branches, after the work at [2] (in particular after > >> > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 merges) > we > >> > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking > >> > > this would be a PTG topic, but I think we'll want to have this > >> > > conversation way before then. There is a tripleo-heat-templates > >> > > example in the etherpad that shows what we can expect from the job > >> > > runs (note the exact jobs that run will vary per repo and even > between > >> > > changes depending on the files touched - but generally there will be > >> > > duplication between 8 and 9 jobs). > >> > > > >> > > The current proposal is that we keep a minimal subset on C8 Wallaby, > >> > > with C9 wallaby having the full set of jobs. > >> > > > >> > > Two factors that will affect our decisions are (there may be more?) > >> > > > >> > > i) Upgrades requirements - are we supporting upgrading to Wallaby > on > >> > > 8? For example, the coming undercloud-upgrade-ffu job will be train > 8 > >> > > to wallaby 8. In which case we definitely need to keep at least some > >> > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby > >> > > completely). > >> > > >> > I think trying to keep up with the moving target of 8-stream would be > counterproductive given that we are pinned downstream to RHEL 8.4 for > Train. We do have a pending patch to merge a job for > 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be > nice to keep but if it becomes troublesome then I?d suggest we remove it. > This job adds some value to us (and has added quite a bit already in > preparing it and making it pass), but if 8-stream starts making it fail, or > it starts failing due to database migrations then we?ll have to shift that > job downstream and deal with the consequences of only ever finding upgrade > bugs post-merge. > >> > > >> > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack > services is not supported upstream, so we can not expect any upstream jobs > to reliability support this process. If/when there is broader support for > enabling FFU support officially in OpenStack services, then we can > reconsider this position. > >> > > >> > >> Hi Jesse, > >> > >> thanks for replying here too. > >> > >> As shaped by the discussions yesterday and for the benefit of others - > >> our current plan is to keep the c8 undercloud-ffu job and in fact that > >> will be the *only* job we will keep on C8 (if standalone ffu upgrade > >> becomes a thing then that will also be 8 only). At least, we'll make a > >> best effort to keep this and we can re-evaluate once we hit blocking > >> issues. > >> > >> In order to maintain this C8 gate we'll also need to maintain the > >> equivalent periodic version for the integration line. So we'll have a > >> C8 integration line running the undercloud-ffu job and possibly one > >> ovb job (TBD, likely featureset 1). > >> > >> To be clear, our plan is to remove *all* centos8 check/gate jobs, > >> except the undercloud-ffu. We are not planning to keep any 'base' set > >> of c8 jobs unless someone can make a case for their value. > >> > >> The tripleo-ci team will start implementing the removal of c8/wallaby > >> jobs once we switch to use c9 for the imports which should happen > >> within the next week or so. > >> > >> regards, marios > >> > >> > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is 8 > >> > > but this will soon switch. > >> > > > >> > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla > standalone > >> > > (no scenarios? some subset of them?), undercloud-ffu, minor update. > >> > > > >> > > This is just to start the conversation so please reply if you have > >> > > thoughts or comments about any of the above. > >> > > > >> > > We are planning to discuss this in the coming tripleo-ci community > >> > > call this coming Tuesday at 1330 UTC - meeting link at [4] so please > >> > > join us if you can and would like to participate, > >> > > > >> > > regards, marios > >> > > > >> > > [1] https://review.opendev.org/q/topic:c8_teardown_master > >> > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates > >> > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 > >> > > [4] https://meet.google.com/bqx-xwht-wky > >> > > > >> > > > >> > > >> > >> > > On Wed, Mar 2, 2022 at 9:44 AM Jiri Podivin wrote: > > > > Hi, > > Sorry I'm replying this late, but this line raised some questions for me: > > > To be clear, our plan is to remove *all* centos8 check/gate jobs, > > > > It seems to imply that all centos8 jobs in both pipelines will be > removed, across stable branches. > > Does this mean up to and including train? > > It was my understanding that everything older than wallaby will remain > as it is. > > > > > > On Wed, Mar 2, 2022 at 8:04 AM Marios Andreou wrote: > >> > >> On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius > wrote: > >> > > >> > Hi folks, response in-line. > >> > > >> > > On 24 Feb 2022, at 15:11, Marios Andreou wrote: > >> > > > >> > > Hello TripleO o/ > >> > > > >> > > During the last period the tripleo-ci team has been working on > getting > >> > > tripleo CI jobs to run centos-9. > >> > > > >> > > On master branches across all TripleO repos we now run C9 > exclusively > >> > > (since [1]). > >> > > > >> > > On Wallaby branches, after the work at [2] (in particular after > >> > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 merges) > we > >> > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking > >> > > this would be a PTG topic, but I think we'll want to have this > >> > > conversation way before then. There is a tripleo-heat-templates > >> > > example in the etherpad that shows what we can expect from the job > >> > > runs (note the exact jobs that run will vary per repo and even > between > >> > > changes depending on the files touched - but generally there will be > >> > > duplication between 8 and 9 jobs). > >> > > > >> > > The current proposal is that we keep a minimal subset on C8 Wallaby, > >> > > with C9 wallaby having the full set of jobs. > >> > > > >> > > Two factors that will affect our decisions are (there may be more?) > >> > > > >> > > i) Upgrades requirements - are we supporting upgrading to Wallaby > on > >> > > 8? For example, the coming undercloud-upgrade-ffu job will be train > 8 > >> > > to wallaby 8. In which case we definitely need to keep at least some > >> > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby > >> > > completely). > >> > > >> > I think trying to keep up with the moving target of 8-stream would be > counterproductive given that we are pinned downstream to RHEL 8.4 for > Train. We do have a pending patch to merge a job for > 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be > nice to keep but if it becomes troublesome then I?d suggest we remove it. > This job adds some value to us (and has added quite a bit already in > preparing it and making it pass), but if 8-stream starts making it fail, or > it starts failing due to database migrations then we?ll have to shift that > job downstream and deal with the consequences of only ever finding upgrade > bugs post-merge. > >> > > >> > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack > services is not supported upstream, so we can not expect any upstream jobs > to reliability support this process. If/when there is broader support for > enabling FFU support officially in OpenStack services, then we can > reconsider this position. > >> > > >> > >> Hi Jesse, > >> > >> thanks for replying here too. > >> > >> As shaped by the discussions yesterday and for the benefit of others - > >> our current plan is to keep the c8 undercloud-ffu job and in fact that > >> will be the *only* job we will keep on C8 (if standalone ffu upgrade > >> becomes a thing then that will also be 8 only). At least, we'll make a > >> best effort to keep this and we can re-evaluate once we hit blocking > >> issues. > >> > >> In order to maintain this C8 gate we'll also need to maintain the > >> equivalent periodic version for the integration line. So we'll have a > >> C8 integration line running the undercloud-ffu job and possibly one > >> ovb job (TBD, likely featureset 1). > >> > >> To be clear, our plan is to remove *all* centos8 check/gate jobs, > >> except the undercloud-ffu. We are not planning to keep any 'base' set > >> of c8 jobs unless someone can make a case for their value. > >> > >> The tripleo-ci team will start implementing the removal of c8/wallaby > >> jobs once we switch to use c9 for the imports which should happen > >> within the next week or so. > >> > >> regards, marios > >> > >> > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is 8 > >> > > but this will soon switch. > >> > > > >> > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla > standalone > >> > > (no scenarios? some subset of them?), undercloud-ffu, minor update. > >> > > > >> > > This is just to start the conversation so please reply if you have > >> > > thoughts or comments about any of the above. > >> > > > >> > > We are planning to discuss this in the coming tripleo-ci community > >> > > call this coming Tuesday at 1330 UTC - meeting link at [4] so please > >> > > join us if you can and would like to participate, > >> > > > >> > > regards, marios > >> > > > >> > > [1] https://review.opendev.org/q/topic:c8_teardown_master > >> > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates > >> > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 > >> > > [4] https://meet.google.com/bqx-xwht-wky > >> > > > >> > > > >> > > >> > >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From mdemaced at redhat.com Wed Mar 2 10:45:01 2022 From: mdemaced at redhat.com (Maysa De Macedo Souza) Date: Wed, 2 Mar 2022 11:45:01 +0100 Subject: [release][requirements][oslo][kuryr][keystone] RFE request In-Reply-To: <20220301205346.nbdx5yi6fbcifeks@mthode.org> References: <116ca5d1-c8e3-d51b-dd98-3b2c444a06b7@est.tech> <20220301205346.nbdx5yi6fbcifeks@mthode.org> Message-ID: Hello, Thanks for catching that. I just checked the Kuryr related patch. On Tue, Mar 1, 2022 at 9:58 PM Matthew Thode wrote: > On 22-03-01 21:43:35, El?d Ill?s wrote: > > Hi, > > > > It turned out that kuryr [1] and keystoneauth [2] deliverables were > missed > > at non-client libraries final release period. So we need to release them > now > > and request Requirement Freeze Exception for them. (PTLs / release > liaisons > > of kuryr and keystoneauth: please review the patches [1][2] as soon as > > possible) > > > > Also I want to mention here oslo.context as well, to formally include it > to > > RFE. (It seems versions >=4.0.0 break things, see thread [3] and [4].) > > > > > > [1] https://review.opendev.org/c/openstack/releases/+/831025 > > [2] https://review.opendev.org/c/openstack/releases/+/831024 > > [3] > http://lists.openstack.org/pipermail/openstack-discuss/2022-February/027455.html > > [4] > > > http://lists.openstack.org/pipermail/openstack-discuss/2022-March/027480.html > > > > > > El?d > > > > Requirements is ok with these and approves. Thanks for the formal email > :D > > -- > Matthew Thode > -------------- next part -------------- An HTML attachment was scrubbed... URL: From zigo at debian.org Wed Mar 2 11:16:44 2022 From: zigo at debian.org (Thomas Goirand) Date: Wed, 2 Mar 2022 12:16:44 +0100 Subject: [horizon][customization] In-Reply-To: <3aa4a3df-fab3-ccc0-5bc3-821cdb80ce67@ovhcloud.com> References: <3aa4a3df-fab3-ccc0-5bc3-821cdb80ce67@ovhcloud.com> Message-ID: <375340aa-227f-8fc2-77f6-c147bf247b98@debian.org> On 2/28/22 09:45, Mathilde Hermet wrote: > Hello Openstack team, > > > I'm currently working on customizing the Horizon (xena) dashboard to use > the OVHcloud assets: logo.svg and logo-splash.svg. > > I've been following this documentation > https://docs.openstack.org/horizon/latest/configuration/themes.html > > I added the ovhcloud theme in > horizon/openstack_dashboard/local/local_settings.py in AVAILABLE_THEMES, > > I created the subdir ovhcloud in horizon/openstack_dashboard/themes and > I created a ovhcloud/img with logo.svg and logo-splash.svg in it. > > I put a _variables.scss and a _styles.scss in /themes/ovhcloud/ > > Then I executed python manage.py collectstatic > > It put all the files in horizon-sources/horizon.../static/themes/ovhcloud > > > But then when trying to connect with the session cookie > theme="ovhcloud", I have a 500 error. > > i don't think it comes from images because i've been puting it in > /horizon-souurces/horizon.../static/dashbaord.img and they were > displayed correctly. > > > Do you have any idea about the misconfiguration I've done ? > > > Best regards, > > > Mathilde Hermet Hi Mathilde, I'm not sure where your error is, but if you want, you can have a look at for exemple: https://packages.debian.org/openstack-dashboard-debian-theme which has it's git over here: https://salsa.debian.org/openstack-team/debian/openstack-dashboard-debian-theme I hope this helps, Cheers, Thomas Goirand (zigo) From mrunge at matthias-runge.de Wed Mar 2 12:37:15 2022 From: mrunge at matthias-runge.de (Matthias Runge) Date: Wed, 2 Mar 2022 13:37:15 +0100 Subject: [horizon][customization] In-Reply-To: <375340aa-227f-8fc2-77f6-c147bf247b98@debian.org> References: <3aa4a3df-fab3-ccc0-5bc3-821cdb80ce67@ovhcloud.com> <375340aa-227f-8fc2-77f6-c147bf247b98@debian.org> Message-ID: On 3/2/22 12:16, Thomas Goirand wrote: >> >> >> Do you have any idea about the misconfiguration I've done ? >> >> >> Best regards, >> >> >> Mathilde Hermet > > Hi Mathilde, > > I'm not sure where your error is, but if you want, you can have a look > at for exemple: > https://packages.debian.org/openstack-dashboard-debian-theme > > which has it's git over here: > https://salsa.debian.org/openstack-team/debian/openstack-dashboard-debian-theme > > > I hope this helps, > > Cheers, > > Thomas Goirand (zigo) > a 500 error is always a server error and should be logged in http error log. I'd take a look at the error log in the next step. Matthias From johfulto at redhat.com Wed Mar 2 13:09:36 2022 From: johfulto at redhat.com (John Fulton) Date: Wed, 2 Mar 2022 08:09:36 -0500 Subject: [TripleO] centos9 jobs only for master, centos 8 & 9 for wallaby In-Reply-To: References: <32D116CA-5034-4557-BEDF-6D7E8C7FDD48@odyssey4.me> Message-ID: Just wanted to add that for Ceph integration in TripleO, these are the important Ceph jobs which run when THT is patched in the main branch: tripleo-ci-centos-9-scenario001-standalone tripleo-ci-centos-9-scenario004-standalone If the Wallaby branch only had centos-9 versions of the above and no centos-8 versions of the above, then I don't see that as a problem. John On Wed, Mar 2, 2022 at 3:08 AM Jiri Podivin wrote: > Great, > Thanks for the clarification, it had me worried for a bit. > > On Wed, Mar 2, 2022 at 9:05 AM Marios Andreou wrote: > >> On Wed, Mar 2, 2022 at 9:44 AM Jiri Podivin wrote: >> > >> > Hi, >> > Sorry I'm replying this late, but this line raised some questions for >> me: >> > > To be clear, our plan is to remove *all* centos8 check/gate jobs, >> > >> > It seems to imply that all centos8 jobs in both pipelines will be >> removed, across stable branches. >> > Does this mean up to and including train? >> > It was my understanding that everything older than wallaby will remain >> as it is. >> > >> >> This discussion is exclusively about wallaby - older stable branches >> will be unaffected. >> >> >> >> >> >> > >> > On Wed, Mar 2, 2022 at 8:04 AM Marios Andreou >> wrote: >> >> >> >> On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius >> wrote: >> >> > >> >> > Hi folks, response in-line. >> >> > >> >> > > On 24 Feb 2022, at 15:11, Marios Andreou >> wrote: >> >> > > >> >> > > Hello TripleO o/ >> >> > > >> >> > > During the last period the tripleo-ci team has been working on >> getting >> >> > > tripleo CI jobs to run centos-9. >> >> > > >> >> > > On master branches across all TripleO repos we now run C9 >> exclusively >> >> > > (since [1]). >> >> > > >> >> > > On Wallaby branches, after the work at [2] (in particular after >> >> > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 >> merges) we >> >> > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking >> >> > > this would be a PTG topic, but I think we'll want to have this >> >> > > conversation way before then. There is a tripleo-heat-templates >> >> > > example in the etherpad that shows what we can expect from the job >> >> > > runs (note the exact jobs that run will vary per repo and even >> between >> >> > > changes depending on the files touched - but generally there will >> be >> >> > > duplication between 8 and 9 jobs). >> >> > > >> >> > > The current proposal is that we keep a minimal subset on C8 >> Wallaby, >> >> > > with C9 wallaby having the full set of jobs. >> >> > > >> >> > > Two factors that will affect our decisions are (there may be more?) >> >> > > >> >> > > i) Upgrades requirements - are we supporting upgrading to Wallaby >> on >> >> > > 8? For example, the coming undercloud-upgrade-ffu job will be >> train 8 >> >> > > to wallaby 8. In which case we definitely need to keep at least >> some >> >> > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby >> >> > > completely). >> >> > >> >> > I think trying to keep up with the moving target of 8-stream would >> be counterproductive given that we are pinned downstream to RHEL 8.4 for >> Train. We do have a pending patch to merge a job for >> 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be >> nice to keep but if it becomes troublesome then I?d suggest we remove it. >> This job adds some value to us (and has added quite a bit already in >> preparing it and making it pass), but if 8-stream starts making it fail, or >> it starts failing due to database migrations then we?ll have to shift that >> job downstream and deal with the consequences of only ever finding upgrade >> bugs post-merge. >> >> > >> >> > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack >> services is not supported upstream, so we can not expect any upstream jobs >> to reliability support this process. If/when there is broader support for >> enabling FFU support officially in OpenStack services, then we can >> reconsider this position. >> >> > >> >> >> >> Hi Jesse, >> >> >> >> thanks for replying here too. >> >> >> >> As shaped by the discussions yesterday and for the benefit of others - >> >> our current plan is to keep the c8 undercloud-ffu job and in fact that >> >> will be the *only* job we will keep on C8 (if standalone ffu upgrade >> >> becomes a thing then that will also be 8 only). At least, we'll make a >> >> best effort to keep this and we can re-evaluate once we hit blocking >> >> issues. >> >> >> >> In order to maintain this C8 gate we'll also need to maintain the >> >> equivalent periodic version for the integration line. So we'll have a >> >> C8 integration line running the undercloud-ffu job and possibly one >> >> ovb job (TBD, likely featureset 1). >> >> >> >> To be clear, our plan is to remove *all* centos8 check/gate jobs, >> >> except the undercloud-ffu. We are not planning to keep any 'base' set >> >> of c8 jobs unless someone can make a case for their value. >> >> >> >> The tripleo-ci team will start implementing the removal of c8/wallaby >> >> jobs once we switch to use c9 for the imports which should happen >> >> within the next week or so. >> >> >> >> regards, marios >> >> >> >> > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is >> 8 >> >> > > but this will soon switch. >> >> > > >> >> > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla >> standalone >> >> > > (no scenarios? some subset of them?), undercloud-ffu, minor update. >> >> > > >> >> > > This is just to start the conversation so please reply if you have >> >> > > thoughts or comments about any of the above. >> >> > > >> >> > > We are planning to discuss this in the coming tripleo-ci community >> >> > > call this coming Tuesday at 1330 UTC - meeting link at [4] so >> please >> >> > > join us if you can and would like to participate, >> >> > > >> >> > > regards, marios >> >> > > >> >> > > [1] https://review.opendev.org/q/topic:c8_teardown_master >> >> > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates >> >> > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 >> >> > > [4] https://meet.google.com/bqx-xwht-wky >> >> > > >> >> > > >> >> > >> >> >> >> >> >> On Wed, Mar 2, 2022 at 9:44 AM Jiri Podivin wrote: >> > >> > Hi, >> > Sorry I'm replying this late, but this line raised some questions for >> me: >> > > To be clear, our plan is to remove *all* centos8 check/gate jobs, >> > >> > It seems to imply that all centos8 jobs in both pipelines will be >> removed, across stable branches. >> > Does this mean up to and including train? >> > It was my understanding that everything older than wallaby will remain >> as it is. >> > >> > >> > On Wed, Mar 2, 2022 at 8:04 AM Marios Andreou >> wrote: >> >> >> >> On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius >> wrote: >> >> > >> >> > Hi folks, response in-line. >> >> > >> >> > > On 24 Feb 2022, at 15:11, Marios Andreou >> wrote: >> >> > > >> >> > > Hello TripleO o/ >> >> > > >> >> > > During the last period the tripleo-ci team has been working on >> getting >> >> > > tripleo CI jobs to run centos-9. >> >> > > >> >> > > On master branches across all TripleO repos we now run C9 >> exclusively >> >> > > (since [1]). >> >> > > >> >> > > On Wallaby branches, after the work at [2] (in particular after >> >> > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 >> merges) we >> >> > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking >> >> > > this would be a PTG topic, but I think we'll want to have this >> >> > > conversation way before then. There is a tripleo-heat-templates >> >> > > example in the etherpad that shows what we can expect from the job >> >> > > runs (note the exact jobs that run will vary per repo and even >> between >> >> > > changes depending on the files touched - but generally there will >> be >> >> > > duplication between 8 and 9 jobs). >> >> > > >> >> > > The current proposal is that we keep a minimal subset on C8 >> Wallaby, >> >> > > with C9 wallaby having the full set of jobs. >> >> > > >> >> > > Two factors that will affect our decisions are (there may be more?) >> >> > > >> >> > > i) Upgrades requirements - are we supporting upgrading to Wallaby >> on >> >> > > 8? For example, the coming undercloud-upgrade-ffu job will be >> train 8 >> >> > > to wallaby 8. In which case we definitely need to keep at least >> some >> >> > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby >> >> > > completely). >> >> > >> >> > I think trying to keep up with the moving target of 8-stream would >> be counterproductive given that we are pinned downstream to RHEL 8.4 for >> Train. We do have a pending patch to merge a job for >> 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be >> nice to keep but if it becomes troublesome then I?d suggest we remove it. >> This job adds some value to us (and has added quite a bit already in >> preparing it and making it pass), but if 8-stream starts making it fail, or >> it starts failing due to database migrations then we?ll have to shift that >> job downstream and deal with the consequences of only ever finding upgrade >> bugs post-merge. >> >> > >> >> > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack >> services is not supported upstream, so we can not expect any upstream jobs >> to reliability support this process. If/when there is broader support for >> enabling FFU support officially in OpenStack services, then we can >> reconsider this position. >> >> > >> >> >> >> Hi Jesse, >> >> >> >> thanks for replying here too. >> >> >> >> As shaped by the discussions yesterday and for the benefit of others - >> >> our current plan is to keep the c8 undercloud-ffu job and in fact that >> >> will be the *only* job we will keep on C8 (if standalone ffu upgrade >> >> becomes a thing then that will also be 8 only). At least, we'll make a >> >> best effort to keep this and we can re-evaluate once we hit blocking >> >> issues. >> >> >> >> In order to maintain this C8 gate we'll also need to maintain the >> >> equivalent periodic version for the integration line. So we'll have a >> >> C8 integration line running the undercloud-ffu job and possibly one >> >> ovb job (TBD, likely featureset 1). >> >> >> >> To be clear, our plan is to remove *all* centos8 check/gate jobs, >> >> except the undercloud-ffu. We are not planning to keep any 'base' set >> >> of c8 jobs unless someone can make a case for their value. >> >> >> >> The tripleo-ci team will start implementing the removal of c8/wallaby >> >> jobs once we switch to use c9 for the imports which should happen >> >> within the next week or so. >> >> >> >> regards, marios >> >> >> >> > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is >> 8 >> >> > > but this will soon switch. >> >> > > >> >> > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla >> standalone >> >> > > (no scenarios? some subset of them?), undercloud-ffu, minor update. >> >> > > >> >> > > This is just to start the conversation so please reply if you have >> >> > > thoughts or comments about any of the above. >> >> > > >> >> > > We are planning to discuss this in the coming tripleo-ci community >> >> > > call this coming Tuesday at 1330 UTC - meeting link at [4] so >> please >> >> > > join us if you can and would like to participate, >> >> > > >> >> > > regards, marios >> >> > > >> >> > > [1] https://review.opendev.org/q/topic:c8_teardown_master >> >> > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates >> >> > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 >> >> > > [4] https://meet.google.com/bqx-xwht-wky >> >> > > >> >> > > >> >> > >> >> >> >> >> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: From marios at redhat.com Wed Mar 2 13:20:47 2022 From: marios at redhat.com (Marios Andreou) Date: Wed, 2 Mar 2022 15:20:47 +0200 Subject: [TripleO] centos9 jobs only for master, centos 8 & 9 for wallaby In-Reply-To: References: <32D116CA-5034-4557-BEDF-6D7E8C7FDD48@odyssey4.me> Message-ID: On Wed, Mar 2, 2022 at 3:09 PM John Fulton wrote: > > Just wanted to add that for Ceph integration in TripleO, these are the important Ceph jobs which run when THT is patched in the main branch: > > tripleo-ci-centos-9-scenario001-standalone > tripleo-ci-centos-9-scenario004-standalone > > If the Wallaby branch only had centos-9 versions of the above and no centos-8 versions of the above, then I don't see that as a problem. > ack thanks for joining the tripleo-ci community call yesterday and for confirming here Yes we will continue to have the centos-9 versions of those jobs triggering as before (i.e. same files: matches etc) regards, marios > John > > On Wed, Mar 2, 2022 at 3:08 AM Jiri Podivin wrote: >> >> Great, >> Thanks for the clarification, it had me worried for a bit. >> >> On Wed, Mar 2, 2022 at 9:05 AM Marios Andreou wrote: >>> >>> On Wed, Mar 2, 2022 at 9:44 AM Jiri Podivin wrote: >>> > >>> > Hi, >>> > Sorry I'm replying this late, but this line raised some questions for me: >>> > > To be clear, our plan is to remove *all* centos8 check/gate jobs, >>> > >>> > It seems to imply that all centos8 jobs in both pipelines will be removed, across stable branches. >>> > Does this mean up to and including train? >>> > It was my understanding that everything older than wallaby will remain as it is. >>> > >>> >>> This discussion is exclusively about wallaby - older stable branches >>> will be unaffected. >>> >>> >>> >>> >>> >>> > >>> > On Wed, Mar 2, 2022 at 8:04 AM Marios Andreou wrote: >>> >> >>> >> On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius wrote: >>> >> > >>> >> > Hi folks, response in-line. >>> >> > >>> >> > > On 24 Feb 2022, at 15:11, Marios Andreou wrote: >>> >> > > >>> >> > > Hello TripleO o/ >>> >> > > >>> >> > > During the last period the tripleo-ci team has been working on getting >>> >> > > tripleo CI jobs to run centos-9. >>> >> > > >>> >> > > On master branches across all TripleO repos we now run C9 exclusively >>> >> > > (since [1]). >>> >> > > >>> >> > > On Wallaby branches, after the work at [2] (in particular after >>> >> > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 merges) we >>> >> > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking >>> >> > > this would be a PTG topic, but I think we'll want to have this >>> >> > > conversation way before then. There is a tripleo-heat-templates >>> >> > > example in the etherpad that shows what we can expect from the job >>> >> > > runs (note the exact jobs that run will vary per repo and even between >>> >> > > changes depending on the files touched - but generally there will be >>> >> > > duplication between 8 and 9 jobs). >>> >> > > >>> >> > > The current proposal is that we keep a minimal subset on C8 Wallaby, >>> >> > > with C9 wallaby having the full set of jobs. >>> >> > > >>> >> > > Two factors that will affect our decisions are (there may be more?) >>> >> > > >>> >> > > i) Upgrades requirements - are we supporting upgrading to Wallaby on >>> >> > > 8? For example, the coming undercloud-upgrade-ffu job will be train 8 >>> >> > > to wallaby 8. In which case we definitely need to keep at least some >>> >> > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby >>> >> > > completely). >>> >> > >>> >> > I think trying to keep up with the moving target of 8-stream would be counterproductive given that we are pinned downstream to RHEL 8.4 for Train. We do have a pending patch to merge a job for 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be nice to keep but if it becomes troublesome then I?d suggest we remove it. This job adds some value to us (and has added quite a bit already in preparing it and making it pass), but if 8-stream starts making it fail, or it starts failing due to database migrations then we?ll have to shift that job downstream and deal with the consequences of only ever finding upgrade bugs post-merge. >>> >> > >>> >> > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack services is not supported upstream, so we can not expect any upstream jobs to reliability support this process. If/when there is broader support for enabling FFU support officially in OpenStack services, then we can reconsider this position. >>> >> > >>> >> >>> >> Hi Jesse, >>> >> >>> >> thanks for replying here too. >>> >> >>> >> As shaped by the discussions yesterday and for the benefit of others - >>> >> our current plan is to keep the c8 undercloud-ffu job and in fact that >>> >> will be the *only* job we will keep on C8 (if standalone ffu upgrade >>> >> becomes a thing then that will also be 8 only). At least, we'll make a >>> >> best effort to keep this and we can re-evaluate once we hit blocking >>> >> issues. >>> >> >>> >> In order to maintain this C8 gate we'll also need to maintain the >>> >> equivalent periodic version for the integration line. So we'll have a >>> >> C8 integration line running the undercloud-ffu job and possibly one >>> >> ovb job (TBD, likely featureset 1). >>> >> >>> >> To be clear, our plan is to remove *all* centos8 check/gate jobs, >>> >> except the undercloud-ffu. We are not planning to keep any 'base' set >>> >> of c8 jobs unless someone can make a case for their value. >>> >> >>> >> The tripleo-ci team will start implementing the removal of c8/wallaby >>> >> jobs once we switch to use c9 for the imports which should happen >>> >> within the next week or so. >>> >> >>> >> regards, marios >>> >> >>> >> > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is 8 >>> >> > > but this will soon switch. >>> >> > > >>> >> > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla standalone >>> >> > > (no scenarios? some subset of them?), undercloud-ffu, minor update. >>> >> > > >>> >> > > This is just to start the conversation so please reply if you have >>> >> > > thoughts or comments about any of the above. >>> >> > > >>> >> > > We are planning to discuss this in the coming tripleo-ci community >>> >> > > call this coming Tuesday at 1330 UTC - meeting link at [4] so please >>> >> > > join us if you can and would like to participate, >>> >> > > >>> >> > > regards, marios >>> >> > > >>> >> > > [1] https://review.opendev.org/q/topic:c8_teardown_master >>> >> > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates >>> >> > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 >>> >> > > [4] https://meet.google.com/bqx-xwht-wky >>> >> > > >>> >> > > >>> >> > >>> >> >>> >> >>> >>> On Wed, Mar 2, 2022 at 9:44 AM Jiri Podivin wrote: >>> > >>> > Hi, >>> > Sorry I'm replying this late, but this line raised some questions for me: >>> > > To be clear, our plan is to remove *all* centos8 check/gate jobs, >>> > >>> > It seems to imply that all centos8 jobs in both pipelines will be removed, across stable branches. >>> > Does this mean up to and including train? >>> > It was my understanding that everything older than wallaby will remain as it is. >>> > >>> > >>> > On Wed, Mar 2, 2022 at 8:04 AM Marios Andreou wrote: >>> >> >>> >> On Tue, Mar 1, 2022 at 5:03 PM Jesse Pretorius wrote: >>> >> > >>> >> > Hi folks, response in-line. >>> >> > >>> >> > > On 24 Feb 2022, at 15:11, Marios Andreou wrote: >>> >> > > >>> >> > > Hello TripleO o/ >>> >> > > >>> >> > > During the last period the tripleo-ci team has been working on getting >>> >> > > tripleo CI jobs to run centos-9. >>> >> > > >>> >> > > On master branches across all TripleO repos we now run C9 exclusively >>> >> > > (since [1]). >>> >> > > >>> >> > > On Wallaby branches, after the work at [2] (in particular after >>> >> > > https://review.opendev.org/c/openstack/tripleo-ci/+/830132 merges) we >>> >> > > will run both C8 and C9 jobs. I started an etherpad at [3] thinking >>> >> > > this would be a PTG topic, but I think we'll want to have this >>> >> > > conversation way before then. There is a tripleo-heat-templates >>> >> > > example in the etherpad that shows what we can expect from the job >>> >> > > runs (note the exact jobs that run will vary per repo and even between >>> >> > > changes depending on the files touched - but generally there will be >>> >> > > duplication between 8 and 9 jobs). >>> >> > > >>> >> > > The current proposal is that we keep a minimal subset on C8 Wallaby, >>> >> > > with C9 wallaby having the full set of jobs. >>> >> > > >>> >> > > Two factors that will affect our decisions are (there may be more?) >>> >> > > >>> >> > > i) Upgrades requirements - are we supporting upgrading to Wallaby on >>> >> > > 8? For example, the coming undercloud-upgrade-ffu job will be train 8 >>> >> > > to wallaby 8. In which case we definitely need to keep at least some >>> >> > > subset of 8 jobs (and can't entertain the removal of 8 from Wallaby >>> >> > > completely). >>> >> > >>> >> > I think trying to keep up with the moving target of 8-stream would be counterproductive given that we are pinned downstream to RHEL 8.4 for Train. We do have a pending patch to merge a job for 8-Stream/Train->8-Stream Wallaby for the Undercloud only, which would be nice to keep but if it becomes troublesome then I?d suggest we remove it. This job adds some value to us (and has added quite a bit already in preparing it and making it pass), but if 8-stream starts making it fail, or it starts failing due to database migrations then we?ll have to shift that job downstream and deal with the consequences of only ever finding upgrade bugs post-merge. >>> >> > >>> >> > It is well known that a Fast-Forward Upgrade (FFU) of OpenStack services is not supported upstream, so we can not expect any upstream jobs to reliability support this process. If/when there is broader support for enabling FFU support officially in OpenStack services, then we can reconsider this position. >>> >> > >>> >> >>> >> Hi Jesse, >>> >> >>> >> thanks for replying here too. >>> >> >>> >> As shaped by the discussions yesterday and for the benefit of others - >>> >> our current plan is to keep the c8 undercloud-ffu job and in fact that >>> >> will be the *only* job we will keep on C8 (if standalone ffu upgrade >>> >> becomes a thing then that will also be 8 only). At least, we'll make a >>> >> best effort to keep this and we can re-evaluate once we hit blocking >>> >> issues. >>> >> >>> >> In order to maintain this C8 gate we'll also need to maintain the >>> >> equivalent periodic version for the integration line. So we'll have a >>> >> C8 integration line running the undercloud-ffu job and possibly one >>> >> ovb job (TBD, likely featureset 1). >>> >> >>> >> To be clear, our plan is to remove *all* centos8 check/gate jobs, >>> >> except the undercloud-ffu. We are not planning to keep any 'base' set >>> >> of c8 jobs unless someone can make a case for their value. >>> >> >>> >> The tripleo-ci team will start implementing the removal of c8/wallaby >>> >> jobs once we switch to use c9 for the imports which should happen >>> >> within the next week or so. >>> >> >>> >> regards, marios >>> >> >>> >> > > ii) Are we importing from Wallaby 8 or Wallaby 9? Currently it is 8 >>> >> > > but this will soon switch. >>> >> > > >>> >> > > For the wallaby c8 'subset of jobs' e.g. multinode, vanilla standalone >>> >> > > (no scenarios? some subset of them?), undercloud-ffu, minor update. >>> >> > > >>> >> > > This is just to start the conversation so please reply if you have >>> >> > > thoughts or comments about any of the above. >>> >> > > >>> >> > > We are planning to discuss this in the coming tripleo-ci community >>> >> > > call this coming Tuesday at 1330 UTC - meeting link at [4] so please >>> >> > > join us if you can and would like to participate, >>> >> > > >>> >> > > regards, marios >>> >> > > >>> >> > > [1] https://review.opendev.org/q/topic:c8_teardown_master >>> >> > > [2] https://review.opendev.org/q/topic:c9_wallaby_gates >>> >> > > [3] https://etherpad.opendev.org/p/tripleoci-wallaby-centos-8-9 >>> >> > > [4] https://meet.google.com/bqx-xwht-wky >>> >> > > >>> >> > > >>> >> > >>> >> >>> >> >>> From senrique at redhat.com Wed Mar 2 13:39:52 2022 From: senrique at redhat.com (Sofia Enriquez) Date: Wed, 2 Mar 2022 10:39:52 -0300 Subject: [cinder] Bug deputy report for week of 03-02-2022 Message-ID: This is a bug report from 02-23-2022 to 03-02-2022. Agenda: https://etherpad.opendev.org/p/cinder-bug-squad-meeting ----------------------------------------------------------------------------------------- Low - https://bugs.launchpad.net/cinder/+bug/1962221 "[lightos] Compression handling in lightos driver while default compression enabled." Fix proposed to master. Incomplete - https://bugs.launchpad.net/cinder/+bug/1962607 "IndexError when creating Backups." Fix proposed to master. Cheers -- Sof?a Enriquez she/her Software Engineer Red Hat PnT IRC: @enriquetaso @RedHat Red Hat Red Hat -------------- next part -------------- An HTML attachment was scrubbed... URL: From elod.illes at est.tech Wed Mar 2 14:50:51 2022 From: elod.illes at est.tech (=?UTF-8?B?RWzFkWQgSWxsw6lz?=) Date: Wed, 2 Mar 2022 15:50:51 +0100 Subject: [QA][OpenStackSDK][release] Cycle With Intermediary without recent deliverables Message-ID: Hi, Quick reminder that for deliverables following the cycle-with-intermediary model, the release team will use the latest Yoga release available on release week. The following deliverables have done a Yoga release, but it was not refreshed in the last two months: tempest patrole (this was not yet released at all in Yoga) python-openstackclient (I remember that openstacksdk cannot do another release in Yoga due to issues on master branch. Is this also valid for python-openstackclient?) You should consider making a new one very soon, so that we don't use an outdated version for the final release. El?d Ill?s irc: elodilles -------------- next part -------------- An HTML attachment was scrubbed... URL: From beagles at redhat.com Wed Mar 2 14:59:18 2022 From: beagles at redhat.com (Brent Eagles) Date: Wed, 2 Mar 2022 11:29:18 -0330 Subject: [tripleo] Dynamically generating firewall rules for the haproxy host on behalf of a deployed service Message-ID: Hi all, Designate has a requirement that appears to be a first for TripleO. It needs to deploy miniDNS instances on the internal network but each are accessible from the public network so external bind instances can sync with them. This is done by mapping port numbers on the VIP to each miniDNS instances [1] e.g. 10.0.0.5:16000 --> 172.16.2.184:5354 10.0.0.5:16001 --> 172.16.2.185:5354 10.0.0.5:16002 --> 172.16.2.186:5354 . . . The haproxy configuration is relatively straightforward [2]. There would be no problem if it the haproxy puppet managed the firewall rules, but for typical deployments we appear to disable this [3] and rely on the tripleo.firewall ansible module instead. The complicating factors are: a. the ports range depends on the number of miniDNS instances. b. the miniDNS deployment needs to set firewall rules wherever haproxy happens to live as they are not guaranteed to be on the same host. AFAICT, our ansible based firewall rule mechanism doesn't allow for setting firewall rules on hosts in addition to the ones that the service is being deployed on. Possible solutions I've been considering: a. Have haproxy puppet create firewall rules on the public API network and use tripleo.firewall for everything else. I think we generally don't distinguish which networks firewall rules are applicable to so this would be an "interesting", pervasive change. This is in PID 1's wheelhouse and I'd be interested in how they feel about this. b. Somehow create a jinja-ed firewall rule for haproxy that gets processed at deploy time so details could be filled in using cloud information. This seems like it would be kind of fragile, may not even be possible and depends on details of the framework I'm not familiar with. c. Extend tripleo.firewall or tripleo.iptables to process jinja'd rules - but that's problematic because it needs to be set on the haproxy host, and miniDNS is elsewhere. d. Create some ansible that gets run in the haproxy hosts that generates the firewall rules based on cloud info. I feel this is a bit ugly. Does anyone have any suggestions on other solutions or thoughts on the above? Cheers, Brent 1. https://bugzilla.redhat.com/show_bug.cgi?id=1897319 The TripleO HAProxy instance needs to be configured for BIND->miniDNS AXFR 2. https://review.opendev.org/c/openstack/puppet-tripleo/+/828169 Designate: create proxy ports on external VIP to access miniDNS workers. 3. https://bugs.launchpad.net/tripleo/+bug/1961799 haproxy template tasks to apply IPTables rules are no ops. -- Brent Eagles Principal Software Engineer Red Hat Inc. From kecarter at redhat.com Wed Mar 2 15:41:38 2022 From: kecarter at redhat.com (Kevin Carter) Date: Wed, 2 Mar 2022 09:41:38 -0600 Subject: [tripleo] Dynamically generating firewall rules for the haproxy host on behalf of a deployed service In-Reply-To: References: Message-ID: On Wed, Mar 2, 2022 at 8:59 AM Brent Eagles wrote: > Hi all, > > Designate has a requirement that appears to be a first for TripleO. It > needs to deploy miniDNS instances on the internal network but each are > accessible from the public network so external bind instances can sync > with them. This is done by mapping port numbers on the VIP to each > miniDNS instances [1] e.g. > > 10.0.0.5:16000 --> 172.16.2.184:5354 > 10.0.0.5:16001 --> 172.16.2.185:5354 > 10.0.0.5:16002 --> 172.16.2.186:5354 > . > . > . > > > The haproxy configuration is relatively straightforward [2]. There > would be no problem if it the haproxy puppet managed the firewall rules, > but for typical deployments we appear to disable this [3] and rely on > the tripleo.firewall ansible module instead. > > > The complicating factors are: > > a. the ports range depends on the number of miniDNS instances. > > b. the miniDNS deployment needs to set firewall rules wherever haproxy > happens to live as they are not guaranteed to be on the same host. > AFAICT, our ansible based firewall rule mechanism doesn't allow for > setting firewall rules on hosts in addition to the ones that the service > is being deployed on. > > > Possible solutions I've been considering: > > a. Have haproxy puppet create firewall rules on the public API network > and use tripleo.firewall for everything else. I think we generally don't > distinguish which networks firewall rules are applicable to so this > would be an "interesting", pervasive change. This is in PID 1's > wheelhouse and I'd be interested in how they feel about this. > > This seems like the easiest path forward. That said, within the DF we're actively attempting to reduce our dependence on puppet. As such, while the other options may be ugly or even more complex, I think they're worth pursuing especially if it means we have less puppet to maintain. > b. Somehow create a jinja-ed firewall rule for haproxy that gets > processed at deploy time so details could be filled in using cloud > information. This seems like it would be kind of fragile, may not even > be possible and depends on details of the framework I'm not familiar > with. > > c. Extend tripleo.firewall or tripleo.iptables to process jinja'd rules > - but that's problematic because it needs to be set on the haproxy host, > and miniDNS is elsewhere. > > d. Create some ansible that gets run in the haproxy hosts that > generates the firewall rules based on cloud info. I feel this is a bit > ugly. > > > Does anyone have any suggestions on other solutions or thoughts on the > above? > > We make heavy use of YAQL to organize our firewall rules for tripleo.firewall, is there something we can do there? We could also transition some of our firewall setup to use hostvars / facts across the deployment and to pull back information for specific nodes, or all nodes within a given group; maybe we can make better use of that within the firewall role to pull back runtime information and compute the needed firewall rules for minidns? > Cheers, > > Brent > > > 1. https://bugzilla.redhat.com/show_bug.cgi?id=1897319 The TripleO > HAProxy instance needs to be configured for BIND->miniDNS AXFR > > 2. https://review.opendev.org/c/openstack/puppet-tripleo/+/828169 > Designate: create proxy ports on external VIP to access miniDNS workers. > > 3. https://bugs.launchpad.net/tripleo/+bug/1961799 haproxy template > tasks to apply IPTables rules are no ops. > > -- > Brent Eagles > Principal Software Engineer > Red Hat Inc. > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From thierry at openstack.org Wed Mar 2 15:49:26 2022 From: thierry at openstack.org (Thierry Carrez) Date: Wed, 2 Mar 2022 16:49:26 +0100 Subject: [largescale-sig] Next meeting: Mar 2nd, 15utc In-Reply-To: <7e4962b0-b479-9ed7-5d69-72a68bcf7a73@openstack.org> References: <7e4962b0-b479-9ed7-5d69-72a68bcf7a73@openstack.org> Message-ID: <9615cc25-7fc8-eec6-37e3-0036bf137cbc@openstack.org> Our meeting just concluded. We discussed the special guest for our next "Large Scale OpenStack" episode on OpenInfra.Live, but also our participation to the PTG in April. In preparation for that, between now and next IRC meeting we will collect topics for the PTG meeting on: https://etherpad.opendev.org/p/large-scale-sig-meeting You can read the meeting logs at: https://meetings.opendev.org/meetings/large_scale_sig/2022/large_scale_sig.2022-03-02-15.09.html Our next IRC meeting will be March 16, at 1500utc on #openstack-operators on OFTC. We will finalize our guest speaker for OILive and (hopefully) the contents of our PTG session. Regards, -- Thierry Carrez (ttx) From tkajinam at redhat.com Wed Mar 2 16:00:15 2022 From: tkajinam at redhat.com (Takashi Kajinami) Date: Thu, 3 Mar 2022 01:00:15 +0900 Subject: [tripleo] Dynamically generating firewall rules for the haproxy host on behalf of a deployed service In-Reply-To: References: Message-ID: I agree with Kevin. What I'd like to add here is that enabling haproxy firewall management by puppet means we have both ansible and puppet to manage firewall rules, and this is likely to cause undesired conjunction in the future. IMO it's better to find out the way to facilitate the requirement within tripleo-ansible(*1). Note that this issue is not specific to Designate and can affect any deployment which has separate nodes for api servers and haproxy, as is mentioned in the bug. I tend to think of a mixed approach of b and c, and gather frontend firewall rules for all services enabled in all roles by yaql, and then put that data into tripleo-firewall, then use that data only if haproxy is enabled in the node. I started drafting the change, but it's too late in my tz to implement such composition beyond roles ... https://review.opendev.org/q/topic:bug%252F1961799 (I hesitate to share this because it implements almost nothing but hope this explains my thought well) (*1) By the way it reminds me that we still require puppet to set up NAT rules, which we also need to deal with, later, as a completely separate topic. On Thu, Mar 3, 2022 at 12:43 AM Kevin Carter wrote: > On Wed, Mar 2, 2022 at 8:59 AM Brent Eagles wrote: > >> Hi all, >> >> Designate has a requirement that appears to be a first for TripleO. It >> needs to deploy miniDNS instances on the internal network but each are >> accessible from the public network so external bind instances can sync >> with them. This is done by mapping port numbers on the VIP to each >> miniDNS instances [1] e.g. >> >> 10.0.0.5:16000 --> 172.16.2.184:5354 >> 10.0.0.5:16001 --> 172.16.2.185:5354 >> 10.0.0.5:16002 --> 172.16.2.186:5354 >> . >> . >> . >> >> >> The haproxy configuration is relatively straightforward [2]. There >> would be no problem if it the haproxy puppet managed the firewall rules, >> but for typical deployments we appear to disable this [3] and rely on >> the tripleo.firewall ansible module instead. >> >> >> The complicating factors are: >> >> a. the ports range depends on the number of miniDNS instances. >> >> b. the miniDNS deployment needs to set firewall rules wherever haproxy >> happens to live as they are not guaranteed to be on the same host. >> AFAICT, our ansible based firewall rule mechanism doesn't allow for >> setting firewall rules on hosts in addition to the ones that the service >> is being deployed on. >> >> >> Possible solutions I've been considering: >> >> a. Have haproxy puppet create firewall rules on the public API network >> and use tripleo.firewall for everything else. I think we generally don't >> distinguish which networks firewall rules are applicable to so this >> would be an "interesting", pervasive change. This is in PID 1's >> wheelhouse and I'd be interested in how they feel about this. >> >> This seems like the easiest path forward. That said, within the DF we're > actively attempting to reduce our dependence on puppet. As such, while the > other options may be ugly or even more complex, I think they're worth > pursuing especially if it means we have less puppet to maintain. > > > >> b. Somehow create a jinja-ed firewall rule for haproxy that gets >> processed at deploy time so details could be filled in using cloud >> information. This seems like it would be kind of fragile, may not even >> be possible and depends on details of the framework I'm not familiar >> with. >> >> c. Extend tripleo.firewall or tripleo.iptables to process jinja'd rules >> - but that's problematic because it needs to be set on the haproxy host, >> and miniDNS is elsewhere. >> >> d. Create some ansible that gets run in the haproxy hosts that >> generates the firewall rules based on cloud info. I feel this is a bit >> ugly. >> >> >> Does anyone have any suggestions on other solutions or thoughts on the >> above? >> >> We make heavy use of YAQL to organize our firewall rules for > tripleo.firewall, is there something we can do there? We could also > transition some of our firewall setup to use hostvars / facts across the > deployment and to pull back information for specific nodes, or all nodes > within a given group; maybe we can make better use of that within the > firewall role to pull back runtime information and compute the needed > firewall rules for minidns? > > > >> Cheers, >> >> Brent >> >> >> 1. https://bugzilla.redhat.com/show_bug.cgi?id=1897319 The TripleO >> HAProxy instance needs to be configured for BIND->miniDNS AXFR >> >> 2. https://review.opendev.org/c/openstack/puppet-tripleo/+/828169 >> Designate: create proxy ports on external VIP to access miniDNS workers. >> >> 3. https://bugs.launchpad.net/tripleo/+bug/1961799 haproxy template >> tasks to apply IPTables rules are no ops. >> >> -- >> Brent Eagles >> Principal Software Engineer >> Red Hat Inc. >> >> -------------- next part -------------- An HTML attachment was scrubbed... URL: From johnsomor at gmail.com Wed Mar 2 16:59:37 2022 From: johnsomor at gmail.com (Michael Johnson) Date: Wed, 2 Mar 2022 08:59:37 -0800 Subject: [designate] Zed PTG planning etherpad Message-ID: Hi Designate community, In preparation for our PTG sessions in April I have created an etherpad for us to add topics for discussion. https://etherpad.opendev.org/p/zed-ptg-designate Please add your attendance and topics. Michael From iurygregory at gmail.com Wed Mar 2 17:29:43 2022 From: iurygregory at gmail.com (Iury Gregory) Date: Wed, 2 Mar 2022 14:29:43 -0300 Subject: [ironic] Zed PTG etherpad Message-ID: Hello ironicers! I've created an etherpad to collect topics for our Zed PTG [1] Feel free to add any topics on it =) Later tomorrow I will be sharing the ideas for possible PTG slots and in the upstream meeting on Monday (March 7) we will announce. [1] https://etherpad.opendev.org/p/ironic-zed-ptg -- *Att[]'sIury Gregory Melo Ferreira * *MSc in Computer Science at UFCG* *Part of the ironic-core and puppet-manager-core team in OpenStack* *Software Engineer at Red Hat Czech* *Social*: https://www.linkedin.com/in/iurygregory *E-mail: iurygregory at gmail.com * -------------- next part -------------- An HTML attachment was scrubbed... URL: From gmann at ghanshyammann.com Wed Mar 2 18:14:04 2022 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Wed, 02 Mar 2022 12:14:04 -0600 Subject: [QA][OpenStackSDK][release] Cycle With Intermediary without recent deliverables In-Reply-To: References: Message-ID: <17f4bd6d4b0.b231163971144.6913564304480452809@ghanshyammann.com> ---- On Wed, 02 Mar 2022 08:50:51 -0600 El?d Ill?s wrote ---- > Hi,Quick reminder that for deliverables following the cycle-with-intermediarymodel, the release team will use the latest Yoga release available onrelease week.The following deliverables have done a Yoga release, but it was notrefreshed in the last two months: tempest patrole (this was not yet released at all in Yoga) python-openstackclient (I remember that openstacksdk cannot do another release in Yoga due to issues on master branch. Is this also valid for python-openstackclient?)You should consider making a new one very soon, so that we don't use anoutdated version for the final release.El?d Ill?sirc: elodilles Thanks for reminder, I have proposed the Patrole release - https://review.opendev.org/c/openstack/releases/+/831565 -gmann > > > > From adivya1.singh at gmail.com Wed Mar 2 20:52:41 2022 From: adivya1.singh at gmail.com (Adivya Singh) Date: Thu, 3 Mar 2022 02:22:41 +0530 Subject: Neutron issue in current Xena Version of openstack Message-ID: hi Team, i am facing issue after installing XENA version of Openstack, where in the Same hosted Network node, some routers work, and some not, The real issue is the router which are not working, All floating IP assigned to the router does not work, but other routers hosted on same L3 agent Works I am seeing this king of error " "Could not generate persistent MAC: No data" , but i am not sure this is impacting the issue Regards Adivya Singh -------------- next part -------------- An HTML attachment was scrubbed... URL: From dsmigiel at redhat.com Wed Mar 2 22:41:08 2022 From: dsmigiel at redhat.com (Dariusz Smigiel) Date: Wed, 2 Mar 2022 14:41:08 -0800 Subject: [TripleO] Gate blocker for Centos 9 Wallaby Message-ID: Currently TripleO gate is experiencing issues with building images [1] We're aware of the problem, and we're trying to resolve this issue. Please withhold from rechecking RDO Third Party CI. Thank you, Dariusz Smigiel (dasm) [1]: https://bugs.launchpad.net/tripleo/+bug/1962786 From zbitter at redhat.com Thu Mar 3 03:03:57 2022 From: zbitter at redhat.com (Zane Bitter) Date: Wed, 2 Mar 2022 22:03:57 -0500 Subject: [election][heat][ptl] PTL non-candidacy In-Reply-To: References: Message-ID: On 8/02/22 00:50, Rico Lin wrote: > Hi all, > > As the PTL election will soon start, I would like to share my statement > on not planning?to run another term?of Heat PTL. And Instead, I > encourage anyone (core reviewer or not) who is interested to put their > name on. > I will definitely still stay around and help with reviews and patches. Hey Rico, Thanks so much for all your work as PTL over many years now. cheers, Zane. From gmann at ghanshyammann.com Thu Mar 3 03:36:22 2022 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Wed, 02 Mar 2022 21:36:22 -0600 Subject: [all][tc] Technical Committee next weekly meeting on Mar 3rd (not 4th) at 1500 UTC In-Reply-To: <17f42dbbe35.e38a85b0295421.3903320940743592043@ghanshyammann.com> References: <17f413dc13b.1112c1076280164.9078145143439661308@ghanshyammann.com> <17f42dbbe35.e38a85b0295421.3903320940743592043@ghanshyammann.com> Message-ID: <17f4dd9a157.12637a06483289.7671218897587973192@ghanshyammann.com> Hello Everyone, Below is the agenda for Tomorrow's TC IRC meeting schedule at 1500 UTC. https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting == Agenda for tomorrow's TC meeting == * Roll call * Follow up on past action items * Welcome new TC members ** http://lists.openstack.org/pipermail/openstack-discuss/2022-February/027412.html * Weekly meeting time check ** https://doodle.com/meeting/organize/id/mep9mGma * Gate health check ** Fixing Zuul config error in OpenStack *** https://etherpad.opendev.org/p/zuul-config-error-openstack * Next step for release cadence ** https://governance.openstack.org/tc/resolutions/20220210-release-cadence-adjustment.html * Z cycle Leaderless projects ** https://etherpad.opendev.org/p/zed-leaderless * PTG Preparation ** http://lists.openstack.org/pipermail/openstack-discuss/2022-February/027381.html ** https://etherpad.opendev.org/p/tc-zed-ptg ** TC + Community Leaders Interaction *** http://lists.openstack.org/pipermail/openstack-discuss/2022-February/027382.html *Dropping tags framework - next steps (yoctozepto) * newsletter ** https://etherpad.opendev.org/p/newsletter-openstack-news * Open Reviews ** https://review.opendev.org/q/projects:openstack/governance+is:open -gmann ---- On Mon, 28 Feb 2022 18:22:51 -0600 Ghanshyam Mann wrote ---- > Sorry I mentioned Mar 4th as the next meeting but it is on Mar 3rd Thursday 15:00 UTC > > -gmann > > ---- On Mon, 28 Feb 2022 10:50:40 -0600 Ghanshyam Mann wrote ---- > > Hello Everyone, > > > > Technical Committee's next weekly meeting is scheduled for Mar 4th at 1500 UTC. > > > > If you would like to add topics for discussion, please add them to the below wiki page by > > Wednesday, Mar 3rd, at 2100 UTC. > > > > https://wiki.openstack.org/wiki/Meetings/TechnicalCommittee#Next_Meeting > > > > -gmann > > > > > > From satish.txt at gmail.com Thu Mar 3 03:47:08 2022 From: satish.txt at gmail.com (Satish Patel) Date: Wed, 2 Mar 2022 22:47:08 -0500 Subject: Openstack HPC storage suggestion Message-ID: Folks, I built a 30 node HPC environment on openstack using mellanox infiniband nic for high speed MPI messaging. So far everything works. Now I am looking for a HPC PFS (parallel file system) similar to Luster which I can mount on all HPC vms to run MPI jobs. I was reading on google and saw some CERN videos and they are using Ceph (cephfs) for PFS. Also i was reading CephFS + Manila is a good choice for HPC on openstack design. Does anyone have any experience with HPC storage for Openstack? Please advice or share your experience :) Thanks in advance. ~S -------------- next part -------------- An HTML attachment was scrubbed... URL: From vikarnatathe at gmail.com Thu Mar 3 04:07:00 2022 From: vikarnatathe at gmail.com (Vikarna Tathe) Date: Thu, 3 Mar 2022 09:37:00 +0530 Subject: nvidia gpu Message-ID: Hello All, Can somebody help me with the procedure for GPU integration with openstack wallaby, Ubuntu 20.04. Below are the card details: da:00.0 VGA compatible controller: NVIDIA Corporation TU102GL [Quadro RTX 6000/8000] Vikarna -------------- next part -------------- An HTML attachment was scrubbed... URL: From jerome.pansanel at iphc.cnrs.fr Thu Mar 3 08:00:49 2022 From: jerome.pansanel at iphc.cnrs.fr (Jerome Pansanel) Date: Thu, 3 Mar 2022 09:00:49 +0100 Subject: nvidia gpu In-Reply-To: References: Message-ID: Dear Virkana, There is an official documentation here: https://docs.openstack.org/project-deploy-guide/charm-deployment-guide/wallaby/app-pci-passthrough-gpu.html Here are additional links: * https://docs.openstack.org/project-deploy-guide/charm-deployment-guide/wallaby/app-pci-passthrough-gpu.html * https://gist.github.com/claudiok/890ab6dfe76fa45b30081e58038a9215 Best, Jerome Le 03/03/2022 ? 05:07, Vikarna Tathe a ?crit?: > Hello All, > > Can somebody help me with the procedure for GPU integration with > openstack wallaby, Ubuntu 20.04. > > Below are the card details: > da:00.0 VGA compatible controller: NVIDIA Corporation TU102GL [Quadro > RTX 6000/8000] > > Vikarna -- Jerome Pansanel, PhD Technical Director at IPHC Technical Coordinator at France Grilles Head of Scientific Cloud Infrastructure (SCIGNE) at IPHC IPHC || GSM: +33 (0)6 25 19 24 43 23 rue du Loess, BP 28 || Tel: +33 (0)3 88 10 66 24 F-67037 STRASBOURG Cedex 2 || Fax: +33 (0)3 88 10 62 34 From rdhasman at redhat.com Thu Mar 3 08:11:02 2022 From: rdhasman at redhat.com (Rajat Dhasmana) Date: Thu, 3 Mar 2022 13:41:02 +0530 Subject: [cinder] PTG Planning Etherpad Message-ID: Hello Argonauts, As you might be aware that the Zed PTG is approaching so I've created an etherpad for the details of the PTG and adding topics. To summarize, PTG will take place from 4-8 April, 2022 and below are the details for Cinder PTG: Date: 5-8 April, 2022 TIme: 1300-1700 UTC Connection info will be shared prior to the meeting on the etherpad and the cinder channel on IRC. Most probably it will be on BlueJeans (as has been in the past) but suggestions are welcome. Etherpad: https://etherpad.opendev.org/p/zed-ptg-cinder-planning Regarding the timings, as discussed in the cinder meeting twice, the attendees had no concern and it is the preferred time for the PTG since past 3 virtual PTGs. If there are any concerns, do let me know and we can also schedule your topic as convenient for your timezone. Lastly, I would encourage everyone to add topics as soon as they can since it makes the planning/arranging of topics easier and better. Thanks and regards Rajat Dhasmana -------------- next part -------------- An HTML attachment was scrubbed... URL: From mahendra.paipuri at cnrs.fr Thu Mar 3 08:23:29 2022 From: mahendra.paipuri at cnrs.fr (Mahendra Paipuri) Date: Thu, 3 Mar 2022 09:23:29 +0100 Subject: Openstack HPC storage suggestion In-Reply-To: References: Message-ID: <11fc1add-94f4-94a9-c47b-d50672708da0@cnrs.fr> Hello, We are quite interested in this too. When we were looking into existing solutions, we found there has been some work in integrating LUSTRE into OpenStack [1], [2]. I remember coming across some Openinfra talks on developing a manila backend driver for LUSTRE. I am not quite sure if this project is still ongoing. Manila already provides a backend driver for GPFS [3] (unfortunately GPFS is not opensource) to readily integrate it into Openstack. Manila supports GlusterFS and CephFS as well but they do not have RDMA support (if I am not wrong). This is pretty much what we found. Would be glad explore more solutions if someone knows any. Cheers - Mahendra [1] https://docs.google.com/presentation/d/1kGRzcdVQX95abei1bDVoRzxyC02i89_m5_sOfp8Aq6o/htmlpresent [2] https://www.openstack.org/videos/summits/barcelona-2016/lustre-integration-for-hpc-on-openstack-at-cambridge-and-monash [3] https://docs.openstack.org/manila/latest/admin/gpfs_driver.html On 03/03/2022 04:47, Satish Patel wrote: > Folks, > > I built a 30 node HPC environment on openstack using mellanox > infiniband nic for high speed MPI messaging. So far everything works. > Now I am looking for a HPC PFS (parallel file system)? similar to > Luster which I can mount on all HPC vms to run MPI jobs. > > I was reading on google and saw some CERN videos and they are using > Ceph (cephfs) for PFS. Also i was reading CephFS?+ Manila is a good > choice for HPC on openstack design. > > Does anyone have any experience with HPC storage for Openstack? Please > advice or share your experience :)? Thanks in advance. > > ~S From mark at stackhpc.com Thu Mar 3 09:11:24 2022 From: mark at stackhpc.com (Mark Goddard) Date: Thu, 3 Mar 2022 09:11:24 +0000 Subject: Openstack HPC storage suggestion In-Reply-To: <11fc1add-94f4-94a9-c47b-d50672708da0@cnrs.fr> References: <11fc1add-94f4-94a9-c47b-d50672708da0@cnrs.fr> Message-ID: I would recommend attending the Scientific SIG meetings if you are looking for people with this kind of information. https://wiki.openstack.org/wiki/Scientific_SIG Mark On Thu, 3 Mar 2022 at 08:28, Mahendra Paipuri wrote: > > Hello, > > We are quite interested in this too. When we were looking into existing > solutions, we found there has been some work in integrating LUSTRE into > OpenStack [1], [2]. I remember coming across some Openinfra talks on > developing a manila backend driver for LUSTRE. I am not quite sure if > this project is still ongoing. Manila already provides a backend driver > for GPFS [3] (unfortunately GPFS is not opensource) to readily integrate > it into Openstack. Manila supports GlusterFS and CephFS as well but they > do not have RDMA support (if I am not wrong). > > This is pretty much what we found. Would be glad explore more solutions > if someone knows any. > > Cheers > > - > > Mahendra > > [1] > https://docs.google.com/presentation/d/1kGRzcdVQX95abei1bDVoRzxyC02i89_m5_sOfp8Aq6o/htmlpresent > > [2] > https://www.openstack.org/videos/summits/barcelona-2016/lustre-integration-for-hpc-on-openstack-at-cambridge-and-monash > > [3] https://docs.openstack.org/manila/latest/admin/gpfs_driver.html > > On 03/03/2022 04:47, Satish Patel wrote: > > Folks, > > > > I built a 30 node HPC environment on openstack using mellanox > > infiniband nic for high speed MPI messaging. So far everything works. > > Now I am looking for a HPC PFS (parallel file system) similar to > > Luster which I can mount on all HPC vms to run MPI jobs. > > > > I was reading on google and saw some CERN videos and they are using > > Ceph (cephfs) for PFS. Also i was reading CephFS + Manila is a good > > choice for HPC on openstack design. > > > > Does anyone have any experience with HPC storage for Openstack? Please > > advice or share your experience :) Thanks in advance. > > > > ~S > From Molka.Gharbaoui at santannapisa.it Thu Mar 3 09:29:15 2022 From: Molka.Gharbaoui at santannapisa.it (Molka Gharbaoui) Date: Thu, 3 Mar 2022 09:29:15 +0000 Subject: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> Message-ID: Thank you Matthias for your reply. Unfortunately, from the logs I still cannot understand what is the issue. Could someone kindly indicate me how ceilometer and gnocchi should be configured to get the data? For example in this link: https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html the endpoints should use the following URL: http://controller:8041 while in my case the 8041 port is not even listening and the API indicated in the dashboard for Gnocchi is http://x.x.x.x/metric. Thank you in advance. ________________________________ Da: Matthias Runge Inviato: luned? 28 febbraio 2022 19:37 A: openstack-discuss at lists.openstack.org Oggetto: Re: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi On 2/28/22 17:58, Molka Gharbaoui wrote: > Hi all, > > I installed devstack and ceilometer by adding the following lines in the > local.conf file: > > enable_plugin ceilometer https://opendev.org/openstack/ceilometer.git > CEILOMETER_BACKEND=gnocchi > > Devstack (yoga) works correctly (I am able to create images, instances, > etc.). The problem is that I cannot visualize the metrics. Using > "gnocchi metric list" and "gnocchi resource list" commands the result is > empty. > > When I use the "--debug" option, no errors are plotted but the body > response is empty. > > "ceilometer-upgrade" gives the following error: DEBUG > ceilometer.cmd.storage Upgrade Gnocchi resource types upgrade > /opt/stack/ceilometer/ceilometer/cmd/storage.py:42 > > "ceilometer-polling" prints "The resource could not be found". > > I have a running instance. > > Could you help me in retrieving the metrics from openstack? That means, that ceilometer apparently either doesn't get any updates from the underlying services OR it doesn't send the metrics over to gnocchi. You should check the ceilometer logs for errors. Also gnocchi logs could reveal issues. Matthias -------------- next part -------------- An HTML attachment was scrubbed... URL: From zyklenfrei at gmail.com Thu Mar 3 09:30:52 2022 From: zyklenfrei at gmail.com (Manuel Holtgrewe) Date: Thu, 3 Mar 2022 10:30:52 +0100 Subject: Openstack HPC storage suggestion In-Reply-To: <11fc1add-94f4-94a9-c47b-d50672708da0@cnrs.fr> References: <11fc1add-94f4-94a9-c47b-d50672708da0@cnrs.fr> Message-ID: Hi, I guess it really depends on what HPC means to you ;-) Do your users schedule nova instances? Can you actually do Infiniband RDMA between nova VMs? Or are your users scheduling ironic instances via nova? We have an openstack setup based on kayobe/kolla where we: - Deploy 8 hyperconverged OpenStack nova libvirt compute+Ceph storage (mostly for block storage and cephfs comes in handy for shared file systems, e.g., for shared state in Slurm) servers. - Deploy 300+ bare metal compute nodes via ironic. - Use Slurm for the job scheduler. - Setup everything after the bare OS with Ansible. - Have login nodes and slurm scheduler etc. run in nova VMs. - Only have Ethernet interconnect (100/25GbE for central switches and recent servers and 40/10GbE for older servers). So we are using OpenStack nova+libvirt as a unified way of deploying virtual and physical machines and then install them as we would do with normal servers. That's a bit non-cloudy (as far as I have learned) but works really well for us. Now to your HPC storage question... that I hope to answer a bit indirectly. The main advantage for using manila with CephFS (that I can see) is that you get the openstack goodies of API and Horizon clients for managing shares. I guess this is mostly useful if you want to have cloud features for your HPC such as users allocating storage in self-service to their nova/ironic machines. We come from a classic HPC setting where the partitioning of the system is not done by creating multiple nodes/clusters by users in self-service, but rather administration provides a bare metal cluster with Slurm scheduler. Users log in to head nodes and submit jobs to the compute nodes to Slurm. Thus Slurm does the managing of resources and users can allocate single cores up to the whole cluster. So in our use case there would not be a major advantage of using manila for our storage as we would primarily have one export that gives access to the whole storage ;-). We currently have an old GPFS storage that we mount on all nodes via Ansible. We are currently migrating to using an additional, dedicated NVME-based ceph cluster (that is not hyperconverged with our compute) and that we would also mount via Ansible. As we essentially only have a single share on this storage managing the share with manila would be more trouble than work. The much more important part will be setting up Ceph appropriately and tuning it to perform well (e.g., using the IO500 benchmark as the croit people demonstrate here [1]). I guess that does not really answer your question but I hope that it gives useful perspective to you and maybe others. You can look at the work of the wonderful people of StackHPC that provide commercial services around Openstack/Kayobe/Kolla/Ironic for HPC setups. There are a couple of interesting videos involving their staff and they have very useful information on the web site as well. And as Mark Goddard just wrote you can have a look at the Scientific SIG where they are involved as well. (I'm not in any way related to StackHPC, I just really like their work). Best wishes, Manuel [1] https://croit.io/blog/ceph-performance-test-and-optimization On Thu, Mar 3, 2022 at 9:25 AM Mahendra Paipuri wrote: > > Hello, > > We are quite interested in this too. When we were looking into existing > solutions, we found there has been some work in integrating LUSTRE into > OpenStack [1], [2]. I remember coming across some Openinfra talks on > developing a manila backend driver for LUSTRE. I am not quite sure if > this project is still ongoing. Manila already provides a backend driver > for GPFS [3] (unfortunately GPFS is not opensource) to readily integrate > it into Openstack. Manila supports GlusterFS and CephFS as well but they > do not have RDMA support (if I am not wrong). > > This is pretty much what we found. Would be glad explore more solutions > if someone knows any. > > Cheers > > - > > Mahendra > > [1] > https://docs.google.com/presentation/d/1kGRzcdVQX95abei1bDVoRzxyC02i89_m5_sOfp8Aq6o/htmlpresent > > [2] > https://www.openstack.org/videos/summits/barcelona-2016/lustre-integration-for-hpc-on-openstack-at-cambridge-and-monash > > [3] https://docs.openstack.org/manila/latest/admin/gpfs_driver.html > > On 03/03/2022 04:47, Satish Patel wrote: > > Folks, > > > > I built a 30 node HPC environment on openstack using mellanox > > infiniband nic for high speed MPI messaging. So far everything works. > > Now I am looking for a HPC PFS (parallel file system) similar to > > Luster which I can mount on all HPC vms to run MPI jobs. > > > > I was reading on google and saw some CERN videos and they are using > > Ceph (cephfs) for PFS. Also i was reading CephFS + Manila is a good > > choice for HPC on openstack design. > > > > Does anyone have any experience with HPC storage for Openstack? Please > > advice or share your experience :) Thanks in advance. > > > > ~S > From rafaelweingartner at gmail.com Thu Mar 3 10:24:37 2022 From: rafaelweingartner at gmail.com (=?UTF-8?Q?Rafael_Weing=C3=A4rtner?=) Date: Thu, 3 Mar 2022 07:24:37 -0300 Subject: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> Message-ID: This situation can be caused by a lot of different things. - "gnocchi resource-type list" will show you the resource-types configured. Make sure you have configured the resource types that Ceilometer is pushing resources to - Make sure you have enabled metrics to be collected via polling.yml file - Make sure you have configured the "gnocchi_resources.yaml" file, and that the link between metrics, resources, and attributes to be pushed are correctly created - If want to generate metrics based on events, you then need to check : meters.yaml file, to see if everything is fine there - Make sure Ceilometer is pushing metrics and resource attributes to gnocchi. When the push is executed you will see the following DEBUG log messages: - Processing sample [%s] for resource ID [%s] - %d measures posted against %d metrics through %d resources - metric %s is not handled by Gnocchi (this one will appear if the metric is not configured to be pushed in the "gnocchi_resources.yaml") On Thu, Mar 3, 2022 at 6:33 AM Molka Gharbaoui < Molka.Gharbaoui at santannapisa.it> wrote: > Thank you Matthias for your reply. > > Unfortunately, from the logs I still cannot understand what is the issue. > > Could someone kindly indicate me how ceilometer and gnocchi should be > configured to get the data? > > For example in this link: > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html the > endpoints should use the following URL: http://controller:8041 while in > my case the 8041 port is not even listening and the API indicated in the > dashboard for Gnocchi is http:// x.x.x.x/metric. > > Thank you in advance. > > > > ------------------------------ > *Da:* Matthias Runge > *Inviato:* luned? 28 febbraio 2022 19:37 > *A:* openstack-discuss at lists.openstack.org < > openstack-discuss at lists.openstack.org> > *Oggetto:* Re: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > On 2/28/22 17:58, Molka Gharbaoui wrote: > > Hi all, > > > > I installed devstack and ceilometer by adding the following lines in the > > local.conf file: > > > > enable_plugin ceilometer https://opendev.org/openstack/ceilometer.git > > CEILOMETER_BACKEND=gnocchi > > > > Devstack (yoga) works correctly (I am able to create images, instances, > > etc.). The problem is that I cannot visualize the metrics. Using > > "gnocchi metric list" and "gnocchi resource list" commands the result is > > empty. > > > > When I use the "--debug" option, no errors are plotted but the body > > response is empty. > > > > "ceilometer-upgrade" gives the following error: DEBUG > > ceilometer.cmd.storage Upgrade Gnocchi resource types upgrade > > /opt/stack/ceilometer/ceilometer/cmd/storage.py:42 > > > > "ceilometer-polling" prints "The resource could not be found". > > > > I have a running instance. > > > > Could you help me in retrieving the metrics from openstack? > > That means, that ceilometer apparently either doesn't get any updates > from the underlying services OR it doesn't send the metrics over to > gnocchi. > > You should check the ceilometer logs for errors. Also gnocchi logs could > reveal issues. > > Matthias > > > -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From dtantsur at redhat.com Thu Mar 3 10:26:54 2022 From: dtantsur at redhat.com (Dmitry Tantsur) Date: Thu, 3 Mar 2022 11:26:54 +0100 Subject: [hwsig] sushy-oem-idrac for Hardware Vendor SIG In-Reply-To: References: Message-ID: Hi, I'm not a hardware vendor SIG member, but with my ironic hat on, it does make sense. Dmitry On Fri, Feb 25, 2022 at 12:46 PM Jaunteva, Aija wrote: > Hi, > > > > I would like to propose moving sushy-oem-idrac [1] that is used in Ironic > project to Hardware Vendor SIG. sushy-oem-idrac is iDRAC specific OEM > extension of vendor-independent sushy [2] library and both use Redfish to > communicate with BMCs. > > > > Any feedback welcome. > > > > Regards, > > Aija > > > > [1] https://opendev.org/x/sushy-oem-idrac/ > > [2] https://opendev.org/openstack/sushy > > Internal Use - Confidential > -- Red Hat GmbH, https://de.redhat.com/ , Registered seat: Grasbrunn, Commercial register: Amtsgericht Muenchen, HRB 153243, Managing Directors: Charles Cachera, Brian Klemm, Laurie Krebs, Michael O'Neill -------------- next part -------------- An HTML attachment was scrubbed... URL: From katonalala at gmail.com Thu Mar 3 10:31:44 2022 From: katonalala at gmail.com (Lajos Katona) Date: Thu, 3 Mar 2022 11:31:44 +0100 Subject: Neutron issue in current Xena Version of openstack In-Reply-To: References: Message-ID: Hi, Could you give some more details how do you use Neutron, with OVN, or OVS? The error you pasted seems to be from syslog, am I right? Regards Lajos Katona (lajoskatona) Adivya Singh ezt ?rta (id?pont: 2022. m?rc. 2., Sze, 21:59): > hi Team, > > i am facing issue after installing XENA version of Openstack, where in the > Same hosted Network node, some routers work, and some not, The real issue > is the router which are not working, All floating IP assigned to the router > does not work, but other routers hosted on same L3 agent Works > > I am seeing this king of error " "Could not generate persistent MAC: No > data" , but i am not sure this is impacting the issue > > Regards > Adivya Singh > -------------- next part -------------- An HTML attachment was scrubbed... URL: From mrunge at matthias-runge.de Thu Mar 3 10:59:10 2022 From: mrunge at matthias-runge.de (Matthias Runge) Date: Thu, 3 Mar 2022 11:59:10 +0100 Subject: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> Message-ID: <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> On 3/3/22 10:29, Molka Gharbaoui wrote: > Thank you Matthias for your reply. > > Unfortunately, from the logs I still cannot understand what is the issue. > > Could someone kindly indicate me how ceilometer and gnocchi should be > configured to get the data? > > For example in this link: > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > ?the > endpoints should use the following URL: http://controller:8041 > ?while in my case the 8041 port is not even > listening and the API indicated in the dashboard for Gnocchi is http:// > x.x.x.x/metric. > > Thank you in advance. > OpenStack configuration is complex. The guide does not remotely touch on production environments. Please do not confuse devstack with production envionments. Also the doc you linked is not for devstack itself. The flow in general is: data is generated by the services like nova or glance. Ceilometer listens on the bus and also polls for information. The info is sent then over (as configured in publishers.yaml) to gnocchi. Gnocchi has an api, ceilometer has not. Reading your response, this makes me suspicious if there is gnocchi running. From the initial question it suggests it is, from your most recent response, it is not. Matthias From Molka.Gharbaoui at santannapisa.it Thu Mar 3 12:23:27 2022 From: Molka.Gharbaoui at santannapisa.it (Molka Gharbaoui) Date: Thu, 3 Mar 2022 12:23:27 +0000 Subject: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: Hi Matthias, Sorry for the confusion. Gnocchi is running and I am able to execute commands like "gnocchi metric list" and "gnocchi resource list" without errors even though the response is empty. Also checking mysql, the gnocchi database is correctly created. By the link I previously shared I wanted just to highlight that there are some differences between what was written there and between the environemnt I installed. I was searching there because I would have expected that Devstack installs and configures every component in automatic with no or minimum external intervention (in fact if I don't consider the Telemetry aspect, the rest works perfectly fine). I will follow Rafael guidelines to better undrstand where is the issue. Thanks. ________________________________ Da: Matthias Runge Inviato: gioved? 3 marzo 2022 11:59 A: openstack-discuss at lists.openstack.org Oggetto: Re: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi On 3/3/22 10:29, Molka Gharbaoui wrote: > Thank you Matthias for your reply. > > Unfortunately, from the logs I still cannot understand what is the issue. > > Could someone kindly indicate me how ceilometer and gnocchi should be > configured to get the data? > > For example in this link: > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > the > endpoints should use the following URL: http://controller:8041 > while in my case the 8041 port is not even > listening and the API indicated in the dashboard for Gnocchi is http:// > x.x.x.x/metric. > > Thank you in advance. > OpenStack configuration is complex. The guide does not remotely touch on production environments. Please do not confuse devstack with production envionments. Also the doc you linked is not for devstack itself. The flow in general is: data is generated by the services like nova or glance. Ceilometer listens on the bus and also polls for information. The info is sent then over (as configured in publishers.yaml) to gnocchi. Gnocchi has an api, ceilometer has not. Reading your response, this makes me suspicious if there is gnocchi running. From the initial question it suggests it is, from your most recent response, it is not. Matthias -------------- next part -------------- An HTML attachment was scrubbed... URL: From dsmigiel at redhat.com Thu Mar 3 13:32:33 2022 From: dsmigiel at redhat.com (Dariusz Smigiel) Date: Thu, 3 Mar 2022 05:32:33 -0800 Subject: [TripleO] Gate blocker for Centos 9 Wallaby In-Reply-To: References: Message-ID: > Currently TripleO gate is experiencing issues with building images [1] > We're aware of the problem, and we're trying to resolve this issue. > > Please withhold from rechecking RDO Third Party CI. > The fix has been merged. [2[ We should be out of the woods now. [2]: https://review.opendev.org/c/openstack/tripleo-common/+/831580 Thanks, dasm > > [1]: https://bugs.launchpad.net/tripleo/+bug/1962786 From satish.txt at gmail.com Thu Mar 3 14:58:43 2022 From: satish.txt at gmail.com (Satish Patel) Date: Thu, 3 Mar 2022 09:58:43 -0500 Subject: Openstack HPC storage suggestion In-Reply-To: References: <11fc1add-94f4-94a9-c47b-d50672708da0@cnrs.fr> Message-ID: Thank you folks, Your inputs are very valuable. Yes, StackHPC is a great source and whenever I search anything about HPC i endup on their website :) Thank you Mark for great work. I would sure like to come and listen to the SIG meetup. 8 year ago I was part of a large HPC project where we had 2000 physical compute nodes coupled with Infiniband fabric (no openstack) and we were using an LSF scheduler with Luster and ZFS for backend storage. This is what I am trying to do with openstack HPC. We built kolla-ansible using 40 compute nodes which include 20 IB nodes which have infiniband fabric, 10 GPU nodes and the remaining Memory nodes which have lots of memory. My clients are students so mostly they are looking for virtualized HPC where they can reserve resources and build their own cluster using Slurm, Condor or Spark whatever they want. We are using all virtualized so Infiniband mellanox configure for SRIOV passthrough to expose directly to VM. We have a VM-to-VM infiniband network for high speed messaging. I have put all those stuff in my blog here: https://satishdotpatel.github.io/HPC-on-openstack/ Now I want to use storage to meet my needs, i don't know at present because I don't know how and what application they are going to use. let's say i start with CephFS and folk are not happy then i have to change storage to Luster or maybe something else. Another question: what people use from scratch for HPC ( assuming it's going to be a local disk because of high I/O rate). On Thu, Mar 3, 2022 at 4:33 AM Manuel Holtgrewe wrote: > Hi, > > I guess it really depends on what HPC means to you ;-) > > Do your users schedule nova instances? Can you actually do Infiniband > RDMA between nova VMs? Or are your users scheduling ironic instances > via nova? > > We have an openstack setup based on kayobe/kolla where we: > > - Deploy 8 hyperconverged OpenStack nova libvirt compute+Ceph storage > (mostly for block storage and cephfs comes in handy for shared file > systems, e.g., for shared state in Slurm) servers. > - Deploy 300+ bare metal compute nodes via ironic. > - Use Slurm for the job scheduler. > - Setup everything after the bare OS with Ansible. > - Have login nodes and slurm scheduler etc. run in nova VMs. > - Only have Ethernet interconnect (100/25GbE for central switches and > recent servers and 40/10GbE for older servers). > > So we are using OpenStack nova+libvirt as a unified way of deploying > virtual and physical machines and then install them as we would do > with normal servers. That's a bit non-cloudy (as far as I have > learned) but works really well for us. > > Now to your HPC storage question... that I hope to answer a bit indirectly. > > The main advantage for using manila with CephFS (that I can see) is > that you get the openstack goodies of API and Horizon clients for > managing shares. I guess this is mostly useful if you want to have > cloud features for your HPC such as users allocating storage in > self-service to their nova/ironic machines. We come from a classic HPC > setting where the partitioning of the system is not done by creating > multiple nodes/clusters by users in self-service, but rather > administration provides a bare metal cluster with Slurm scheduler. > Users log in to head nodes and submit jobs to the compute nodes to > Slurm. Thus Slurm does the managing of resources and users can > allocate single cores up to the whole cluster. So in our use case > there would not be a major advantage of using manila for our storage > as we would primarily have one export that gives access to the whole > storage ;-). > > We currently have an old GPFS storage that we mount on all nodes via > Ansible. We are currently migrating to using an additional, dedicated > NVME-based ceph cluster (that is not hyperconverged with our compute) > and that we would also mount via Ansible. As we essentially only have > a single share on this storage managing the share with manila would be > more trouble than work. The much more important part will be setting > up Ceph appropriately and tuning it to perform well (e.g., using the > IO500 benchmark as the croit people demonstrate here [1]). > > I guess that does not really answer your question but I hope that it > gives useful perspective to you and maybe others. > > You can look at the work of the wonderful people of StackHPC that > provide commercial services around Openstack/Kayobe/Kolla/Ironic for > HPC setups. There are a couple of interesting videos involving their > staff and they have very useful information on the web site as well. > And as Mark Goddard just wrote you can have a look at the Scientific > SIG where they are involved as well. (I'm not in any way related to > StackHPC, I just really like their work). > > Best wishes, > Manuel > > [1] https://croit.io/blog/ceph-performance-test-and-optimization > > On Thu, Mar 3, 2022 at 9:25 AM Mahendra Paipuri > wrote: > > > > Hello, > > > > We are quite interested in this too. When we were looking into existing > > solutions, we found there has been some work in integrating LUSTRE into > > OpenStack [1], [2]. I remember coming across some Openinfra talks on > > developing a manila backend driver for LUSTRE. I am not quite sure if > > this project is still ongoing. Manila already provides a backend driver > > for GPFS [3] (unfortunately GPFS is not opensource) to readily integrate > > it into Openstack. Manila supports GlusterFS and CephFS as well but they > > do not have RDMA support (if I am not wrong). > > > > This is pretty much what we found. Would be glad explore more solutions > > if someone knows any. > > > > Cheers > > > > - > > > > Mahendra > > > > [1] > > > https://docs.google.com/presentation/d/1kGRzcdVQX95abei1bDVoRzxyC02i89_m5_sOfp8Aq6o/htmlpresent > > > > [2] > > > https://www.openstack.org/videos/summits/barcelona-2016/lustre-integration-for-hpc-on-openstack-at-cambridge-and-monash > > > > [3] https://docs.openstack.org/manila/latest/admin/gpfs_driver.html > > > > On 03/03/2022 04:47, Satish Patel wrote: > > > Folks, > > > > > > I built a 30 node HPC environment on openstack using mellanox > > > infiniband nic for high speed MPI messaging. So far everything works. > > > Now I am looking for a HPC PFS (parallel file system) similar to > > > Luster which I can mount on all HPC vms to run MPI jobs. > > > > > > I was reading on google and saw some CERN videos and they are using > > > Ceph (cephfs) for PFS. Also i was reading CephFS + Manila is a good > > > choice for HPC on openstack design. > > > > > > Does anyone have any experience with HPC storage for Openstack? Please > > > advice or share your experience :) Thanks in advance. > > > > > > ~S > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From katonalala at gmail.com Thu Mar 3 20:44:51 2022 From: katonalala at gmail.com (Lajos Katona) Date: Thu, 3 Mar 2022 21:44:51 +0100 Subject: [neutron] Drivers meeting agenda - 04.03.2022. Message-ID: Hi Neutron Drivers, The agenda for tomorrow's drivers meeting is at [1]. * Gratuitous ARPs are not sent during master transition: ( https://bugs.launchpad.net/neutron/+bug/1952907) ** For this one there are a few wip patches: ** https://review.opendev.org/c/openstack/neutron/+/821434 ** https://review.opendev.org/c/openstack/neutron/+/821433 ** https://review.opendev.org/c/openstack/neutron/+/712474 * On Demand agenda: (ralonsoh): https://bugs.launchpad.net/neutron/+bug/1962465. I want to discuss the scope of this bug and the possible fix. [1] https://wiki.openstack.org/wiki/Meetings/NeutronDrivers#Agenda See you at the meeting tomorrow. Lajos Katona (lajoskatona) -------------- next part -------------- An HTML attachment was scrubbed... URL: From iurygregory at gmail.com Fri Mar 4 03:23:04 2022 From: iurygregory at gmail.com (Iury Gregory) Date: Fri, 4 Mar 2022 03:23:04 +0000 Subject: [ironic] Zed PTG slots Message-ID: Hello Ironicers! For the Zed PTG I'm planning to have some slots that will allow APAC contributors to participate in the discussions. Keep in mind that we have some guidelines we should try to follow: 1. Cross project discussions (like SIGs or support project teams) should be scheduled towards the start of the week so that any discussions that might shape those of other teams happen first. 2. No team should sign up for more than 4 hours per UTC day to help keep participants actively engaged. 3. No team should sign up for more than 16 hours across all time slots to avoid burning out our contributors and to enable participation in multiple teams discussions. Based on this I have the following proposals: (Feel free to provide more ideas also) =) *Option #1* *16hrs* Monday till Thursday 1h (APAC night + EMEA morning) 2hrs (EMEA afternoon + NA morning) 1hr (NA afternoon/night + APAC morning) *Option #2* *15hrs* Tuesday till Friday 1h (APAC night + EMEA morning) 2hrs (EMEA afternoon + NA morning) 1hr (NA afternoon/night + APAC morning) * On Friday the NA + APAC session won't happen because is saturday for APAC contributors * *Option #3* *16hrs* Monday till Thursday 3hrs (EMEA afternoon + NA morning) 1hr (NA afternoon/night + APAC morning) *Option #4* *15hrs* Tuesday till Friday 3hrs (EMEA afternoon + NA morning) 1hr (NA afternoon/night + APAC morning) * On Friday the NA + APAC session won't happen because is saturday for APAC contributors * Please choose two options and we will vote during our upstream meeting on March 7th. Thank you! -- *Att[]'sIury Gregory Melo Ferreira * *MSc in Computer Science at UFCG* *Part of the ironic-core and puppet-manager-core team in OpenStack* *Software Engineer at Red Hat Czech* *Social*: https://www.linkedin.com/in/iurygregory *E-mail: iurygregory at gmail.com * -------------- next part -------------- An HTML attachment was scrubbed... URL: From hanguangyu2 at gmail.com Fri Mar 4 07:54:18 2022 From: hanguangyu2 at gmail.com (=?UTF-8?B?6Z+p5YWJ5a6H?=) Date: Fri, 4 Mar 2022 15:54:18 +0800 Subject: [Ironic] No suitable device was found for deployment In-Reply-To: References: <04af20bb-143b-9b56-6787-d5ac6ab0d38c@cern.ch> <3209d9fc-8db1-6dca-5736-402c6f082fcc@cern.ch> Message-ID: Hi Julia and Arne, > Did cleaning fail at any given point with these machines? Sorry that I didn't describe it clearly. Actually, clean is alway success in ironic log, deploying is failed. I just wonder if the cleaning phase did something that caused disk identification problems. And I have found the RAID config menu. In my machine, I need to prase "Ctrl + R" when RAID interface appear in machine boot. Thank you very much~! In RAID config menu, I found that the state of three disk is UB(unconfig bad). So, if I use the Ironic service to install an operating system for a server that has three hard disks in the 'JBOD' state, is there anything I should pay attention to or operate? If I don't do something for this, deploying stage give me an error 'No suitable device was found for deployment' and 'lslbk' is empty. After clean successed and deploy faided, the disk state is "Unconfigured bad". best wishes to you, Han Guangyu Julia Kreger ?2022?3?1??? 22:06??? > > On Mon, Feb 28, 2022 at 1:12 AM Arne Wiebalck wrote: > > > > Hi Guangyu, > > > > I am not aware of anything in the Ironic Python Agent that > > would remove disks from the system in a way that they would > > not be visible after a reboot (apart from, as mentioned before, > > the clean up of a hardware RAID in a way the IPA is not able > > to see any devices after). > > > > How about trying to access and configure the hardware RAID with > > the corresponding tool from the RAM disk you booted into from the > > USB stick? Install the tool and see if it detects the controller. > > > > The very first step before doing anything with Ironic is to > > get the disks back or understand why they are not visible. > > > > Did cleaning fail at any given point with these machines? > > If you have physical access, try disconnecting all of the drives, and > then powering up the machine and see if you can get into the firmware > configuration screen with control-h. If you can, remove all of the > prior configuration or disk volumes. They will look like they are in > error states most likely. If your unable to get into this screen, I > would be worried about your disk controller card. If your able to > clear everything out of the controller, power off, try re-inserting > drives, and see what happens. See if the controller can view/interact > with the drives. If it sees no drives, then my next paragraph is > likely the case. > > The disks sound like they might be in security locked state which will > likely require a desktop SATA disk controller to remedy by attaching > and manually removing from a security locked state. Megaraid > controllers can't recognize security locked devices (most controllers > and especially ones labeled "raid controllers" can't handle it) when > in pass-through mode, but I've never heard of security lock commands > actually getting through to the device with those controllers in > pass-through mode. If the card was in raid mode to begin with, then it > likely never did anything involving secure erase as the controller > should not be offering that as a feature of provided disks to the OS. > > > Cheers, > > Arne > > > > On 28.02.22 09:28, ??? wrote: > > > Hi Arne, > > > > > > I didn't find hardware RAID config option during the initial boot > > > sequence. Ctrl+H is unresponsive in this computer. I just saw "Press > > > Del to enter firmware configuration, press F3 to enter boot menu, and > > > press F12 to enter network boot". And I press 'Del' to enter the BIOS. > > > But I didn't find RAID config menu in BIOS. Sorry that I have poor > > > knowledge about this. > > > > > > And now, even though I installed the operating system manually using a > > > USB stick, I still couldn't find the hard drive. Is there anything > > > that ironic-agent did in the clean phase that would have caused this > > > problem? > > > > > > I wonder if this is a thinking pointto solve the problem. Now, my idea > > > is to first find a way to manually configure RAID. Do you agree with > > > this? And than, whether RAID configurations are still cleared in the > > > Clean phase if clean phase will do this? > > > > > > Sorry that I have so much confuse. > > > > > > love you, > > > Guangyu > > > > > > Arne Wiebalck ?2022?2?14??? 15:59??? > > >> > > >> Hi Guangyu, > > >> > > >> It seems like Julia had the right idea and the disks > > >> are not visible since the RAID controller does not > > >> expose anything to the operating system. This seems > > >> to be confirmed by you booting into the CentOS7 image. > > >> > > >> What I would suggest to try next is to look for the > > >> hardware RAID config option during the initial boot > > >> sequence to enter the RAID config menu (there should be > > >> a message quite early on, and maybe Ctrl-H is needed > > >> to enter the menu). > > >> > > >> Once there, manually configure the disks as JBODs or > > >> create a RAID device. Upon reboot this should be visible > > >> and accessible as a device. Maybe check from your CentOS7 > > >> image again. If the devices are there, Ironic should > > >> also be able to deploy on them (for this you can remove > > >> the RAID config you added). > > >> > > >> It depends a little on what your goal is, but I would > > >> try this first to see if you can make a device visible > > >> and if the Ironic deploy bit works, before trying to > > >> configure the hardware RAID via Ironic. > > >> > > >> Cheers, > > >> Arne > > >> > > >> On 14.02.22 03:20, ??? wrote: > > >>> Hi Arne and Julia, > > >>> > > >>> You make me feel so warm. Best wishes to you. > > >>> > > >>> I have tried to boot the node into a CentOS7, but it still coundnot to > > >>> find disk. And sorry that I didn't notice the RAID card. > > >>> > > >>> # lspci -v > > >>> ... > > >>> 23:00.0 RAID bus controller: Broadcom / LSI MegaRAID SAS-3 3108 > > >>> [Invader] (rev 02) > > >>> Subsystem: Broadcom / LSI MegaRAID SAS 9361-8i > > >>> Flags: bus master, fast devsel, latency 0, IRQ -2147483648, NUMA node 1 > > >>> I/O ports at 3000 [size=256] > > >>> Memory at e9900000 (64-bit, non-prefetchable) [size=64K] > > >>> Memory at e9700000 (64-bit, non-prefetchable) [size=1M] > > >>> Expansion ROM at e9800000 [disabled] [size=1M] > > >>> Capabilities: [50] Power Management version 3 > > >>> Capabilities: [68] Express Endpoint, MSI 00 > > >>> Capabilities: [d0] Vital Product Data > > >>> Capabilities: [a8] MSI: Enable- Count=1/1 Maskable+ 64bit+ > > >>> Capabilities: [c0] MSI-X: Enable+ Count=97 Masked- > > >>> Capabilities: [100] Advanced Error Reporting > > >>> Capabilities: [1e0] #19 > > >>> Capabilities: [1c0] Power Budgeting > > >>> Capabilities: [148] Alternative Routing-ID Interpretation (ARI) > > >>> Kernel driver in use: megaraid_sas > > >>> Kernel modules: megaraid_sas > > >>> ... > > >>> > > >>> I try to config raid fallowing > > >>> https://docs.openstack.org/ironic/latest/admin/raid.html > > >>> by `baremetal node set $NODE_UUID --target-raid-config raid.json`. The > > >>> server have three same disk(Western Digital DC HA210 2TB SATA 6GB/s) > > >>> # cat raid.json > > >>> { > > >>> "logical_disks": [ > > >>> { > > >>> "size_gb": "MAX", > > >>> "raid_level": "0", > > >>> "is_root_volume": true > > >>> } > > >>> ] > > >>> } > > >>> > > >>> But Ironic still coundn't see disk. I still got > > >>> ``` > > >>> ## In deploy images > > >>> # journalctl -fxeu ironic-python-agent > > >>> Feb 14 02:17:22 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 > > >>> 02:17:22.863 2329 WARNING root [-] Path /dev/disk/by-path is > > >>> inaccessible, /dev/disk/by-path/* version of block device name is > > >>> unavailable Cause: [Errno 2] No such file or directory: > > >>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or > > >>> directory: '/dev/disk/by-path' > > >>> Feb 14 02:17:44 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 > > >>> 02:17:44.391 2329 ERROR root [-] Unexpected error dispatching > > >>> get_os_install_device to manager > > >>> > >>> 0x7efbf4da2208>: Error finding the disk or partition device to deploy > > >>> the image onto: No suitable device was found for deployment - root > > >>> device hints were not provided and all found block devices are smaller > > >>> than 4294967296B.: ironic_python_agent.errors.DeviceNotFound: Error > > >>> finding the disk or partition device to deploy the image onto: No > > >>> suitable device was found for deployment - root device hints were not > > >>> provided and all found block devices are smaller than 4294967296B. > > >>> ``` > > >>> > > >>> I don't know if it's a lack of a RAID card driver or a lack of a disk > > >>> driver or a lack of RAID configuration. Could you have some idea about > > >>> this question? > > >>> > > >>> love you, > > >>> Han Guangyu > > >>> > > >>> > > >>> Julia Kreger ?2022?2?10??? 23:11??? > > >>> > > >>>> > > >>>> If the disk controllers *are* enumerated in the kernel log, which is > > >>>> something to also look for, then the disks themselves may be in some > > >>>> weird state like security locked. Generally this shows up as the > > >>>> operating system kind of sees the disk and the SATA port connected but > > >>>> can't really access it. This is also an exceptionally rare state to > > >>>> find one's self in. > > >>>> > > >>>> More common, especially in enterprise grade hardware: If the disk > > >>>> controller is actually a raid controller, and there are no raid > > >>>> volumes configured, then the operating system likely cannot see the > > >>>> underlying disks and turn that into a usable block device. I've seen a > > >>>> couple drivers over the years which expose hints of disks in the > > >>>> kernel log and without raid configuration in the cards, the drivers > > >>>> can't present usable block devices to the operating system system. > > >>>> > > >>>> -Julia > > >>>> > > >>>> On Thu, Feb 10, 2022 at 3:17 AM Arne Wiebalck wrote: > > >>>>> > > >>>>> Hi Guangyu, > > >>>>> > > >>>>> No worries about asking questions, this is what the mailing > > >>>>> list is for :) > > >>>>> > > >>>>> Just to clarify, you do not have to set root device hints, > > >>>>> it also works without (with the algorithm I mentioned). > > >>>>> However, hints help to define the exact device and/or make > > >>>>> deployment more predictable/repeatable. > > >>>>> > > >>>>> If it is really a driver problem, it is an issue with the > > >>>>> operating system of the image you use, i.e. CentOS8. Some > > >>>>> drivers were removed from 7 to 8, and we have seen issues > > >>>>> with specific drive models as well. > > >>>>> > > >>>>> You can try to build your own IPA images as described in > > >>>>> [1], e.g. to add your ssh key to be able to log into the > > >>>>> IPA to debug further, and to eventually include drivers > > >>>>> (if you can identify them and they are available for CentOS8). > > >>>>> > > >>>>> Another option may be to add another (newer) disk model to > > >>>>> the server, just to confirm it is the disk model/driver which > > >>>>> is the cause. > > >>>>> > > >>>>> You could also try to boot the node into a CentOS7 (and then > > >>>>> a CentOS8) live image to confirm it can see the disks at all. > > >>>>> > > >>>>> Hope this helps! > > >>>>> Arne > > >>>>> > > >>>>> [1] > > >>>>> https://docs.openstack.org/ironic-python-agent-builder/latest/admin/dib.html > > >>>>> > > >>>>> > > >>>>> On 10.02.22 11:15, ??? wrote: > > >>>>>> Hi Arne, > > >>>>>> > > >>>>>> Thank you very much for your response. Love you. You take away a lot > > >>>>>> of my confusion. > > >>>>>> > > >>>>>> You are right, I didn't set 'root device'. And Ironic also can not see > > >>>>>> disk, the content of the 'lsblk' file in the deploy los is emply. > > >>>>>> I tried to set 'root device', but because ironic can't find any disk, > > >>>>>> the deploy still filed. > > >>>>>> > > >>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 > > >>>>>> 09:51:55.045 2324 WARNING root [-] Path /dev/disk/by-path is > > >>>>>> inaccessible, /dev/disk/by-path/* version of block device name is > > >>>>>> unavailable Cause: [Errno 2] No such file or directory: > > >>>>>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or > > >>>>>> directory: '/dev/disk/by-path' > > >>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 > > >>>>>> 09:51:55.056 2324 WARNING ironic_lib.utils [-] No device found that > > >>>>>> matches the root device hints {'wwn': '0x50014EE2691D724C'}: > > >>>>>> StopIteration > > >>>>>> > > >>>>>> Sorry to bother you, I'm a newcomer of Ironic and I didn't find > > >>>>>> information about it on google. > > >>>>>> > > >>>>>> The bare metal node have three same disk(Western Digital DC HA210 2TB > > >>>>>> SATA 6GB/s). Where I can confirm whether ironic-python-agent supports > > >>>>>> this disk? > > >>>>>> > > >>>>>> And If Ironic cannot find disk since the corresponding drivers in the > > >>>>>> IPA image are missing, do you know how to resolve it? I have used the > > >>>>>> latest deploy images in > > >>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/ > > >>>>>> . Do I need to find and manually add driver in the source code or > > >>>>>> ramdisk(That was difficult tome)? > > >>>>>> > > >>>>>> Love you. > > >>>>>> > > >>>>>> Cheers, > > >>>>>> Guangyu > > >>>>>> > > >>>>>> Arne Wiebalck ?2022?2?10??? 15:51??? > > >>>>>>> > > >>>>>>> Hi Guangyu, > > >>>>>>> > > >>>>>>> The error indicates that Ironic was not able to find > > >>>>>>> a device where it could deploy the image to. > > >>>>>>> > > >>>>>>> To find a device, Ironic will use 'root device' > > >>>>>>> hints [1], usually set by the admin on a node. If that > > >>>>>>> does not yield anything, Ironic will loop over all > > >>>>>>> block devices and pick the smallest which is larger > > >>>>>>> than 4GB (and order them alphabetically). > > >>>>>>> > > >>>>>>> If you have disks in your server which are larger than > > >>>>>>> 4GB, one potential explanation is that Ironic cannot see them, > > >>>>>>> e.g. since the corresponding drivers in the IPA image are missing. > > >>>>>>> The logs you posted seem to confirm something along those > > >>>>>>> lines. > > >>>>>>> > > >>>>>>> Check the content of the 'lsblk' file in the deploy logs which > > >>>>>>> you can find in the tar archive in /var/log/ironic/deploy/ > > >>>>>>> on the controller for your deployment attempt to see what > > >>>>>>> devices Ironic has access to. > > >>>>>>> > > >>>>>>> Cheers, > > >>>>>>> Arne > > >>>>>>> > > >>>>>>> > > >>>>>>> [1] https://docs.openstack.org/ironic/latest/install/advanced.html#root-device-hints > > >>>>>>> > > >>>>>>> On 10.02.22 02:50, ??? wrote: > > >>>>>>>> Dear all, > > >>>>>>>> > > >>>>>>>> I have a OpenStack Victoria environment, and tried to use ironic > > >>>>>>>> manage bare metal. But I got "- root device hints were not provided > > >>>>>>>> and all found block devices are smaller than 4294967296B." in deploy > > >>>>>>>> stage. > > >>>>>>>> > > >>>>>>>> 2022-02-09 17:57:56.492 3908982 ERROR > > >>>>>>>> ironic.drivers.modules.agent_base [-] Agent returned error for deploy > > >>>>>>>> step {'step': 'write_image', 'priority': 80, 'argsinfo': None, > > >>>>>>>> 'interface': 'deploy'} on node cc68c450-ce54-4e1c-be04-8b0a6169ef92 : > > >>>>>>>> No suitable device was found for deployment - root device hints were > > >>>>>>>> not provided and all found block devices are smaller than > > >>>>>>>> 4294967296B.. > > >>>>>>>> > > >>>>>>>> I used "openstack server create --flavor my-baremetal-flavor --nic > > >>>>>>>> net-id=$net_id --image $image testing" to deploy bare metal node. I > > >>>>>>>> download deploy images(ipa-centos8-master.kernel and > > >>>>>>>> ipa-centos8-master.initramfs) in > > >>>>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/. > > >>>>>>>> > > >>>>>>>> The baremetal node info and flavor info as following: > > >>>>>>>> https://paste.opendev.org/show/bV7lgO6RkNQY6ZGPbT2e/ > > >>>>>>>> Ironic configure file as following: > > >>>>>>>> https://paste.opendev.org/show/bTgY9Kpn7KWqwQl73aEa/ > > >>>>>>>> Ironic-conductor log: https://paste.opendev.org/show/bFKZYlXmccxNxU8lEogk/ > > >>>>>>>> The log of ironic-python-agent in bare metal node: > > >>>>>>>> https://paste.opendev.org/show/btAuaMuV2IutV2Pa7YIa/ > > >>>>>>>> > > >>>>>>>> I see some old discussion about this, such as: > > >>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1312187. But those > > >>>>>>>> discussions took place a long time ago, not version V, and no solution > > >>>>>>>> was seen. > > >>>>>>>> > > >>>>>>>> Does anyone know how to solve this problem? I would appreciate any > > >>>>>>>> kind of guidance or help. > > >>>>>>>> > > >>>>>>>> Thank you, > > >>>>>>>> Han Guangyu > > >>>>>>>> > > >>>>> From arne.wiebalck at cern.ch Fri Mar 4 08:37:19 2022 From: arne.wiebalck at cern.ch (Arne Wiebalck) Date: Fri, 4 Mar 2022 09:37:19 +0100 Subject: [Ironic] No suitable device was found for deployment In-Reply-To: References: <04af20bb-143b-9b56-6787-d5ac6ab0d38c@cern.ch> <3209d9fc-8db1-6dca-5736-402c6f082fcc@cern.ch> Message-ID: <3542bf93-7e9c-9af6-88e5-99ec6a3fe34f@cern.ch> Hi Guangyu, I would think cleaning succeeds even if there are no disks: the loop to clean the disks is simply empty, so nothing to do, success! :) Deployment then fails since it needs a disk to deploy on. For my understanding: You reconfigured the disks into JBOD state and then retried to deploy (which failed and the disks fell back into UB state)? JBOD mode should work, but is usually not the main mode h/w RAID controllers work in. One thing to try is to actually configure a RAID-0 or RAID-1 device from your three disks and retry to deploy. I am not totally sure if Ironic would try to remove such a h/w RAID config during cleaning, but Julia will be able to tell. Cheers, Arne On 04.03.22 08:54, ??? wrote: > Hi Julia and Arne, > >> Did cleaning fail at any given point with these machines? > Sorry that I didn't describe it clearly. Actually, clean is alway > success in ironic log, deploying is failed. I just wonder if the > cleaning phase did something that caused disk identification problems. > > And I have found the RAID config menu. In my machine, I need to prase > "Ctrl + R" when RAID interface appear in machine boot. > Thank you very much~! In RAID config menu, I found that the state of > three disk is UB(unconfig bad). > > So, if I use the Ironic service to install an operating system for a > server that has three hard disks in the 'JBOD' state, is there > anything I should pay attention to or operate? If I don't do something > for this, deploying stage give me an error 'No suitable device was > found for deployment' and 'lslbk' is empty. After clean successed and > deploy faided, the disk state is "Unconfigured bad". > > best wishes to you, > Han Guangyu > > Julia Kreger ?2022?3?1??? 22:06??? > > >> >> On Mon, Feb 28, 2022 at 1:12 AM Arne Wiebalck wrote: >>> >>> Hi Guangyu, >>> >>> I am not aware of anything in the Ironic Python Agent that >>> would remove disks from the system in a way that they would >>> not be visible after a reboot (apart from, as mentioned before, >>> the clean up of a hardware RAID in a way the IPA is not able >>> to see any devices after). >>> >>> How about trying to access and configure the hardware RAID with >>> the corresponding tool from the RAM disk you booted into from the >>> USB stick? Install the tool and see if it detects the controller. >>> >>> The very first step before doing anything with Ironic is to >>> get the disks back or understand why they are not visible. >>> >> >> Did cleaning fail at any given point with these machines? >> >> If you have physical access, try disconnecting all of the drives, and >> then powering up the machine and see if you can get into the firmware >> configuration screen with control-h. If you can, remove all of the >> prior configuration or disk volumes. They will look like they are in >> error states most likely. If your unable to get into this screen, I >> would be worried about your disk controller card. If your able to >> clear everything out of the controller, power off, try re-inserting >> drives, and see what happens. See if the controller can view/interact >> with the drives. If it sees no drives, then my next paragraph is >> likely the case. >> >> The disks sound like they might be in security locked state which will >> likely require a desktop SATA disk controller to remedy by attaching >> and manually removing from a security locked state. Megaraid >> controllers can't recognize security locked devices (most controllers >> and especially ones labeled "raid controllers" can't handle it) when >> in pass-through mode, but I've never heard of security lock commands >> actually getting through to the device with those controllers in >> pass-through mode. If the card was in raid mode to begin with, then it >> likely never did anything involving secure erase as the controller >> should not be offering that as a feature of provided disks to the OS. >> >>> Cheers, >>> Arne >>> >>> On 28.02.22 09:28, ??? wrote: >>>> Hi Arne, >>>> >>>> I didn't find hardware RAID config option during the initial boot >>>> sequence. Ctrl+H is unresponsive in this computer. I just saw "Press >>>> Del to enter firmware configuration, press F3 to enter boot menu, and >>>> press F12 to enter network boot". And I press 'Del' to enter the BIOS. >>>> But I didn't find RAID config menu in BIOS. Sorry that I have poor >>>> knowledge about this. >>>> >>>> And now, even though I installed the operating system manually using a >>>> USB stick, I still couldn't find the hard drive. Is there anything >>>> that ironic-agent did in the clean phase that would have caused this >>>> problem? >>>> >>>> I wonder if this is a thinking pointto solve the problem. Now, my idea >>>> is to first find a way to manually configure RAID. Do you agree with >>>> this? And than, whether RAID configurations are still cleared in the >>>> Clean phase if clean phase will do this? >>>> >>>> Sorry that I have so much confuse. >>>> >>>> love you, >>>> Guangyu >>>> >>>> Arne Wiebalck ?2022?2?14??? 15:59??? >>>>> >>>>> Hi Guangyu, >>>>> >>>>> It seems like Julia had the right idea and the disks >>>>> are not visible since the RAID controller does not >>>>> expose anything to the operating system. This seems >>>>> to be confirmed by you booting into the CentOS7 image. >>>>> >>>>> What I would suggest to try next is to look for the >>>>> hardware RAID config option during the initial boot >>>>> sequence to enter the RAID config menu (there should be >>>>> a message quite early on, and maybe Ctrl-H is needed >>>>> to enter the menu). >>>>> >>>>> Once there, manually configure the disks as JBODs or >>>>> create a RAID device. Upon reboot this should be visible >>>>> and accessible as a device. Maybe check from your CentOS7 >>>>> image again. If the devices are there, Ironic should >>>>> also be able to deploy on them (for this you can remove >>>>> the RAID config you added). >>>>> >>>>> It depends a little on what your goal is, but I would >>>>> try this first to see if you can make a device visible >>>>> and if the Ironic deploy bit works, before trying to >>>>> configure the hardware RAID via Ironic. >>>>> >>>>> Cheers, >>>>> Arne >>>>> >>>>> On 14.02.22 03:20, ??? wrote: >>>>>> Hi Arne and Julia, >>>>>> >>>>>> You make me feel so warm. Best wishes to you. >>>>>> >>>>>> I have tried to boot the node into a CentOS7, but it still coundnot to >>>>>> find disk. And sorry that I didn't notice the RAID card. >>>>>> >>>>>> # lspci -v >>>>>> ... >>>>>> 23:00.0 RAID bus controller: Broadcom / LSI MegaRAID SAS-3 3108 >>>>>> [Invader] (rev 02) >>>>>> Subsystem: Broadcom / LSI MegaRAID SAS 9361-8i >>>>>> Flags: bus master, fast devsel, latency 0, IRQ -2147483648, NUMA node 1 >>>>>> I/O ports at 3000 [size=256] >>>>>> Memory at e9900000 (64-bit, non-prefetchable) [size=64K] >>>>>> Memory at e9700000 (64-bit, non-prefetchable) [size=1M] >>>>>> Expansion ROM at e9800000 [disabled] [size=1M] >>>>>> Capabilities: [50] Power Management version 3 >>>>>> Capabilities: [68] Express Endpoint, MSI 00 >>>>>> Capabilities: [d0] Vital Product Data >>>>>> Capabilities: [a8] MSI: Enable- Count=1/1 Maskable+ 64bit+ >>>>>> Capabilities: [c0] MSI-X: Enable+ Count=97 Masked- >>>>>> Capabilities: [100] Advanced Error Reporting >>>>>> Capabilities: [1e0] #19 >>>>>> Capabilities: [1c0] Power Budgeting >>>>>> Capabilities: [148] Alternative Routing-ID Interpretation (ARI) >>>>>> Kernel driver in use: megaraid_sas >>>>>> Kernel modules: megaraid_sas >>>>>> ... >>>>>> >>>>>> I try to config raid fallowing >>>>>> https://docs.openstack.org/ironic/latest/admin/raid.html >>>>>> by `baremetal node set $NODE_UUID --target-raid-config raid.json`. The >>>>>> server have three same disk(Western Digital DC HA210 2TB SATA 6GB/s) >>>>>> # cat raid.json >>>>>> { >>>>>> "logical_disks": [ >>>>>> { >>>>>> "size_gb": "MAX", >>>>>> "raid_level": "0", >>>>>> "is_root_volume": true >>>>>> } >>>>>> ] >>>>>> } >>>>>> >>>>>> But Ironic still coundn't see disk. I still got >>>>>> ``` >>>>>> ## In deploy images >>>>>> # journalctl -fxeu ironic-python-agent >>>>>> Feb 14 02:17:22 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 >>>>>> 02:17:22.863 2329 WARNING root [-] Path /dev/disk/by-path is >>>>>> inaccessible, /dev/disk/by-path/* version of block device name is >>>>>> unavailable Cause: [Errno 2] No such file or directory: >>>>>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or >>>>>> directory: '/dev/disk/by-path' >>>>>> Feb 14 02:17:44 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 >>>>>> 02:17:44.391 2329 ERROR root [-] Unexpected error dispatching >>>>>> get_os_install_device to manager >>>>>> >>>>> 0x7efbf4da2208>: Error finding the disk or partition device to deploy >>>>>> the image onto: No suitable device was found for deployment - root >>>>>> device hints were not provided and all found block devices are smaller >>>>>> than 4294967296B.: ironic_python_agent.errors.DeviceNotFound: Error >>>>>> finding the disk or partition device to deploy the image onto: No >>>>>> suitable device was found for deployment - root device hints were not >>>>>> provided and all found block devices are smaller than 4294967296B. >>>>>> ``` >>>>>> >>>>>> I don't know if it's a lack of a RAID card driver or a lack of a disk >>>>>> driver or a lack of RAID configuration. Could you have some idea about >>>>>> this question? >>>>>> >>>>>> love you, >>>>>> Han Guangyu >>>>>> >>>>>> >>>>>> Julia Kreger ?2022?2?10??? 23:11??? >>>>>> >>>>>>> >>>>>>> If the disk controllers *are* enumerated in the kernel log, which is >>>>>>> something to also look for, then the disks themselves may be in some >>>>>>> weird state like security locked. Generally this shows up as the >>>>>>> operating system kind of sees the disk and the SATA port connected but >>>>>>> can't really access it. This is also an exceptionally rare state to >>>>>>> find one's self in. >>>>>>> >>>>>>> More common, especially in enterprise grade hardware: If the disk >>>>>>> controller is actually a raid controller, and there are no raid >>>>>>> volumes configured, then the operating system likely cannot see the >>>>>>> underlying disks and turn that into a usable block device. I've seen a >>>>>>> couple drivers over the years which expose hints of disks in the >>>>>>> kernel log and without raid configuration in the cards, the drivers >>>>>>> can't present usable block devices to the operating system system. >>>>>>> >>>>>>> -Julia >>>>>>> >>>>>>> On Thu, Feb 10, 2022 at 3:17 AM Arne Wiebalck wrote: >>>>>>>> >>>>>>>> Hi Guangyu, >>>>>>>> >>>>>>>> No worries about asking questions, this is what the mailing >>>>>>>> list is for :) >>>>>>>> >>>>>>>> Just to clarify, you do not have to set root device hints, >>>>>>>> it also works without (with the algorithm I mentioned). >>>>>>>> However, hints help to define the exact device and/or make >>>>>>>> deployment more predictable/repeatable. >>>>>>>> >>>>>>>> If it is really a driver problem, it is an issue with the >>>>>>>> operating system of the image you use, i.e. CentOS8. Some >>>>>>>> drivers were removed from 7 to 8, and we have seen issues >>>>>>>> with specific drive models as well. >>>>>>>> >>>>>>>> You can try to build your own IPA images as described in >>>>>>>> [1], e.g. to add your ssh key to be able to log into the >>>>>>>> IPA to debug further, and to eventually include drivers >>>>>>>> (if you can identify them and they are available for CentOS8). >>>>>>>> >>>>>>>> Another option may be to add another (newer) disk model to >>>>>>>> the server, just to confirm it is the disk model/driver which >>>>>>>> is the cause. >>>>>>>> >>>>>>>> You could also try to boot the node into a CentOS7 (and then >>>>>>>> a CentOS8) live image to confirm it can see the disks at all. >>>>>>>> >>>>>>>> Hope this helps! >>>>>>>> Arne >>>>>>>> >>>>>>>> [1] >>>>>>>> https://docs.openstack.org/ironic-python-agent-builder/latest/admin/dib.html >>>>>>>> >>>>>>>> >>>>>>>> On 10.02.22 11:15, ??? wrote: >>>>>>>>> Hi Arne, >>>>>>>>> >>>>>>>>> Thank you very much for your response. Love you. You take away a lot >>>>>>>>> of my confusion. >>>>>>>>> >>>>>>>>> You are right, I didn't set 'root device'. And Ironic also can not see >>>>>>>>> disk, the content of the 'lsblk' file in the deploy los is emply. >>>>>>>>> I tried to set 'root device', but because ironic can't find any disk, >>>>>>>>> the deploy still filed. >>>>>>>>> >>>>>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 >>>>>>>>> 09:51:55.045 2324 WARNING root [-] Path /dev/disk/by-path is >>>>>>>>> inaccessible, /dev/disk/by-path/* version of block device name is >>>>>>>>> unavailable Cause: [Errno 2] No such file or directory: >>>>>>>>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or >>>>>>>>> directory: '/dev/disk/by-path' >>>>>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 >>>>>>>>> 09:51:55.056 2324 WARNING ironic_lib.utils [-] No device found that >>>>>>>>> matches the root device hints {'wwn': '0x50014EE2691D724C'}: >>>>>>>>> StopIteration >>>>>>>>> >>>>>>>>> Sorry to bother you, I'm a newcomer of Ironic and I didn't find >>>>>>>>> information about it on google. >>>>>>>>> >>>>>>>>> The bare metal node have three same disk(Western Digital DC HA210 2TB >>>>>>>>> SATA 6GB/s). Where I can confirm whether ironic-python-agent supports >>>>>>>>> this disk? >>>>>>>>> >>>>>>>>> And If Ironic cannot find disk since the corresponding drivers in the >>>>>>>>> IPA image are missing, do you know how to resolve it? I have used the >>>>>>>>> latest deploy images in >>>>>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/ >>>>>>>>> . Do I need to find and manually add driver in the source code or >>>>>>>>> ramdisk(That was difficult tome)? >>>>>>>>> >>>>>>>>> Love you. >>>>>>>>> >>>>>>>>> Cheers, >>>>>>>>> Guangyu >>>>>>>>> >>>>>>>>> Arne Wiebalck ?2022?2?10??? 15:51??? >>>>>>>>>> >>>>>>>>>> Hi Guangyu, >>>>>>>>>> >>>>>>>>>> The error indicates that Ironic was not able to find >>>>>>>>>> a device where it could deploy the image to. >>>>>>>>>> >>>>>>>>>> To find a device, Ironic will use 'root device' >>>>>>>>>> hints [1], usually set by the admin on a node. If that >>>>>>>>>> does not yield anything, Ironic will loop over all >>>>>>>>>> block devices and pick the smallest which is larger >>>>>>>>>> than 4GB (and order them alphabetically). >>>>>>>>>> >>>>>>>>>> If you have disks in your server which are larger than >>>>>>>>>> 4GB, one potential explanation is that Ironic cannot see them, >>>>>>>>>> e.g. since the corresponding drivers in the IPA image are missing. >>>>>>>>>> The logs you posted seem to confirm something along those >>>>>>>>>> lines. >>>>>>>>>> >>>>>>>>>> Check the content of the 'lsblk' file in the deploy logs which >>>>>>>>>> you can find in the tar archive in /var/log/ironic/deploy/ >>>>>>>>>> on the controller for your deployment attempt to see what >>>>>>>>>> devices Ironic has access to. >>>>>>>>>> >>>>>>>>>> Cheers, >>>>>>>>>> Arne >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> [1] https://docs.openstack.org/ironic/latest/install/advanced.html#root-device-hints >>>>>>>>>> >>>>>>>>>> On 10.02.22 02:50, ??? wrote: >>>>>>>>>>> Dear all, >>>>>>>>>>> >>>>>>>>>>> I have a OpenStack Victoria environment, and tried to use ironic >>>>>>>>>>> manage bare metal. But I got "- root device hints were not provided >>>>>>>>>>> and all found block devices are smaller than 4294967296B." in deploy >>>>>>>>>>> stage. >>>>>>>>>>> >>>>>>>>>>> 2022-02-09 17:57:56.492 3908982 ERROR >>>>>>>>>>> ironic.drivers.modules.agent_base [-] Agent returned error for deploy >>>>>>>>>>> step {'step': 'write_image', 'priority': 80, 'argsinfo': None, >>>>>>>>>>> 'interface': 'deploy'} on node cc68c450-ce54-4e1c-be04-8b0a6169ef92 : >>>>>>>>>>> No suitable device was found for deployment - root device hints were >>>>>>>>>>> not provided and all found block devices are smaller than >>>>>>>>>>> 4294967296B.. >>>>>>>>>>> >>>>>>>>>>> I used "openstack server create --flavor my-baremetal-flavor --nic >>>>>>>>>>> net-id=$net_id --image $image testing" to deploy bare metal node. I >>>>>>>>>>> download deploy images(ipa-centos8-master.kernel and >>>>>>>>>>> ipa-centos8-master.initramfs) in >>>>>>>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/. >>>>>>>>>>> >>>>>>>>>>> The baremetal node info and flavor info as following: >>>>>>>>>>> https://paste.opendev.org/show/bV7lgO6RkNQY6ZGPbT2e/ >>>>>>>>>>> Ironic configure file as following: >>>>>>>>>>> https://paste.opendev.org/show/bTgY9Kpn7KWqwQl73aEa/ >>>>>>>>>>> Ironic-conductor log: https://paste.opendev.org/show/bFKZYlXmccxNxU8lEogk/ >>>>>>>>>>> The log of ironic-python-agent in bare metal node: >>>>>>>>>>> https://paste.opendev.org/show/btAuaMuV2IutV2Pa7YIa/ >>>>>>>>>>> >>>>>>>>>>> I see some old discussion about this, such as: >>>>>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1312187. But those >>>>>>>>>>> discussions took place a long time ago, not version V, and no solution >>>>>>>>>>> was seen. >>>>>>>>>>> >>>>>>>>>>> Does anyone know how to solve this problem? I would appreciate any >>>>>>>>>>> kind of guidance or help. >>>>>>>>>>> >>>>>>>>>>> Thank you, >>>>>>>>>>> Han Guangyu >>>>>>>>>>> >>>>>>>> From hanguangyu2 at gmail.com Fri Mar 4 09:25:11 2022 From: hanguangyu2 at gmail.com (=?UTF-8?B?6Z+p5YWJ5a6H?=) Date: Fri, 4 Mar 2022 17:25:11 +0800 Subject: [Ironic] No suitable device was found for deployment In-Reply-To: References: <04af20bb-143b-9b56-6787-d5ac6ab0d38c@cern.ch> <3209d9fc-8db1-6dca-5736-402c6f082fcc@cern.ch> Message-ID: Hi Julia, Sorry that my last email didn't reploy some question in your email. When I get into RAID config menu, I said disk state is "unconfig bad". And the more info is that Virtual Drive Management displayed "No Configuration Present!". But I cound not modify disk state in RAID confug menu. Even if I moved the disk to other server, and used the software in operating system to modify it, it still coudn't to be modify. such as: ```shell # MegaCli -PDList -a0 Adapter #0 Enclosure Device ID: 252 Slot Number: 1 Enclosure position: 0 Device Id: 9 WWN: 50014EE2BE72FEBF Sequence Number: 2 Media Error Count: 0 Other Error Count: 0 Predictive Failure Count: 0 Last Predictive Failure Event Seq Number: 0 PD Type: SATA Raw Size: 0 KB [0x0 Sectors] Non Coerced Size: 0 KB [0x0 Sectors] Coerced Size: 0 KB [0x0 Sectors] Firmware state: Unconfigured(bad) Device Firmware Level: WA09 Shield Counter: 0 Successful diagnostics completion on : N/A SAS Address(0): 0x4433221101000000 Connected Port Number: 0(path0) Inquiry Data: ATA HGST HUS722T2TALWA09WCC6N4HZV9SX FDE Enable: Disable Secured: Unsecured Locked: Unlocked Needs EKM Attention: No Foreign State: None Device Speed: Unknown Link Speed: 6.0Gb/s Media Type: Hard Disk Device Drive Temperature :0C (32.00 F) PI Eligibility: No Drive is formatted for PI information: No PI: No PI Drive's write cache : Disabled Drive's NCQ setting : Enabled Port-0 : Port status: Active Port's Linkspeed: 6.0Gb/s Drive has flagged a S.M.A.R.T alert : No Enclosure Device ID: 252 Slot Number: 2 Enclosure position: 0 Device Id: 10 WWN: 50014EE2691D724C Sequence Number: 2 Media Error Count: 0 Other Error Count: 0 Predictive Failure Count: 0 Last Predictive Failure Event Seq Number: 0 PD Type: SATA Raw Size: 0 KB [0x0 Sectors] Non Coerced Size: 0 KB [0x0 Sectors] Coerced Size: 0 KB [0x0 Sectors] Firmware state: Unconfigured(bad) Device Firmware Level: WA09 Shield Counter: 0 Successful diagnostics completion on : N/A SAS Address(0): 0x4433221102000000 Connected Port Number: 1(path0) Inquiry Data: ATA HGST HUS722T2TALWA09WCC6N4HZVTV5 FDE Enable: Disable Secured: Unsecured Locked: Unlocked Needs EKM Attention: No Foreign State: None Device Speed: Unknown Link Speed: 6.0Gb/s Media Type: Hard Disk Device Drive Temperature :0C (32.00 F) PI Eligibility: No Drive is formatted for PI information: No PI: No PI Drive's write cache : Disabled Drive's NCQ setting : Enabled Port-0 : Port status: Active Port's Linkspeed: 6.0Gb/s Drive has flagged a S.M.A.R.T alert : No Enclosure Device ID: 252 Slot Number: 3 Enclosure position: 0 Device Id: 11 WWN: 50014EE2BE733A5E Sequence Number: 2 Media Error Count: 0 Other Error Count: 0 Predictive Failure Count: 0 Last Predictive Failure Event Seq Number: 0 PD Type: SATA Raw Size: 1.819 TB [0xe8e088b0 Sectors] Non Coerced Size: 1.818 TB [0xe8d088b0 Sectors] Coerced Size: 1.818 TB [0xe8d00000 Sectors] Firmware state: JBOD Device Firmware Level: WA09 Shield Counter: 0 Successful diagnostics completion on : N/A SAS Address(0): 0x4433221103000000 Connected Port Number: 2(path0) Inquiry Data: WCC6N0KX0HJD HGST HUS722T2TALA604 RAGNWA09 FDE Enable: Disable Secured: Unsecured Locked: Unlocked Needs EKM Attention: No Foreign State: None Device Speed: 6.0Gb/s Link Speed: 6.0Gb/s Media Type: Hard Disk Device Drive Temperature :34C (93.20 F) PI Eligibility: No Drive is formatted for PI information: No PI: No PI Drive's write cache : Enabled Drive's NCQ setting : Enabled Port-0 : Port status: Active Port's Linkspeed: 6.0Gb/s Drive has flagged a S.M.A.R.T alert : No Exit Code: 0x00 ``` ```shell # MegaCli -PDList -a0 | grep state Firmware state: Unconfigured(bad) Firmware state: Unconfigured(bad) Firmware state: JBOD # MegaCli -PDMakeGood -PhysDrv[252:2] -a0 Adapter: 0: Failed to change PD state at EnclId-252 SlotId-2. Exit Code: 0x01 ``` And the "security locked state" is a idea that I continue. I will try to determine if it is in this state and find a way to disarm it. Thank you very much Julia, cheers, Han Guangyu Julia Kreger ?2022?3?1??? 22:06??? > > On Mon, Feb 28, 2022 at 1:12 AM Arne Wiebalck wrote: > > > > Hi Guangyu, > > > > I am not aware of anything in the Ironic Python Agent that > > would remove disks from the system in a way that they would > > not be visible after a reboot (apart from, as mentioned before, > > the clean up of a hardware RAID in a way the IPA is not able > > to see any devices after). > > > > How about trying to access and configure the hardware RAID with > > the corresponding tool from the RAM disk you booted into from the > > USB stick? Install the tool and see if it detects the controller. > > > > The very first step before doing anything with Ironic is to > > get the disks back or understand why they are not visible. > > > > Did cleaning fail at any given point with these machines? > > If you have physical access, try disconnecting all of the drives, and > then powering up the machine and see if you can get into the firmware > configuration screen with control-h. If you can, remove all of the > prior configuration or disk volumes. They will look like they are in > error states most likely. If your unable to get into this screen, I > would be worried about your disk controller card. If your able to > clear everything out of the controller, power off, try re-inserting > drives, and see what happens. See if the controller can view/interact > with the drives. If it sees no drives, then my next paragraph is > likely the case. > > The disks sound like they might be in security locked state which will > likely require a desktop SATA disk controller to remedy by attaching > and manually removing from a security locked state. Megaraid > controllers can't recognize security locked devices (most controllers > and especially ones labeled "raid controllers" can't handle it) when > in pass-through mode, but I've never heard of security lock commands > actually getting through to the device with those controllers in > pass-through mode. If the card was in raid mode to begin with, then it > likely never did anything involving secure erase as the controller > should not be offering that as a feature of provided disks to the OS. > > > Cheers, > > Arne > > > > On 28.02.22 09:28, ??? wrote: > > > Hi Arne, > > > > > > I didn't find hardware RAID config option during the initial boot > > > sequence. Ctrl+H is unresponsive in this computer. I just saw "Press > > > Del to enter firmware configuration, press F3 to enter boot menu, and > > > press F12 to enter network boot". And I press 'Del' to enter the BIOS. > > > But I didn't find RAID config menu in BIOS. Sorry that I have poor > > > knowledge about this. > > > > > > And now, even though I installed the operating system manually using a > > > USB stick, I still couldn't find the hard drive. Is there anything > > > that ironic-agent did in the clean phase that would have caused this > > > problem? > > > > > > I wonder if this is a thinking pointto solve the problem. Now, my idea > > > is to first find a way to manually configure RAID. Do you agree with > > > this? And than, whether RAID configurations are still cleared in the > > > Clean phase if clean phase will do this? > > > > > > Sorry that I have so much confuse. > > > > > > love you, > > > Guangyu > > > > > > Arne Wiebalck ?2022?2?14??? 15:59??? > > >> > > >> Hi Guangyu, > > >> > > >> It seems like Julia had the right idea and the disks > > >> are not visible since the RAID controller does not > > >> expose anything to the operating system. This seems > > >> to be confirmed by you booting into the CentOS7 image. > > >> > > >> What I would suggest to try next is to look for the > > >> hardware RAID config option during the initial boot > > >> sequence to enter the RAID config menu (there should be > > >> a message quite early on, and maybe Ctrl-H is needed > > >> to enter the menu). > > >> > > >> Once there, manually configure the disks as JBODs or > > >> create a RAID device. Upon reboot this should be visible > > >> and accessible as a device. Maybe check from your CentOS7 > > >> image again. If the devices are there, Ironic should > > >> also be able to deploy on them (for this you can remove > > >> the RAID config you added). > > >> > > >> It depends a little on what your goal is, but I would > > >> try this first to see if you can make a device visible > > >> and if the Ironic deploy bit works, before trying to > > >> configure the hardware RAID via Ironic. > > >> > > >> Cheers, > > >> Arne > > >> > > >> On 14.02.22 03:20, ??? wrote: > > >>> Hi Arne and Julia, > > >>> > > >>> You make me feel so warm. Best wishes to you. > > >>> > > >>> I have tried to boot the node into a CentOS7, but it still coundnot to > > >>> find disk. And sorry that I didn't notice the RAID card. > > >>> > > >>> # lspci -v > > >>> ... > > >>> 23:00.0 RAID bus controller: Broadcom / LSI MegaRAID SAS-3 3108 > > >>> [Invader] (rev 02) > > >>> Subsystem: Broadcom / LSI MegaRAID SAS 9361-8i > > >>> Flags: bus master, fast devsel, latency 0, IRQ -2147483648, NUMA node 1 > > >>> I/O ports at 3000 [size=256] > > >>> Memory at e9900000 (64-bit, non-prefetchable) [size=64K] > > >>> Memory at e9700000 (64-bit, non-prefetchable) [size=1M] > > >>> Expansion ROM at e9800000 [disabled] [size=1M] > > >>> Capabilities: [50] Power Management version 3 > > >>> Capabilities: [68] Express Endpoint, MSI 00 > > >>> Capabilities: [d0] Vital Product Data > > >>> Capabilities: [a8] MSI: Enable- Count=1/1 Maskable+ 64bit+ > > >>> Capabilities: [c0] MSI-X: Enable+ Count=97 Masked- > > >>> Capabilities: [100] Advanced Error Reporting > > >>> Capabilities: [1e0] #19 > > >>> Capabilities: [1c0] Power Budgeting > > >>> Capabilities: [148] Alternative Routing-ID Interpretation (ARI) > > >>> Kernel driver in use: megaraid_sas > > >>> Kernel modules: megaraid_sas > > >>> ... > > >>> > > >>> I try to config raid fallowing > > >>> https://docs.openstack.org/ironic/latest/admin/raid.html > > >>> by `baremetal node set $NODE_UUID --target-raid-config raid.json`. The > > >>> server have three same disk(Western Digital DC HA210 2TB SATA 6GB/s) > > >>> # cat raid.json > > >>> { > > >>> "logical_disks": [ > > >>> { > > >>> "size_gb": "MAX", > > >>> "raid_level": "0", > > >>> "is_root_volume": true > > >>> } > > >>> ] > > >>> } > > >>> > > >>> But Ironic still coundn't see disk. I still got > > >>> ``` > > >>> ## In deploy images > > >>> # journalctl -fxeu ironic-python-agent > > >>> Feb 14 02:17:22 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 > > >>> 02:17:22.863 2329 WARNING root [-] Path /dev/disk/by-path is > > >>> inaccessible, /dev/disk/by-path/* version of block device name is > > >>> unavailable Cause: [Errno 2] No such file or directory: > > >>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or > > >>> directory: '/dev/disk/by-path' > > >>> Feb 14 02:17:44 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 > > >>> 02:17:44.391 2329 ERROR root [-] Unexpected error dispatching > > >>> get_os_install_device to manager > > >>> > >>> 0x7efbf4da2208>: Error finding the disk or partition device to deploy > > >>> the image onto: No suitable device was found for deployment - root > > >>> device hints were not provided and all found block devices are smaller > > >>> than 4294967296B.: ironic_python_agent.errors.DeviceNotFound: Error > > >>> finding the disk or partition device to deploy the image onto: No > > >>> suitable device was found for deployment - root device hints were not > > >>> provided and all found block devices are smaller than 4294967296B. > > >>> ``` > > >>> > > >>> I don't know if it's a lack of a RAID card driver or a lack of a disk > > >>> driver or a lack of RAID configuration. Could you have some idea about > > >>> this question? > > >>> > > >>> love you, > > >>> Han Guangyu > > >>> > > >>> > > >>> Julia Kreger ?2022?2?10??? 23:11??? > > >>> > > >>>> > > >>>> If the disk controllers *are* enumerated in the kernel log, which is > > >>>> something to also look for, then the disks themselves may be in some > > >>>> weird state like security locked. Generally this shows up as the > > >>>> operating system kind of sees the disk and the SATA port connected but > > >>>> can't really access it. This is also an exceptionally rare state to > > >>>> find one's self in. > > >>>> > > >>>> More common, especially in enterprise grade hardware: If the disk > > >>>> controller is actually a raid controller, and there are no raid > > >>>> volumes configured, then the operating system likely cannot see the > > >>>> underlying disks and turn that into a usable block device. I've seen a > > >>>> couple drivers over the years which expose hints of disks in the > > >>>> kernel log and without raid configuration in the cards, the drivers > > >>>> can't present usable block devices to the operating system system. > > >>>> > > >>>> -Julia > > >>>> > > >>>> On Thu, Feb 10, 2022 at 3:17 AM Arne Wiebalck wrote: > > >>>>> > > >>>>> Hi Guangyu, > > >>>>> > > >>>>> No worries about asking questions, this is what the mailing > > >>>>> list is for :) > > >>>>> > > >>>>> Just to clarify, you do not have to set root device hints, > > >>>>> it also works without (with the algorithm I mentioned). > > >>>>> However, hints help to define the exact device and/or make > > >>>>> deployment more predictable/repeatable. > > >>>>> > > >>>>> If it is really a driver problem, it is an issue with the > > >>>>> operating system of the image you use, i.e. CentOS8. Some > > >>>>> drivers were removed from 7 to 8, and we have seen issues > > >>>>> with specific drive models as well. > > >>>>> > > >>>>> You can try to build your own IPA images as described in > > >>>>> [1], e.g. to add your ssh key to be able to log into the > > >>>>> IPA to debug further, and to eventually include drivers > > >>>>> (if you can identify them and they are available for CentOS8). > > >>>>> > > >>>>> Another option may be to add another (newer) disk model to > > >>>>> the server, just to confirm it is the disk model/driver which > > >>>>> is the cause. > > >>>>> > > >>>>> You could also try to boot the node into a CentOS7 (and then > > >>>>> a CentOS8) live image to confirm it can see the disks at all. > > >>>>> > > >>>>> Hope this helps! > > >>>>> Arne > > >>>>> > > >>>>> [1] > > >>>>> https://docs.openstack.org/ironic-python-agent-builder/latest/admin/dib.html > > >>>>> > > >>>>> > > >>>>> On 10.02.22 11:15, ??? wrote: > > >>>>>> Hi Arne, > > >>>>>> > > >>>>>> Thank you very much for your response. Love you. You take away a lot > > >>>>>> of my confusion. > > >>>>>> > > >>>>>> You are right, I didn't set 'root device'. And Ironic also can not see > > >>>>>> disk, the content of the 'lsblk' file in the deploy los is emply. > > >>>>>> I tried to set 'root device', but because ironic can't find any disk, > > >>>>>> the deploy still filed. > > >>>>>> > > >>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 > > >>>>>> 09:51:55.045 2324 WARNING root [-] Path /dev/disk/by-path is > > >>>>>> inaccessible, /dev/disk/by-path/* version of block device name is > > >>>>>> unavailable Cause: [Errno 2] No such file or directory: > > >>>>>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or > > >>>>>> directory: '/dev/disk/by-path' > > >>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 > > >>>>>> 09:51:55.056 2324 WARNING ironic_lib.utils [-] No device found that > > >>>>>> matches the root device hints {'wwn': '0x50014EE2691D724C'}: > > >>>>>> StopIteration > > >>>>>> > > >>>>>> Sorry to bother you, I'm a newcomer of Ironic and I didn't find > > >>>>>> information about it on google. > > >>>>>> > > >>>>>> The bare metal node have three same disk(Western Digital DC HA210 2TB > > >>>>>> SATA 6GB/s). Where I can confirm whether ironic-python-agent supports > > >>>>>> this disk? > > >>>>>> > > >>>>>> And If Ironic cannot find disk since the corresponding drivers in the > > >>>>>> IPA image are missing, do you know how to resolve it? I have used the > > >>>>>> latest deploy images in > > >>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/ > > >>>>>> . Do I need to find and manually add driver in the source code or > > >>>>>> ramdisk(That was difficult tome)? > > >>>>>> > > >>>>>> Love you. > > >>>>>> > > >>>>>> Cheers, > > >>>>>> Guangyu > > >>>>>> > > >>>>>> Arne Wiebalck ?2022?2?10??? 15:51??? > > >>>>>>> > > >>>>>>> Hi Guangyu, > > >>>>>>> > > >>>>>>> The error indicates that Ironic was not able to find > > >>>>>>> a device where it could deploy the image to. > > >>>>>>> > > >>>>>>> To find a device, Ironic will use 'root device' > > >>>>>>> hints [1], usually set by the admin on a node. If that > > >>>>>>> does not yield anything, Ironic will loop over all > > >>>>>>> block devices and pick the smallest which is larger > > >>>>>>> than 4GB (and order them alphabetically). > > >>>>>>> > > >>>>>>> If you have disks in your server which are larger than > > >>>>>>> 4GB, one potential explanation is that Ironic cannot see them, > > >>>>>>> e.g. since the corresponding drivers in the IPA image are missing. > > >>>>>>> The logs you posted seem to confirm something along those > > >>>>>>> lines. > > >>>>>>> > > >>>>>>> Check the content of the 'lsblk' file in the deploy logs which > > >>>>>>> you can find in the tar archive in /var/log/ironic/deploy/ > > >>>>>>> on the controller for your deployment attempt to see what > > >>>>>>> devices Ironic has access to. > > >>>>>>> > > >>>>>>> Cheers, > > >>>>>>> Arne > > >>>>>>> > > >>>>>>> > > >>>>>>> [1] https://docs.openstack.org/ironic/latest/install/advanced.html#root-device-hints > > >>>>>>> > > >>>>>>> On 10.02.22 02:50, ??? wrote: > > >>>>>>>> Dear all, > > >>>>>>>> > > >>>>>>>> I have a OpenStack Victoria environment, and tried to use ironic > > >>>>>>>> manage bare metal. But I got "- root device hints were not provided > > >>>>>>>> and all found block devices are smaller than 4294967296B." in deploy > > >>>>>>>> stage. > > >>>>>>>> > > >>>>>>>> 2022-02-09 17:57:56.492 3908982 ERROR > > >>>>>>>> ironic.drivers.modules.agent_base [-] Agent returned error for deploy > > >>>>>>>> step {'step': 'write_image', 'priority': 80, 'argsinfo': None, > > >>>>>>>> 'interface': 'deploy'} on node cc68c450-ce54-4e1c-be04-8b0a6169ef92 : > > >>>>>>>> No suitable device was found for deployment - root device hints were > > >>>>>>>> not provided and all found block devices are smaller than > > >>>>>>>> 4294967296B.. > > >>>>>>>> > > >>>>>>>> I used "openstack server create --flavor my-baremetal-flavor --nic > > >>>>>>>> net-id=$net_id --image $image testing" to deploy bare metal node. I > > >>>>>>>> download deploy images(ipa-centos8-master.kernel and > > >>>>>>>> ipa-centos8-master.initramfs) in > > >>>>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/. > > >>>>>>>> > > >>>>>>>> The baremetal node info and flavor info as following: > > >>>>>>>> https://paste.opendev.org/show/bV7lgO6RkNQY6ZGPbT2e/ > > >>>>>>>> Ironic configure file as following: > > >>>>>>>> https://paste.opendev.org/show/bTgY9Kpn7KWqwQl73aEa/ > > >>>>>>>> Ironic-conductor log: https://paste.opendev.org/show/bFKZYlXmccxNxU8lEogk/ > > >>>>>>>> The log of ironic-python-agent in bare metal node: > > >>>>>>>> https://paste.opendev.org/show/btAuaMuV2IutV2Pa7YIa/ > > >>>>>>>> > > >>>>>>>> I see some old discussion about this, such as: > > >>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1312187. But those > > >>>>>>>> discussions took place a long time ago, not version V, and no solution > > >>>>>>>> was seen. > > >>>>>>>> > > >>>>>>>> Does anyone know how to solve this problem? I would appreciate any > > >>>>>>>> kind of guidance or help. > > >>>>>>>> > > >>>>>>>> Thank you, > > >>>>>>>> Han Guangyu > > >>>>>>>> > > >>>>> From lokendrarathour at gmail.com Fri Mar 4 09:40:13 2022 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Fri, 4 Mar 2022 15:10:13 +0530 Subject: Fwd: [Triple0] Upgrade Error In-Reply-To: <87333433-63c3-6b52-ab36-8de16dff277d@redhat.com> References: <87333433-63c3-6b52-ab36-8de16dff277d@redhat.com> Message-ID: Hi Harald / Team, Thanks for the support here. It did help in the classification of the issue and we saw that we need to re-render the templates for an upgrade to proceed. After re-rendering we saw the diff of DNS related in the templates. Now we have completed the next few steps and are facing issues when running the below commands: openstack overcloud upgrade run --limit Controller --playbook upgrade_steps_playbook.yaml It gives the error as : *2022-03-04 14:29:56.571690 | 525400cb-2e90-4086-e760-00000000009b | TASK | Ensure DNF modules have the right stream* *2022-03-04 14:29:57.723661 | 525400cb-2e90-4086-e760-00000000009b | FATAL | Ensure DNF modules have the right stream | overcloud-novacompute-0 | item={'module': 'container-tools', 'stream': '3.0'} | error={"ansible_loop_var": "item", "changed": false, "item": {"module": "container-tools", "stream": "3.0"}, "module_stderr": "Traceback (most recent call last):\n File \"\", line 102, in \n File \"\", line 94, in _ansiballz_main\n File \"\", line 40, in invoke_module\n File \"/usr/lib64/python3.6/runpy.py\", line 205, in run_module\n return _run_module_code(code, init_globals, run_name, mod_spec)\n File \"/usr/lib64/python3.6/runpy.py\", line 96, in _run_module_code\n mod_name, mod_spec, pkg_name, script_name)\n File \"/usr/lib64/python3.6/runpy.py\", line 85, in _run_code\n exec(code, run_globals)\n File \"/tmp/ansible_dnf_payload_3lb54omj/ansible_dnf_payload.zip/ansible/modules/packaging/os/dnf.py\", line 1370, in \n File \"/tmp/ansible_dnf_payload_3lb54omj/ansible_dnf_payload.zip/ansible/modules/packaging/os/dnf.py\", line 1359, in main\n File \"/tmp/ansible_dnf_payload_3lb54omj/ansible_dnf_payload.zip/ansible/modules/packaging/os/dnf.py\", line 1338, in run\n File \"/tmp/ansible_dnf_payload_3lb54omj/ansible_dnf_payload.zip/ansible/modules/packaging/os/dnf.py\", line 996, in ensure\n File \"/usr/lib/python3.6/site-packages/dnf/module/module_base.py\", line 144, in install\n \"Installing module from Fail-Safe repository is not allowed\"))\ndnf.exceptions.Error: Installing module from Fail-Safe repository is not allowed\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1}* *Few things to note:* - *the document followed * Upgrading to a Next Major Release ? TripleO 3.0.0 documentation (openstack.org) - The command which was executed to run upgrade prepare: - *openstack overcloud upgrade prepare --templates \ -n /home/stack/templates/network_data.yaml \ -r /home/stack/templates/roles_data.yaml \ -e /home/stack/templates/environment.yaml \ -e /home/stack/templates/environments/network-isolation.yaml \ -e /home/stack/templates/environments/network-environment.yaml \ -e /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-conductor.yaml \ -e /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml \ -e /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-overcloud.yaml \ -e /home/stack/templates/ironic-config.yaml \ -e /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml \ -e /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ -e /home/stack/containers-prepare-parameter.yaml* Please advice over the same issue. Best Regards, Lokendra On Sun, Feb 27, 2022 at 6:50 AM Harald Jensas wrote: > On 2/25/22 12:21, Lokendra Rathour wrote: > > Hi Team, > > > > we are trying to upgrade *Triple0 Train to Ussuri.* > > Undercloud upgrade is done successfully. > > *_OverCloud Upgrade:_* > > As a first step in overcloud upgrade, we are running " *openstack > > overcloud upgrade prepare* " command. > > which is throwing following error: > > > > > overcloud: ERROR: Property error: > > resources.InternalApiVirtualIP.properties: Unknown Property DnsName\n > > File \"/usr/lib/python3.6/site-packages/heat/common/wsgi.py\", line > > 891, in __call__\n request, **action_args)\n File > > \"/usr/lib/python3.6/site-packages/heat/common/wsgi.py\", line 964, in > > dispatch\n return method(*args, **kwargs)\n File > > \"/usr/lib/python3.6/site-packages/heat/api/openstack/v1/util.py\", line > > 56, in handle_stack_method\n return handler(controller, req, > > **kwargs)\n File > > \"/usr/lib/python3.6/site-packages/heat/api/openstack/v1/stacks.py\", > > line 633, in validate_template\n raise > > exc.HTTPBadRequest(result['Error'])\n", "success": false}* > > > > The DnsName property was added to templats in Ussuri, this change: > https://review.opendev.org/c/openstack/tripleo-heat-templates/+/715883 > > > Somewhere in you environment files you are mapping > `OS::TripleO::Network::Ports::InternalApiVipPort` to a local template > file instead of using the in-tree template. > > You either need to add DnsName paramter to your local VipPort template, > or change the resource_registry so that you map to the in-tree port > template. > `/usr/share/openstack-tripleo-heat-template/network/ports/internal_api.yaml` > > is the in-tree template you want. > > > > *_Command used to run the overcloud upgrade prepare:_* > > > > *(undercloud) [stack at undercloud ~]$ cat upgrade_overcloud_prepare.sh > > openstack overcloud upgrade prepare --templates \ > > -r /home/stack/templates/roles_data.yaml \ > > -n /home/stack/templates/network_data.yaml \ > > -e /home/stack/templates/environment.yaml \ > > -e /home/stack/templates/environments/network-isolation.yaml \ > > Most likely the mapping is in this network-isolation.yaml. > If you use the in-tree environment > `/usr/share/openstack-tripleo-heat-tempaltes/envioronments/network-isolation.yaml` > > instead you should not run into these type of issues on upgrade. > > > -e /home/stack/templates/environments/network-environment.yaml \ > > I would also recommend to use the in-tree network-environment.yaml > -e > > /usr/share/openstack-tripleo-heat-tempaltes/envioronments/network-environment.yaml > > Then add another file to override anything in the in-tree > network-environment.yaml. > -e /home/stack/templates/environments/network-environment-overrides.yaml > > > -e > > > /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-conductor.yaml > > > \ > > -e > > > /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml > > > \ > > -e > > > /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-overcloud.yaml > > > \ > > -e /home/stack/templates/ironic-config.yaml \ > > -e > > /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml \ > > -e > > /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ > > -e /home/stack/templates/init-repo.yaml \ > > -e /home/stack/containers-prepare-parameter.yaml > > > -- > Harald > > -- ~ Lokendra -------------- next part -------------- An HTML attachment was scrubbed... URL: From elfosardo at gmail.com Fri Mar 4 10:36:47 2022 From: elfosardo at gmail.com (Riccardo Pittau) Date: Fri, 4 Mar 2022 11:36:47 +0100 Subject: [ironic] Zed PTG slots In-Reply-To: References: Message-ID: Thanks Iury! My choices are for options #1 and #3 Ciao! Riccardo On Fri, Mar 4, 2022 at 4:28 AM Iury Gregory wrote: > Hello Ironicers! > > For the Zed PTG I'm planning to have some slots that will allow APAC > contributors to participate in the discussions. > > Keep in mind that we have some guidelines we should try to follow: > 1. Cross project discussions (like SIGs or support project teams) should > be scheduled towards the start of the week so that any discussions that > might shape those of other teams happen first. > 2. No team should sign up for more than 4 hours per UTC day to help keep > participants actively engaged. > 3. No team should sign up for more than 16 hours across all time slots to > avoid burning out our contributors and to enable participation in multiple > teams discussions. > > Based on this I have the following proposals: (Feel free to provide more > ideas also) =) > *Option #1* *16hrs* Monday till Thursday > 1h (APAC night + EMEA morning) > 2hrs (EMEA afternoon + NA morning) > 1hr (NA afternoon/night + APAC morning) > > *Option #2* *15hrs* Tuesday till Friday > 1h (APAC night + EMEA morning) > 2hrs (EMEA afternoon + NA morning) > 1hr (NA afternoon/night + APAC morning) > * On Friday the NA + APAC session won't happen because is saturday for > APAC contributors * > > *Option #3* *16hrs* Monday till Thursday > 3hrs (EMEA afternoon + NA morning) > 1hr (NA afternoon/night + APAC morning) > > *Option #4* *15hrs* Tuesday till Friday > 3hrs (EMEA afternoon + NA morning) > 1hr (NA afternoon/night + APAC morning) > * On Friday the NA + APAC session won't happen because is saturday for > APAC contributors * > > Please choose two options and we will vote during our upstream meeting on > March 7th. > > Thank you! > > -- > > > *Att[]'sIury Gregory Melo Ferreira * > *MSc in Computer Science at UFCG* > *Part of the ironic-core and puppet-manager-core team in OpenStack* > *Software Engineer at Red Hat Czech* > *Social*: https://www.linkedin.com/in/iurygregory > *E-mail: iurygregory at gmail.com * > -------------- next part -------------- An HTML attachment was scrubbed... URL: From sbauza at redhat.com Fri Mar 4 10:51:36 2022 From: sbauza at redhat.com (Sylvain Bauza) Date: Fri, 4 Mar 2022 11:51:36 +0100 Subject: [nova][placement] Zed PTG preparation Message-ID: Hey folks, We have until Mar 11th [1]to communicate about our needs for the Zed PTG (I've already signed up our presence by the survey) .As you can see in [2], there are already projects that booked some timeslots. This time, I won't ask you to tell me over Doodle which times you can be around, but I'd rather propose you 3 options : Option A "The short one" : we ask for 3 days from Wednesday to Friday with 4 hours each (13UTC-17UTC) Option B "The week used": we ask for 4 days from Tuesday to Friday with 3 hours each (14UTC-17UTC) Option C "The largest one" : we ask for 4 days from Tuesday to Friday with 4 hours each (13UTC-17UTC) Option D "The Asian-nice one" : we ask for Wed to Friday with 3 hours each (14UTC-17UTC) but we also ask for Thursday with 2 hours (6UTC-8UTC) Please start to think about those options as we'll discuss this on next Tuesday meeting [3] and eventually we will vote on it. I also created a new etherpad for our PTG : https://etherpad.opendev.org/p/nova-zed-ptg If you want, you can start to collect some topics but also if you aren't able to attend next Nova meeting, you can also start to vote on the option above in there. Voila, that's it. Thanks and let's discuss it again on next meeting, -Sylvain [1] http://lists.openstack.org/pipermail/openstack-discuss/2022-February/027051.html [2] https://ethercalc.openstack.org/7yxdas7suqnd [3] https://meetings.opendev.org/#Nova_Team_Meeting -------------- next part -------------- An HTML attachment was scrubbed... URL: From juliaashleykreger at gmail.com Fri Mar 4 14:03:09 2022 From: juliaashleykreger at gmail.com (Julia Kreger) Date: Fri, 4 Mar 2022 06:03:09 -0800 Subject: [Ironic] No suitable device was found for deployment In-Reply-To: References: <04af20bb-143b-9b56-6787-d5ac6ab0d38c@cern.ch> <3209d9fc-8db1-6dca-5736-402c6f082fcc@cern.ch> Message-ID: Seems like your issue is in raid controller configuration. And based on the MegaCli output, I doubt the disks are security locked. I would remove the disks, get into the bios/firmware menus, and try to delete them, then try to re-import the disks. Alternatively, it may be worthwhile to take the disks and attempt to use them in a desktop or something where a MegaRAID controller is not in the mix in case it is something with configuration on the disk as well. Examining outside of a MegaRAID controller would hopefully give you an idea of operational state as well. -Julia On Fri, Mar 4, 2022 at 1:25 AM ??? wrote: > > Hi Julia, > > Sorry that my last email didn't reploy some question in your email. > When I get into RAID config menu, I said disk state is "unconfig bad". > And the more info is that Virtual Drive Management displayed "No > Configuration Present!". But I cound not modify disk state in RAID > confug menu. Even if I moved the disk to other server, and used the > software in operating system to modify it, it still coudn't to be > modify. such as: > ```shell > # MegaCli -PDList -a0 > > Adapter #0 > > Enclosure Device ID: 252 > Slot Number: 1 > Enclosure position: 0 > Device Id: 9 > WWN: 50014EE2BE72FEBF > Sequence Number: 2 > Media Error Count: 0 > Other Error Count: 0 > Predictive Failure Count: 0 > Last Predictive Failure Event Seq Number: 0 > PD Type: SATA > Raw Size: 0 KB [0x0 Sectors] > Non Coerced Size: 0 KB [0x0 Sectors] > Coerced Size: 0 KB [0x0 Sectors] > Firmware state: Unconfigured(bad) > Device Firmware Level: WA09 > Shield Counter: 0 > Successful diagnostics completion on : N/A > SAS Address(0): 0x4433221101000000 > Connected Port Number: 0(path0) > Inquiry Data: ATA HGST HUS722T2TALWA09WCC6N4HZV9SX > FDE Enable: Disable > Secured: Unsecured > Locked: Unlocked > Needs EKM Attention: No > Foreign State: None > Device Speed: Unknown > Link Speed: 6.0Gb/s > Media Type: Hard Disk Device > Drive Temperature :0C (32.00 F) > PI Eligibility: No > Drive is formatted for PI information: No > PI: No PI > Drive's write cache : Disabled > Drive's NCQ setting : Enabled > Port-0 : > Port status: Active > Port's Linkspeed: 6.0Gb/s > Drive has flagged a S.M.A.R.T alert : No > > > > Enclosure Device ID: 252 > Slot Number: 2 > Enclosure position: 0 > Device Id: 10 > WWN: 50014EE2691D724C > Sequence Number: 2 > Media Error Count: 0 > Other Error Count: 0 > Predictive Failure Count: 0 > Last Predictive Failure Event Seq Number: 0 > PD Type: SATA > Raw Size: 0 KB [0x0 Sectors] > Non Coerced Size: 0 KB [0x0 Sectors] > Coerced Size: 0 KB [0x0 Sectors] > Firmware state: Unconfigured(bad) > Device Firmware Level: WA09 > Shield Counter: 0 > Successful diagnostics completion on : N/A > SAS Address(0): 0x4433221102000000 > Connected Port Number: 1(path0) > Inquiry Data: ATA HGST HUS722T2TALWA09WCC6N4HZVTV5 > FDE Enable: Disable > Secured: Unsecured > Locked: Unlocked > Needs EKM Attention: No > Foreign State: None > Device Speed: Unknown > Link Speed: 6.0Gb/s > Media Type: Hard Disk Device > Drive Temperature :0C (32.00 F) > PI Eligibility: No > Drive is formatted for PI information: No > PI: No PI > Drive's write cache : Disabled > Drive's NCQ setting : Enabled > Port-0 : > Port status: Active > Port's Linkspeed: 6.0Gb/s > Drive has flagged a S.M.A.R.T alert : No > > > > Enclosure Device ID: 252 > Slot Number: 3 > Enclosure position: 0 > Device Id: 11 > WWN: 50014EE2BE733A5E > Sequence Number: 2 > Media Error Count: 0 > Other Error Count: 0 > Predictive Failure Count: 0 > Last Predictive Failure Event Seq Number: 0 > PD Type: SATA > Raw Size: 1.819 TB [0xe8e088b0 Sectors] > Non Coerced Size: 1.818 TB [0xe8d088b0 Sectors] > Coerced Size: 1.818 TB [0xe8d00000 Sectors] > Firmware state: JBOD > Device Firmware Level: WA09 > Shield Counter: 0 > Successful diagnostics completion on : N/A > SAS Address(0): 0x4433221103000000 > Connected Port Number: 2(path0) > Inquiry Data: WCC6N0KX0HJD HGST HUS722T2TALA604 > RAGNWA09 > FDE Enable: Disable > Secured: Unsecured > Locked: Unlocked > Needs EKM Attention: No > Foreign State: None > Device Speed: 6.0Gb/s > Link Speed: 6.0Gb/s > Media Type: Hard Disk Device > Drive Temperature :34C (93.20 F) > PI Eligibility: No > Drive is formatted for PI information: No > PI: No PI > Drive's write cache : Enabled > Drive's NCQ setting : Enabled > Port-0 : > Port status: Active > Port's Linkspeed: 6.0Gb/s > Drive has flagged a S.M.A.R.T alert : No > > > > > Exit Code: 0x00 > ``` > ```shell > # MegaCli -PDList -a0 | grep state > Firmware state: Unconfigured(bad) > Firmware state: Unconfigured(bad) > Firmware state: JBOD > # MegaCli -PDMakeGood -PhysDrv[252:2] -a0 > > Adapter: 0: Failed to change PD state at EnclId-252 SlotId-2. > > Exit Code: 0x01 > ``` > > And the "security locked state" is a idea that I continue. I will try > to determine if it is in this state and find a way to disarm it. > > Thank you very much Julia, > cheers, > Han Guangyu > > Julia Kreger ?2022?3?1??? 22:06??? > > > > On Mon, Feb 28, 2022 at 1:12 AM Arne Wiebalck wrote: > > > > > > Hi Guangyu, > > > > > > I am not aware of anything in the Ironic Python Agent that > > > would remove disks from the system in a way that they would > > > not be visible after a reboot (apart from, as mentioned before, > > > the clean up of a hardware RAID in a way the IPA is not able > > > to see any devices after). > > > > > > How about trying to access and configure the hardware RAID with > > > the corresponding tool from the RAM disk you booted into from the > > > USB stick? Install the tool and see if it detects the controller. > > > > > > The very first step before doing anything with Ironic is to > > > get the disks back or understand why they are not visible. > > > > > > > Did cleaning fail at any given point with these machines? > > > > If you have physical access, try disconnecting all of the drives, and > > then powering up the machine and see if you can get into the firmware > > configuration screen with control-h. If you can, remove all of the > > prior configuration or disk volumes. They will look like they are in > > error states most likely. If your unable to get into this screen, I > > would be worried about your disk controller card. If your able to > > clear everything out of the controller, power off, try re-inserting > > drives, and see what happens. See if the controller can view/interact > > with the drives. If it sees no drives, then my next paragraph is > > likely the case. > > > > The disks sound like they might be in security locked state which will > > likely require a desktop SATA disk controller to remedy by attaching > > and manually removing from a security locked state. Megaraid > > controllers can't recognize security locked devices (most controllers > > and especially ones labeled "raid controllers" can't handle it) when > > in pass-through mode, but I've never heard of security lock commands > > actually getting through to the device with those controllers in > > pass-through mode. If the card was in raid mode to begin with, then it > > likely never did anything involving secure erase as the controller > > should not be offering that as a feature of provided disks to the OS. > > > > > Cheers, > > > Arne > > > > > > On 28.02.22 09:28, ??? wrote: > > > > Hi Arne, > > > > > > > > I didn't find hardware RAID config option during the initial boot > > > > sequence. Ctrl+H is unresponsive in this computer. I just saw "Press > > > > Del to enter firmware configuration, press F3 to enter boot menu, and > > > > press F12 to enter network boot". And I press 'Del' to enter the BIOS. > > > > But I didn't find RAID config menu in BIOS. Sorry that I have poor > > > > knowledge about this. > > > > > > > > And now, even though I installed the operating system manually using a > > > > USB stick, I still couldn't find the hard drive. Is there anything > > > > that ironic-agent did in the clean phase that would have caused this > > > > problem? > > > > > > > > I wonder if this is a thinking pointto solve the problem. Now, my idea > > > > is to first find a way to manually configure RAID. Do you agree with > > > > this? And than, whether RAID configurations are still cleared in the > > > > Clean phase if clean phase will do this? > > > > > > > > Sorry that I have so much confuse. > > > > > > > > love you, > > > > Guangyu > > > > > > > > Arne Wiebalck ?2022?2?14??? 15:59??? > > > >> > > > >> Hi Guangyu, > > > >> > > > >> It seems like Julia had the right idea and the disks > > > >> are not visible since the RAID controller does not > > > >> expose anything to the operating system. This seems > > > >> to be confirmed by you booting into the CentOS7 image. > > > >> > > > >> What I would suggest to try next is to look for the > > > >> hardware RAID config option during the initial boot > > > >> sequence to enter the RAID config menu (there should be > > > >> a message quite early on, and maybe Ctrl-H is needed > > > >> to enter the menu). > > > >> > > > >> Once there, manually configure the disks as JBODs or > > > >> create a RAID device. Upon reboot this should be visible > > > >> and accessible as a device. Maybe check from your CentOS7 > > > >> image again. If the devices are there, Ironic should > > > >> also be able to deploy on them (for this you can remove > > > >> the RAID config you added). > > > >> > > > >> It depends a little on what your goal is, but I would > > > >> try this first to see if you can make a device visible > > > >> and if the Ironic deploy bit works, before trying to > > > >> configure the hardware RAID via Ironic. > > > >> > > > >> Cheers, > > > >> Arne > > > >> > > > >> On 14.02.22 03:20, ??? wrote: > > > >>> Hi Arne and Julia, > > > >>> > > > >>> You make me feel so warm. Best wishes to you. > > > >>> > > > >>> I have tried to boot the node into a CentOS7, but it still coundnot to > > > >>> find disk. And sorry that I didn't notice the RAID card. > > > >>> > > > >>> # lspci -v > > > >>> ... > > > >>> 23:00.0 RAID bus controller: Broadcom / LSI MegaRAID SAS-3 3108 > > > >>> [Invader] (rev 02) > > > >>> Subsystem: Broadcom / LSI MegaRAID SAS 9361-8i > > > >>> Flags: bus master, fast devsel, latency 0, IRQ -2147483648, NUMA node 1 > > > >>> I/O ports at 3000 [size=256] > > > >>> Memory at e9900000 (64-bit, non-prefetchable) [size=64K] > > > >>> Memory at e9700000 (64-bit, non-prefetchable) [size=1M] > > > >>> Expansion ROM at e9800000 [disabled] [size=1M] > > > >>> Capabilities: [50] Power Management version 3 > > > >>> Capabilities: [68] Express Endpoint, MSI 00 > > > >>> Capabilities: [d0] Vital Product Data > > > >>> Capabilities: [a8] MSI: Enable- Count=1/1 Maskable+ 64bit+ > > > >>> Capabilities: [c0] MSI-X: Enable+ Count=97 Masked- > > > >>> Capabilities: [100] Advanced Error Reporting > > > >>> Capabilities: [1e0] #19 > > > >>> Capabilities: [1c0] Power Budgeting > > > >>> Capabilities: [148] Alternative Routing-ID Interpretation (ARI) > > > >>> Kernel driver in use: megaraid_sas > > > >>> Kernel modules: megaraid_sas > > > >>> ... > > > >>> > > > >>> I try to config raid fallowing > > > >>> https://docs.openstack.org/ironic/latest/admin/raid.html > > > >>> by `baremetal node set $NODE_UUID --target-raid-config raid.json`. The > > > >>> server have three same disk(Western Digital DC HA210 2TB SATA 6GB/s) > > > >>> # cat raid.json > > > >>> { > > > >>> "logical_disks": [ > > > >>> { > > > >>> "size_gb": "MAX", > > > >>> "raid_level": "0", > > > >>> "is_root_volume": true > > > >>> } > > > >>> ] > > > >>> } > > > >>> > > > >>> But Ironic still coundn't see disk. I still got > > > >>> ``` > > > >>> ## In deploy images > > > >>> # journalctl -fxeu ironic-python-agent > > > >>> Feb 14 02:17:22 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 > > > >>> 02:17:22.863 2329 WARNING root [-] Path /dev/disk/by-path is > > > >>> inaccessible, /dev/disk/by-path/* version of block device name is > > > >>> unavailable Cause: [Errno 2] No such file or directory: > > > >>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or > > > >>> directory: '/dev/disk/by-path' > > > >>> Feb 14 02:17:44 host-10-12-22-74 ironic-python-agent[2329]: 2022-02-14 > > > >>> 02:17:44.391 2329 ERROR root [-] Unexpected error dispatching > > > >>> get_os_install_device to manager > > > >>> > > >>> 0x7efbf4da2208>: Error finding the disk or partition device to deploy > > > >>> the image onto: No suitable device was found for deployment - root > > > >>> device hints were not provided and all found block devices are smaller > > > >>> than 4294967296B.: ironic_python_agent.errors.DeviceNotFound: Error > > > >>> finding the disk or partition device to deploy the image onto: No > > > >>> suitable device was found for deployment - root device hints were not > > > >>> provided and all found block devices are smaller than 4294967296B. > > > >>> ``` > > > >>> > > > >>> I don't know if it's a lack of a RAID card driver or a lack of a disk > > > >>> driver or a lack of RAID configuration. Could you have some idea about > > > >>> this question? > > > >>> > > > >>> love you, > > > >>> Han Guangyu > > > >>> > > > >>> > > > >>> Julia Kreger ?2022?2?10??? 23:11??? > > > >>> > > > >>>> > > > >>>> If the disk controllers *are* enumerated in the kernel log, which is > > > >>>> something to also look for, then the disks themselves may be in some > > > >>>> weird state like security locked. Generally this shows up as the > > > >>>> operating system kind of sees the disk and the SATA port connected but > > > >>>> can't really access it. This is also an exceptionally rare state to > > > >>>> find one's self in. > > > >>>> > > > >>>> More common, especially in enterprise grade hardware: If the disk > > > >>>> controller is actually a raid controller, and there are no raid > > > >>>> volumes configured, then the operating system likely cannot see the > > > >>>> underlying disks and turn that into a usable block device. I've seen a > > > >>>> couple drivers over the years which expose hints of disks in the > > > >>>> kernel log and without raid configuration in the cards, the drivers > > > >>>> can't present usable block devices to the operating system system. > > > >>>> > > > >>>> -Julia > > > >>>> > > > >>>> On Thu, Feb 10, 2022 at 3:17 AM Arne Wiebalck wrote: > > > >>>>> > > > >>>>> Hi Guangyu, > > > >>>>> > > > >>>>> No worries about asking questions, this is what the mailing > > > >>>>> list is for :) > > > >>>>> > > > >>>>> Just to clarify, you do not have to set root device hints, > > > >>>>> it also works without (with the algorithm I mentioned). > > > >>>>> However, hints help to define the exact device and/or make > > > >>>>> deployment more predictable/repeatable. > > > >>>>> > > > >>>>> If it is really a driver problem, it is an issue with the > > > >>>>> operating system of the image you use, i.e. CentOS8. Some > > > >>>>> drivers were removed from 7 to 8, and we have seen issues > > > >>>>> with specific drive models as well. > > > >>>>> > > > >>>>> You can try to build your own IPA images as described in > > > >>>>> [1], e.g. to add your ssh key to be able to log into the > > > >>>>> IPA to debug further, and to eventually include drivers > > > >>>>> (if you can identify them and they are available for CentOS8). > > > >>>>> > > > >>>>> Another option may be to add another (newer) disk model to > > > >>>>> the server, just to confirm it is the disk model/driver which > > > >>>>> is the cause. > > > >>>>> > > > >>>>> You could also try to boot the node into a CentOS7 (and then > > > >>>>> a CentOS8) live image to confirm it can see the disks at all. > > > >>>>> > > > >>>>> Hope this helps! > > > >>>>> Arne > > > >>>>> > > > >>>>> [1] > > > >>>>> https://docs.openstack.org/ironic-python-agent-builder/latest/admin/dib.html > > > >>>>> > > > >>>>> > > > >>>>> On 10.02.22 11:15, ??? wrote: > > > >>>>>> Hi Arne, > > > >>>>>> > > > >>>>>> Thank you very much for your response. Love you. You take away a lot > > > >>>>>> of my confusion. > > > >>>>>> > > > >>>>>> You are right, I didn't set 'root device'. And Ironic also can not see > > > >>>>>> disk, the content of the 'lsblk' file in the deploy los is emply. > > > >>>>>> I tried to set 'root device', but because ironic can't find any disk, > > > >>>>>> the deploy still filed. > > > >>>>>> > > > >>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 > > > >>>>>> 09:51:55.045 2324 WARNING root [-] Path /dev/disk/by-path is > > > >>>>>> inaccessible, /dev/disk/by-path/* version of block device name is > > > >>>>>> unavailable Cause: [Errno 2] No such file or directory: > > > >>>>>> '/dev/disk/by-path': FileNotFoundError: [Errno 2] No such file or > > > >>>>>> directory: '/dev/disk/by-path' > > > >>>>>> Feb 10 09:51:55 host-10-12-22-59 ironic-python-agent[2324]: 2022-02-10 > > > >>>>>> 09:51:55.056 2324 WARNING ironic_lib.utils [-] No device found that > > > >>>>>> matches the root device hints {'wwn': '0x50014EE2691D724C'}: > > > >>>>>> StopIteration > > > >>>>>> > > > >>>>>> Sorry to bother you, I'm a newcomer of Ironic and I didn't find > > > >>>>>> information about it on google. > > > >>>>>> > > > >>>>>> The bare metal node have three same disk(Western Digital DC HA210 2TB > > > >>>>>> SATA 6GB/s). Where I can confirm whether ironic-python-agent supports > > > >>>>>> this disk? > > > >>>>>> > > > >>>>>> And If Ironic cannot find disk since the corresponding drivers in the > > > >>>>>> IPA image are missing, do you know how to resolve it? I have used the > > > >>>>>> latest deploy images in > > > >>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/ > > > >>>>>> . Do I need to find and manually add driver in the source code or > > > >>>>>> ramdisk(That was difficult tome)? > > > >>>>>> > > > >>>>>> Love you. > > > >>>>>> > > > >>>>>> Cheers, > > > >>>>>> Guangyu > > > >>>>>> > > > >>>>>> Arne Wiebalck ?2022?2?10??? 15:51??? > > > >>>>>>> > > > >>>>>>> Hi Guangyu, > > > >>>>>>> > > > >>>>>>> The error indicates that Ironic was not able to find > > > >>>>>>> a device where it could deploy the image to. > > > >>>>>>> > > > >>>>>>> To find a device, Ironic will use 'root device' > > > >>>>>>> hints [1], usually set by the admin on a node. If that > > > >>>>>>> does not yield anything, Ironic will loop over all > > > >>>>>>> block devices and pick the smallest which is larger > > > >>>>>>> than 4GB (and order them alphabetically). > > > >>>>>>> > > > >>>>>>> If you have disks in your server which are larger than > > > >>>>>>> 4GB, one potential explanation is that Ironic cannot see them, > > > >>>>>>> e.g. since the corresponding drivers in the IPA image are missing. > > > >>>>>>> The logs you posted seem to confirm something along those > > > >>>>>>> lines. > > > >>>>>>> > > > >>>>>>> Check the content of the 'lsblk' file in the deploy logs which > > > >>>>>>> you can find in the tar archive in /var/log/ironic/deploy/ > > > >>>>>>> on the controller for your deployment attempt to see what > > > >>>>>>> devices Ironic has access to. > > > >>>>>>> > > > >>>>>>> Cheers, > > > >>>>>>> Arne > > > >>>>>>> > > > >>>>>>> > > > >>>>>>> [1] https://docs.openstack.org/ironic/latest/install/advanced.html#root-device-hints > > > >>>>>>> > > > >>>>>>> On 10.02.22 02:50, ??? wrote: > > > >>>>>>>> Dear all, > > > >>>>>>>> > > > >>>>>>>> I have a OpenStack Victoria environment, and tried to use ironic > > > >>>>>>>> manage bare metal. But I got "- root device hints were not provided > > > >>>>>>>> and all found block devices are smaller than 4294967296B." in deploy > > > >>>>>>>> stage. > > > >>>>>>>> > > > >>>>>>>> 2022-02-09 17:57:56.492 3908982 ERROR > > > >>>>>>>> ironic.drivers.modules.agent_base [-] Agent returned error for deploy > > > >>>>>>>> step {'step': 'write_image', 'priority': 80, 'argsinfo': None, > > > >>>>>>>> 'interface': 'deploy'} on node cc68c450-ce54-4e1c-be04-8b0a6169ef92 : > > > >>>>>>>> No suitable device was found for deployment - root device hints were > > > >>>>>>>> not provided and all found block devices are smaller than > > > >>>>>>>> 4294967296B.. > > > >>>>>>>> > > > >>>>>>>> I used "openstack server create --flavor my-baremetal-flavor --nic > > > >>>>>>>> net-id=$net_id --image $image testing" to deploy bare metal node. I > > > >>>>>>>> download deploy images(ipa-centos8-master.kernel and > > > >>>>>>>> ipa-centos8-master.initramfs) in > > > >>>>>>>> https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/. > > > >>>>>>>> > > > >>>>>>>> The baremetal node info and flavor info as following: > > > >>>>>>>> https://paste.opendev.org/show/bV7lgO6RkNQY6ZGPbT2e/ > > > >>>>>>>> Ironic configure file as following: > > > >>>>>>>> https://paste.opendev.org/show/bTgY9Kpn7KWqwQl73aEa/ > > > >>>>>>>> Ironic-conductor log: https://paste.opendev.org/show/bFKZYlXmccxNxU8lEogk/ > > > >>>>>>>> The log of ironic-python-agent in bare metal node: > > > >>>>>>>> https://paste.opendev.org/show/btAuaMuV2IutV2Pa7YIa/ > > > >>>>>>>> > > > >>>>>>>> I see some old discussion about this, such as: > > > >>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1312187. But those > > > >>>>>>>> discussions took place a long time ago, not version V, and no solution > > > >>>>>>>> was seen. > > > >>>>>>>> > > > >>>>>>>> Does anyone know how to solve this problem? I would appreciate any > > > >>>>>>>> kind of guidance or help. > > > >>>>>>>> > > > >>>>>>>> Thank you, > > > >>>>>>>> Han Guangyu > > > >>>>>>>> > > > >>>>> From iurygregory at gmail.com Fri Mar 4 15:05:40 2022 From: iurygregory at gmail.com (Iury Gregory) Date: Fri, 4 Mar 2022 12:05:40 -0300 Subject: [requirements][release][rfe] sushy 4.1.1 for Yoga Message-ID: Hello Stackers, I would like to request a new release for sushy before creating the yoga branch, It contains some important fixes (e.g it will fix a problem when build sushy in Debian). The sushy lib is only used by ironic, so it shouldn't cause any problems updating the version in requirements. Thanks! -- *Att[]'sIury Gregory Melo Ferreira * *MSc in Computer Science at UFCG* *Part of the ironic-core and puppet-manager-core team in OpenStack* *Software Engineer at Red Hat Czech* *Social*: https://www.linkedin.com/in/iurygregory *E-mail: iurygregory at gmail.com * -------------- next part -------------- An HTML attachment was scrubbed... URL: From iurygregory at gmail.com Fri Mar 4 16:11:51 2022 From: iurygregory at gmail.com (Iury Gregory) Date: Fri, 4 Mar 2022 13:11:51 -0300 Subject: [requirements][release][rfe] sushy 4.1.1 for Yoga In-Reply-To: <20220304152607.tkkvd7qhyl6o5ppu@domaintools.com> References: <20220304152607.tkkvd7qhyl6o5ppu@domaintools.com> Message-ID: Hi Matthew You can find the diff in [1], thanks! [1] https://zuul.opendev.org/t/openstack/build/4894a7e4d02e4892adab5ca4cefc8e89/log/tox/list-changes-results.log#280-408 Em sex., 4 de mar. de 2022 ?s 12:26, Matthew Thode escreveu: > On 22-03-04 12:05:40, Iury Gregory wrote: > > Hello Stackers, > > > > I would like to request a new release for sushy before creating the yoga > > branch, It contains some important fixes (e.g it will fix a problem when > > build sushy in Debian). The sushy lib is only used by ironic, so it > > shouldn't cause any problems updating the version in requirements. > > > > Thanks! > > > > -- > > > > > > *Att[]'sIury Gregory Melo Ferreira * > > *MSc in Computer Science at UFCG* > > *Part of the ironic-core and puppet-manager-core team in OpenStack* > > *Software Engineer at Red Hat Czech* > > *Social*: https://www.linkedin.com/in/iurygregory > > *E-mail: iurygregory at gmail.com * > > Hi, can you link to the diff? > > -- > Matthew Thode > -- *Att[]'sIury Gregory Melo Ferreira * *MSc in Computer Science at UFCG* *Part of the ironic-core and puppet-manager-core team in OpenStack* *Software Engineer at Red Hat Czech* *Social*: https://www.linkedin.com/in/iurygregory *E-mail: iurygregory at gmail.com * -------------- next part -------------- An HTML attachment was scrubbed... URL: From Molka.Gharbaoui at santannapisa.it Fri Mar 4 16:21:45 2022 From: Molka.Gharbaoui at santannapisa.it (Molka Gharbaoui) Date: Fri, 4 Mar 2022 16:21:45 +0000 Subject: R: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: Unfortunately, the local.conf file shared by Matthias gives the same result: no metrics are pushed to gnocchi. Gnocchi database is empty: I can create metrics and resources manually but nothing is pushed from ceilometer to the database. So I guess the problem is at the ceilometer level. Here is an extract of the ceilometer-polling logs: 2022-03-04 16:13:07.904 679128 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of all_pollsters 2022-03-04 16:13:07.904 679128 DEBUG ceilometer.compute.pollsters [-] afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.usage volume: 5368709120 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/__init__.py:108 2022-03-04 16:13:07.907 679128 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of all_pollsters 2022-03-04 16:13:07.908 679128 DEBUG ceilometer.compute.pollsters [-] afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.read.requests volume: 6138 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/__init__.py:108 2022-03-04 16:13:08.038 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.038 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.pool, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.162 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.162 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.vip, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.163 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.outgoing.bytes, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] Skipping tenant, keystone issue: admin endpoint for identity service in RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: admin endpoint for identity service in RegionOne region not found 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster radosgw.containers.objects.size, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster port.receive.packets, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster hardware.network.ip.outgoing.datagrams, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.355 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.355 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.listener, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 And here I report the content of the ceilometer.conf file: [DEFAULT] debug = True transport_url = rabbit://stackrabbit:admin at 10.30.2.194:5672/ [oslo_messaging_notifications] topics = notifications [coordination] backend_url = redis://localhost:6379 [notification] workers = 2 [cache] backend_argument = url:redis://localhost:6379 backend_argument = distributed_lock:True backend_argument = db:0 backend_argument = redis_expiration_time:600 backend = dogpile.cache.redis enabled = True [service_credentials] auth_url = http://10.30.2.194/identity region_name = RegionOne password = admin username = ceilometer project_name = service project_domain_name = default user_domain_id = default auth_type = password [keystone_authtoken] memcached_servers = localhost:11211 cafile = /opt/stack/data/ca-bundle.pem project_domain_name = default project_name = admin user_domain_name = default password = admin username = ceilometer auth_url = http://10.30.2.194/identity interface = public auth_type = password I believe nothing should be changed in the configuration files since Devstack should configure every components automatically. I also checked pipeline.yaml and polling.yaml files and they seem correct. Do you have any idea on how to make ceilometer push metrics to the gnocchi database?? Thanks. ________________________________ Da: Matthias Runge Inviato: gioved? 3 marzo 2022 13:58 A: Molka Gharbaoui Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi On 3/3/22 12:31, Molka Gharbaoui wrote: > Hi Matthias, > > Sorry for the confusion. Gnocchi is running and I am able to execute > commands like "gnocchi metric list" and "gnocchi resource list" without > errors even though the response is empty. Also checking mysql, the > gnocchi database is correctly created. > > By the link I previously shared I wanted just to highlight that there > are some differences between what was written there and between the > environemnt I installed. I was searching there because I would have > expected that Devstack installs and configures every component in > automatic with no or minimum external intervention (in fact if I don't > consider the Telemetry aspect, the rest works perfectly fine). > > I will follow Rafael guidelines to better undrstand where is the issue. > > Thanks. Something I have used in the past is a local.conf like this: [[local|localrc]] ADMIN_PASSWORD=secret DATABASE_PASSWORD=$ADMIN_PASSWORD RABBIT_PASSWORD=$ADMIN_PASSWORD SERVICE_PASSWORD=$ADMIN_PASSWORD IMAGE_URL_SITE="https://download.fedoraproject.org" IMAGE_URL_PATH="/pub/fedora/linux/releases/33/Cloud/x86_64/images/" IMAGE_URL_FILE="Fedora-Cloud-Base-33-1.2.x86_64.qcow2" IMAGE_URLS+=","$IMAGE_URL_SITE$IMAGE_URL_PATH$IMAGE_URL_FILE CEILOMETER_BACKEND=gnocchi USE_PYTHON3="True" VERBOSE="True" VERBOSE_NO_TIMESTAMP="True" # LIBS_FROM_GIT=aodh,ceilometer,cinder,devstack,devstack-gate,dib-utils,diskimage-builder,glance,heat,keystone,neutron,nova,placement,requirements,swift enable_plugin aodh https://opendev.org/openstack/aodh enable_plugin ceilometer https://opendev.org/openstack/ceilometer enable_plugin heat https://opendev.org/openstack/heat INSTALL_TEMPEST=False Hope that helps. The IMAGE_URL_FILE needs probably adjustment, fedora 33 is not supported anymore. Matthias > > > ------------------------------------------------------------------------ > *Da:* Matthias Runge > *Inviato:* gioved? 3 marzo 2022 11:59 > *A:* openstack-discuss at lists.openstack.org > > *Oggetto:* Re: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > On 3/3/22 10:29, Molka Gharbaoui wrote: >> Thank you Matthias for your reply. >> >> Unfortunately, from the logs I still cannot understand what is the issue. >> >> Could someone kindly indicate me how ceilometer and gnocchi should be >> configured to get the data? >> >> For example in this link: >> https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > >> > the > >> endpoints should use the following URL: http://controller:8041 >> > while in my case the > 8041 port is not even >> listening and the API indicated in the dashboard for Gnocchi is http:// >> >x.x.x.x/metric. >> >> Thank you in advance. >> > > OpenStack configuration is complex. The guide does not remotely touch on > production environments. Please do not confuse devstack with production > envionments. Also the doc you linked is not for devstack itself. > > The flow in general is: data is generated by the services like nova or > glance. Ceilometer listens on the bus and also polls for information. > The info is sent then over (as configured in publishers.yaml) to gnocchi. > Gnocchi has an api, ceilometer has not. Reading your response, this > makes me suspicious if there is gnocchi running. From the initial > question it suggests it is, from your most recent response, it is not. > > > Matthias > -------------- next part -------------- An HTML attachment was scrubbed... URL: From rafaelweingartner at gmail.com Fri Mar 4 16:28:50 2022 From: rafaelweingartner at gmail.com (=?UTF-8?Q?Rafael_Weing=C3=A4rtner?=) Date: Fri, 4 Mar 2022 13:28:50 -0300 Subject: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: What do you or what would you expect to see in Gnocchi being pushed by Ceilometer? What are the resources (volumes, VMs, and so) that you created in the OpenStack environment in DevStack? On Fri, Mar 4, 2022 at 1:26 PM Molka Gharbaoui < Molka.Gharbaoui at santannapisa.it> wrote: > Unfortunately, the local.conf file shared by Matthias gives the same > result: no metrics are pushed to gnocchi. > > Gnocchi database is empty: I can create metrics and resources manually but > nothing is pushed from ceilometer to the database. So I guess the problem > is at the ceilometer level. > > Here is an extract of the ceilometer-polling logs: > > 2022-03-04 16:13:07.904 679128 INFO ceilometer.polling.manager [-] Polling > pollster disk.device.usage in the context of all_pollsters > 2022-03-04 16:13:07.904 679128 DEBUG ceilometer.compute.pollsters [-] > afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.usage volume: 5368709120 > _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/_ > _init__.py:108 > 2022-03-04 16:13:07.907 679128 INFO ceilometer.polling.manager [-] Polling > pollster disk.device.read.requests in the context of all_pollsters > 2022-03-04 16:13:07.908 679128 DEBUG ceilometer.compute.pollsters [-] > afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.read.requests volume: > 6138 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/_ > _init__.py:108 > 2022-03-04 16:13:08.038 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.038 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.pool, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.162 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.162 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.vip, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.163 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.outgoing.bytes, no resources found this > cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py: > 177 > 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] > Skipping tenant, keystone issue: admin endpoint for identity service in > RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: > admin endpoint for identity service in RegionOne region not found > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster radosgw.containers.objects.size, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster port.receive.packets, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster hardware.network.ip.outgoing.datagrams, no resources found this > cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py: > 177 > 2022-03-04 16:13:08.355 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.355 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.listener, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > > And here I report the content of the ceilometer.conf file: > [DEFAULT] > > debug = True > transport_url = rabbit://stackrabbit:admin at 10.30.2.194:5672/ > > [oslo_messaging_notifications] > topics = notifications > > [coordination] > backend_url = redis://localhost:6379 > > [notification] > workers = 2 > > [cache] > backend_argument = url:redis://localhost:6379 > backend_argument = distributed_lock:True > backend_argument = db:0 > backend_argument = redis_expiration_time:600 > backend = dogpile.cache.redis > enabled = True > > [service_credentials] > auth_url = http://10.30.2.194/identity > region_name = RegionOne > password = admin > username = ceilometer > project_name = service > project_domain_name = default > user_domain_id = default > auth_type = password > > [keystone_authtoken] > memcached_servers = localhost:11211 > cafile = /opt/stack/data/ca-bundle.pem > project_domain_name = default > project_name = admin > user_domain_name = default > password = admin > username = ceilometer > auth_url = http://10.30.2.194/identity > interface = public > auth_type = password > > I believe nothing should be changed in the configuration files since > Devstack should configure every components automatically. I also checked > pipeline.yaml and polling.yaml files and they seem correct. > > Do you have any idea on how to make ceilometer push metrics to the gnocchi > database?? > > Thanks. > > > ------------------------------ > *Da:* Matthias Runge > *Inviato:* gioved? 3 marzo 2022 13:58 > *A:* Molka Gharbaoui > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > On 3/3/22 12:31, Molka Gharbaoui wrote: > > Hi Matthias, > > > > Sorry for the confusion. Gnocchi is running and I am able to execute > > commands like "gnocchi metric list" and "gnocchi resource list" without > > errors even though the response is empty. Also checking mysql, the > > gnocchi database is correctly created. > > > > By the link I previously shared I wanted just to highlight that there > > are some differences between what was written there and between the > > environemnt I installed. I was searching there because I would have > > expected that Devstack installs and configures every component in > > automatic with no or minimum external intervention (in fact if I don't > > consider the Telemetry aspect, the rest works perfectly fine). > > > > I will follow Rafael guidelines to better undrstand where is the issue. > > > > Thanks. > > > Something I have used in the past is a local.conf like this: > > [[local|localrc]] > ADMIN_PASSWORD=secret > DATABASE_PASSWORD=$ADMIN_PASSWORD > RABBIT_PASSWORD=$ADMIN_PASSWORD > SERVICE_PASSWORD=$ADMIN_PASSWORD > > IMAGE_URL_SITE="https://download.fedoraproject.org" > IMAGE_URL_PATH="/pub/fedora/linux/releases/33/Cloud/x86_64/images/" > IMAGE_URL_FILE="Fedora-Cloud-Base-33-1.2.x86_64.qcow2" > IMAGE_URLS+=","$IMAGE_URL_SITE$IMAGE_URL_PATH$IMAGE_URL_FILE > > CEILOMETER_BACKEND=gnocchi > USE_PYTHON3="True" > > VERBOSE="True" > > VERBOSE_NO_TIMESTAMP="True" > > # > > LIBS_FROM_GIT=aodh,ceilometer,cinder,devstack,devstack-gate,dib-utils,diskimage-builder,glance,heat,keystone,neutron,nova,placement,requirements,swift > > enable_plugin aodh https://opendev.org/openstack/aodh > enable_plugin ceilometer https://opendev.org/openstack/ceilometer > enable_plugin heat https://opendev.org/openstack/heat > > > INSTALL_TEMPEST=False > > > Hope that helps. The IMAGE_URL_FILE needs probably adjustment, fedora 33 > is not supported anymore. > > Matthias > > > > > > > > > ------------------------------------------------------------------------ > > *Da:* Matthias Runge > > *Inviato:* gioved? 3 marzo 2022 11:59 > > *A:* openstack-discuss at lists.openstack.org > > > > *Oggetto:* Re: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > > metrics using ceilometer and gnocchi > > On 3/3/22 10:29, Molka Gharbaoui wrote: > >> Thank you Matthias for your reply. > >> > >> Unfortunately, from the logs I still cannot understand what is the > issue. > >> > >> Could someone kindly indicate me how ceilometer and gnocchi should be > >> configured to get the data? > >> > >> For example in this link: > >> > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html> > > > > >> < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html>> the > > > > >> endpoints should use the following URL: http://controller:8041 < > http://controller:8041> > >> > while in my case the > > 8041 port is not even > >> listening and the API indicated in the dashboard for Gnocchi is http:// > >> >x.x.x.x/metric. > >> > >> Thank you in advance. > >> > > > > OpenStack configuration is complex. The guide does not remotely touch on > > production environments. Please do not confuse devstack with production > > envionments. Also the doc you linked is not for devstack itself. > > > > The flow in general is: data is generated by the services like nova or > > glance. Ceilometer listens on the bus and also polls for information. > > The info is sent then over (as configured in publishers.yaml) to gnocchi. > > Gnocchi has an api, ceilometer has not. Reading your response, this > > makes me suspicious if there is gnocchi running. From the initial > > question it suggests it is, from your most recent response, it is not. > > > > > > Matthias > > > > -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From pierre at stackhpc.com Fri Mar 4 16:36:33 2022 From: pierre at stackhpc.com (Pierre Riteau) Date: Fri, 4 Mar 2022 17:36:33 +0100 Subject: [blazar][release][requirements] blazar-nova and nova requirements Message-ID: Hello release team, blazar-nova is a package including a nova scheduler filter, which must be installed together with nova to be used. The blazar-nova deliverable for Yoga was tagged with 2.4.0 at the time of the final release for non-client libraries. A few days ago, stable/yoga patches were automatically proposed. This is when I discovered that the lower-constraints job was broken because nova bumped some requirements / lower-constraints after we released [1] [2]. Because blazar-nova and nova are installed together, we keep our requirements in sync for testing purposes. Now, I believe it is just the lower-constraints testing that is broken, since nova requirements are still compatible with what blazar-nova requires: nova just requires more recent versions of some packages. I can run "pip install blazar-nova git+ https://opendev.org/openstack/nova.git at master#egg=nova" fine. I fixed the issue on the blazar-nova master branch [3] and proposed a backport to stable/yoga [4], but realised this was probably breaking release policy. Should we leave stable/yoga as released and just drop the lower-constraints job? As an aside, would it be possible to requalify blazar-nova as a cycle-with-rc deliverable for the next release? It is typed as a library but nothing depends on it: if anything, it depends on nova itself. Thanks, Pierre Riteau (priteau) [1] https://review.opendev.org/c/openstack/nova/+/712139 [2] https://review.opendev.org/c/openstack/nova/+/821606 [3] https://review.opendev.org/c/openstack/blazar-nova/+/831741 [4] https://review.opendev.org/c/openstack/blazar-nova/+/831832 -------------- next part -------------- An HTML attachment was scrubbed... URL: From mthode at mthode.org Fri Mar 4 16:37:29 2022 From: mthode at mthode.org (Matthew Thode) Date: Fri, 4 Mar 2022 10:37:29 -0600 Subject: [requirements][release][rfe] sushy 4.1.1 for Yoga In-Reply-To: References: <20220304152607.tkkvd7qhyl6o5ppu@domaintools.com> Message-ID: <20220304163729.piv3pggsfc5ppvb7@mthode.org> On 22-03-04 13:11:51, Iury Gregory wrote: > Hi Matthew > > You can find the diff in [1], thanks! > > [1] > https://zuul.opendev.org/t/openstack/build/4894a7e4d02e4892adab5ca4cefc8e89/log/tox/list-changes-results.log#280-408 > > > Em sex., 4 de mar. de 2022 ?s 12:26, Matthew Thode > escreveu: > > > On 22-03-04 12:05:40, Iury Gregory wrote: > > > Hello Stackers, > > > > > > I would like to request a new release for sushy before creating the yoga > > > branch, It contains some important fixes (e.g it will fix a problem when > > > build sushy in Debian). The sushy lib is only used by ironic, so it > > > shouldn't cause any problems updating the version in requirements. > > > > > > Thanks! > > > > > > -- > > > > > > > > > *Att[]'sIury Gregory Melo Ferreira * > > > *MSc in Computer Science at UFCG* > > > *Part of the ironic-core and puppet-manager-core team in OpenStack* > > > *Software Engineer at Red Hat Czech* > > > *Social*: https://www.linkedin.com/in/iurygregory > > > *E-mail: iurygregory at gmail.com * > > > > Hi, can you link to the diff? > > > > -- > > Matthew Thode > > > > > -- > > > *Att[]'sIury Gregory Melo Ferreira * > *MSc in Computer Science at UFCG* > *Part of the ironic-core and puppet-manager-core team in OpenStack* > *Software Engineer at Red Hat Czech* > *Social*: https://www.linkedin.com/in/iurygregory > *E-mail: iurygregory at gmail.com * Looks good, thanks. -- Matthew Thode From elod.illes at est.tech Fri Mar 4 16:41:16 2022 From: elod.illes at est.tech (=?UTF-8?B?RWzFkWQgSWxsw6lz?=) Date: Fri, 4 Mar 2022 17:41:16 +0100 Subject: [release] Release countdown for week R-3, Mar 07 - 11 Message-ID: Development Focus ----------------- The Release Candidate (RC) deadline is next Thursday, March 10th, 2022. Work should be focused on fixing any release-critical bugs. General Information ------------------- All deliverables released under a cycle-with-rc model should have a first release candidate by the end of the week, from which a stable/yoga branch will be cut. This branch will track the Yoga release. Once stable/yoga has been created, the master branch will be ready to switch to Zed development. While the master branch will no longer be feature-frozen, please prioritize any work necessary for completing Yoga plans.? Release-critical bugfixes will need to be merged in the master branch first, then backported to the stable/yoga branch before a new release candidate can be proposed. Actions ------- Early in the week, the release team will be proposing RC1 patches for all cycle-with-rc projects, using the latest commit from the master branch. If your team is ready to go for cutting RC1, please let us know by leaving a +1 on these patches. If there are still a few more patches needed before RC1, you can -1 the patch and update it later in the week with the new commit hash you would like to use. Remember, stable/yoga branches will be created with this, so you will want to make sure you have what you need included to avoid needing to backport changes from the master branch (which will technically then be Zed) to this stable branch for any additional RCs before the final release. The release team will also be proposing releases for any deliverable following a cycle-with-intermediary model that has not produced any Yoga release so far. Finally, now is a good time to finalize release highlights if you have not already done so. Release highlights help shape the messaging around the release and make sure that your work is properly represented. Upcoming Deadlines & Dates -------------------------- RC1 deadline: March 10th, 2022 (R-3 week) Final RC deadline: March 24th, 2022 (R-1 week) Final Yoga release: March 30th, 2022 Next PTG: April 4 - 8, 2022 (virtual) El?d Ill?s irc: elodilles -------------- next part -------------- An HTML attachment was scrubbed... URL: From Molka.Gharbaoui at santannapisa.it Fri Mar 4 16:49:34 2022 From: Molka.Gharbaoui at santannapisa.it (Molka Gharbaoui) Date: Fri, 4 Mar 2022 16:49:34 +0000 Subject: R: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: I have a running instance and I would like to retrieve some statistics on it like cpu usage, memory usage, packets received and sent. ________________________________ Da: Rafael Weing?rtner Inviato: venerd? 4 marzo 2022 17:28 A: Molka Gharbaoui Cc: Openstack Discuss List Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi What do you or what would you expect to see in Gnocchi being pushed by Ceilometer? What are the resources (volumes, VMs, and so) that you created in the OpenStack environment in DevStack? On Fri, Mar 4, 2022 at 1:26 PM Molka Gharbaoui > wrote: Unfortunately, the local.conf file shared by Matthias gives the same result: no metrics are pushed to gnocchi. Gnocchi database is empty: I can create metrics and resources manually but nothing is pushed from ceilometer to the database. So I guess the problem is at the ceilometer level. Here is an extract of the ceilometer-polling logs: 2022-03-04 16:13:07.904 679128 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of all_pollsters 2022-03-04 16:13:07.904 679128 DEBUG ceilometer.compute.pollsters [-] afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.usage volume: 5368709120 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/__init__.py:108 2022-03-04 16:13:07.907 679128 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of all_pollsters 2022-03-04 16:13:07.908 679128 DEBUG ceilometer.compute.pollsters [-] afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.read.requests volume: 6138 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/__init__.py:108 2022-03-04 16:13:08.038 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.038 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.pool, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.162 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.162 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.vip, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.163 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.outgoing.bytes, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] Skipping tenant, keystone issue: admin endpoint for identity service in RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: admin endpoint for identity service in RegionOne region not found 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster radosgw.containers.objects.size, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster port.receive.packets, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster hardware.network.ip.outgoing.datagrams, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.355 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.355 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.listener, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 And here I report the content of the ceilometer.conf file: [DEFAULT] debug = True transport_url = rabbit://stackrabbit:admin at 10.30.2.194:5672/ [oslo_messaging_notifications] topics = notifications [coordination] backend_url = redis://localhost:6379 [notification] workers = 2 [cache] backend_argument = url:redis://localhost:6379 backend_argument = distributed_lock:True backend_argument = db:0 backend_argument = redis_expiration_time:600 backend = dogpile.cache.redis enabled = True [service_credentials] auth_url = http://10.30.2.194/identity region_name = RegionOne password = admin username = ceilometer project_name = service project_domain_name = default user_domain_id = default auth_type = password [keystone_authtoken] memcached_servers = localhost:11211 cafile = /opt/stack/data/ca-bundle.pem project_domain_name = default project_name = admin user_domain_name = default password = admin username = ceilometer auth_url = http://10.30.2.194/identity interface = public auth_type = password I believe nothing should be changed in the configuration files since Devstack should configure every components automatically. I also checked pipeline.yaml and polling.yaml files and they seem correct. Do you have any idea on how to make ceilometer push metrics to the gnocchi database?? Thanks. ________________________________ Da: Matthias Runge > Inviato: gioved? 3 marzo 2022 13:58 A: Molka Gharbaoui > Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi On 3/3/22 12:31, Molka Gharbaoui wrote: > Hi Matthias, > > Sorry for the confusion. Gnocchi is running and I am able to execute > commands like "gnocchi metric list" and "gnocchi resource list" without > errors even though the response is empty. Also checking mysql, the > gnocchi database is correctly created. > > By the link I previously shared I wanted just to highlight that there > are some differences between what was written there and between the > environemnt I installed. I was searching there because I would have > expected that Devstack installs and configures every component in > automatic with no or minimum external intervention (in fact if I don't > consider the Telemetry aspect, the rest works perfectly fine). > > I will follow Rafael guidelines to better undrstand where is the issue. > > Thanks. Something I have used in the past is a local.conf like this: [[local|localrc]] ADMIN_PASSWORD=secret DATABASE_PASSWORD=$ADMIN_PASSWORD RABBIT_PASSWORD=$ADMIN_PASSWORD SERVICE_PASSWORD=$ADMIN_PASSWORD IMAGE_URL_SITE="https://download.fedoraproject.org" IMAGE_URL_PATH="/pub/fedora/linux/releases/33/Cloud/x86_64/images/" IMAGE_URL_FILE="Fedora-Cloud-Base-33-1.2.x86_64.qcow2" IMAGE_URLS+=","$IMAGE_URL_SITE$IMAGE_URL_PATH$IMAGE_URL_FILE CEILOMETER_BACKEND=gnocchi USE_PYTHON3="True" VERBOSE="True" VERBOSE_NO_TIMESTAMP="True" # LIBS_FROM_GIT=aodh,ceilometer,cinder,devstack,devstack-gate,dib-utils,diskimage-builder,glance,heat,keystone,neutron,nova,placement,requirements,swift enable_plugin aodh https://opendev.org/openstack/aodh enable_plugin ceilometer https://opendev.org/openstack/ceilometer enable_plugin heat https://opendev.org/openstack/heat INSTALL_TEMPEST=False Hope that helps. The IMAGE_URL_FILE needs probably adjustment, fedora 33 is not supported anymore. Matthias > > > ------------------------------------------------------------------------ > *Da:* Matthias Runge > > *Inviato:* gioved? 3 marzo 2022 11:59 > *A:* openstack-discuss at lists.openstack.org > > > *Oggetto:* Re: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > On 3/3/22 10:29, Molka Gharbaoui wrote: >> Thank you Matthias for your reply. >> >> Unfortunately, from the logs I still cannot understand what is the issue. >> >> Could someone kindly indicate me how ceilometer and gnocchi should be >> configured to get the data? >> >> For example in this link: >> https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > >> > the > >> endpoints should use the following URL: http://controller:8041 >> > while in my case the > 8041 port is not even >> listening and the API indicated in the dashboard for Gnocchi is http:// >> >x.x.x.x/metric. >> >> Thank you in advance. >> > > OpenStack configuration is complex. The guide does not remotely touch on > production environments. Please do not confuse devstack with production > envionments. Also the doc you linked is not for devstack itself. > > The flow in general is: data is generated by the services like nova or > glance. Ceilometer listens on the bus and also polls for information. > The info is sent then over (as configured in publishers.yaml) to gnocchi. > Gnocchi has an api, ceilometer has not. Reading your response, this > makes me suspicious if there is gnocchi running. From the initial > question it suggests it is, from your most recent response, it is not. > > > Matthias > -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From rafaelweingartner at gmail.com Fri Mar 4 16:56:01 2022 From: rafaelweingartner at gmail.com (=?UTF-8?Q?Rafael_Weing=C3=A4rtner?=) Date: Fri, 4 Mar 2022 13:56:01 -0300 Subject: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: I see. Is the Ceilometer compute agent up and running? On Fri, Mar 4, 2022 at 1:54 PM Molka Gharbaoui < Molka.Gharbaoui at santannapisa.it> wrote: > I have a running instance and I would like to retrieve some statistics on > it like cpu usage, memory usage, packets received and sent. > ------------------------------ > *Da:* Rafael Weing?rtner > *Inviato:* venerd? 4 marzo 2022 17:28 > *A:* Molka Gharbaoui > *Cc:* Openstack Discuss List > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > What do you or what would you expect to see in Gnocchi being pushed by > Ceilometer? > What are the resources (volumes, VMs, and so) that you created in the > OpenStack environment in DevStack? > > > On Fri, Mar 4, 2022 at 1:26 PM Molka Gharbaoui < > Molka.Gharbaoui at santannapisa.it> wrote: > > Unfortunately, the local.conf file shared by Matthias gives the same > result: no metrics are pushed to gnocchi. > > Gnocchi database is empty: I can create metrics and resources manually but > nothing is pushed from ceilometer to the database. So I guess the problem > is at the ceilometer level. > > Here is an extract of the ceilometer-polling logs: > > 2022-03-04 16:13:07.904 679128 INFO ceilometer.polling.manager [-] Polling > pollster disk.device.usage in the context of all_pollsters > 2022-03-04 16:13:07.904 679128 DEBUG ceilometer.compute.pollsters [-] > afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.usage volume: 5368709120 > _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/_ > _init__.py:108 > 2022-03-04 16:13:07.907 679128 INFO ceilometer.polling.manager [-] Polling > pollster disk.device.read.requests in the context of all_pollsters > 2022-03-04 16:13:07.908 679128 DEBUG ceilometer.compute.pollsters [-] > afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.read.requests volume: > 6138 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/_ > _init__.py:108 > 2022-03-04 16:13:08.038 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.038 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.pool, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.162 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.162 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.vip, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.163 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.outgoing.bytes, no resources found this > cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py: > 177 > 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] > Skipping tenant, keystone issue: admin endpoint for identity service in > RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: > admin endpoint for identity service in RegionOne region not found > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster radosgw.containers.objects.size, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster port.receive.packets, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster hardware.network.ip.outgoing.datagrams, no resources found this > cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py: > 177 > 2022-03-04 16:13:08.355 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.355 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.listener, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > > And here I report the content of the ceilometer.conf file: > [DEFAULT] > > debug = True > transport_url = rabbit://stackrabbit:admin at 10.30.2.194:5672/ > > [oslo_messaging_notifications] > topics = notifications > > [coordination] > backend_url = redis://localhost:6379 > > [notification] > workers = 2 > > [cache] > backend_argument = url:redis://localhost:6379 > backend_argument = distributed_lock:True > backend_argument = db:0 > backend_argument = redis_expiration_time:600 > backend = dogpile.cache.redis > enabled = True > > [service_credentials] > auth_url = http://10.30.2.194/identity > region_name = RegionOne > password = admin > username = ceilometer > project_name = service > project_domain_name = default > user_domain_id = default > auth_type = password > > [keystone_authtoken] > memcached_servers = localhost:11211 > cafile = /opt/stack/data/ca-bundle.pem > project_domain_name = default > project_name = admin > user_domain_name = default > password = admin > username = ceilometer > auth_url = http://10.30.2.194/identity > interface = public > auth_type = password > > I believe nothing should be changed in the configuration files since > Devstack should configure every components automatically. I also checked > pipeline.yaml and polling.yaml files and they seem correct. > > Do you have any idea on how to make ceilometer push metrics to the gnocchi > database?? > > Thanks. > > > ------------------------------ > *Da:* Matthias Runge > *Inviato:* gioved? 3 marzo 2022 13:58 > *A:* Molka Gharbaoui > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > On 3/3/22 12:31, Molka Gharbaoui wrote: > > Hi Matthias, > > > > Sorry for the confusion. Gnocchi is running and I am able to execute > > commands like "gnocchi metric list" and "gnocchi resource list" without > > errors even though the response is empty. Also checking mysql, the > > gnocchi database is correctly created. > > > > By the link I previously shared I wanted just to highlight that there > > are some differences between what was written there and between the > > environemnt I installed. I was searching there because I would have > > expected that Devstack installs and configures every component in > > automatic with no or minimum external intervention (in fact if I don't > > consider the Telemetry aspect, the rest works perfectly fine). > > > > I will follow Rafael guidelines to better undrstand where is the issue. > > > > Thanks. > > > Something I have used in the past is a local.conf like this: > > [[local|localrc]] > ADMIN_PASSWORD=secret > DATABASE_PASSWORD=$ADMIN_PASSWORD > RABBIT_PASSWORD=$ADMIN_PASSWORD > SERVICE_PASSWORD=$ADMIN_PASSWORD > > IMAGE_URL_SITE="https://download.fedoraproject.org" > IMAGE_URL_PATH="/pub/fedora/linux/releases/33/Cloud/x86_64/images/" > IMAGE_URL_FILE="Fedora-Cloud-Base-33-1.2.x86_64.qcow2" > IMAGE_URLS+=","$IMAGE_URL_SITE$IMAGE_URL_PATH$IMAGE_URL_FILE > > CEILOMETER_BACKEND=gnocchi > USE_PYTHON3="True" > > VERBOSE="True" > > VERBOSE_NO_TIMESTAMP="True" > > # > > LIBS_FROM_GIT=aodh,ceilometer,cinder,devstack,devstack-gate,dib-utils,diskimage-builder,glance,heat,keystone,neutron,nova,placement,requirements,swift > > enable_plugin aodh https://opendev.org/openstack/aodh > enable_plugin ceilometer https://opendev.org/openstack/ceilometer > enable_plugin heat https://opendev.org/openstack/heat > > > INSTALL_TEMPEST=False > > > Hope that helps. The IMAGE_URL_FILE needs probably adjustment, fedora 33 > is not supported anymore. > > Matthias > > > > > > > > > ------------------------------------------------------------------------ > > *Da:* Matthias Runge > > *Inviato:* gioved? 3 marzo 2022 11:59 > > *A:* openstack-discuss at lists.openstack.org > > > > *Oggetto:* Re: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > > metrics using ceilometer and gnocchi > > On 3/3/22 10:29, Molka Gharbaoui wrote: > >> Thank you Matthias for your reply. > >> > >> Unfortunately, from the logs I still cannot understand what is the > issue. > >> > >> Could someone kindly indicate me how ceilometer and gnocchi should be > >> configured to get the data? > >> > >> For example in this link: > >> > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html> > > > > >> < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html>> the > > > > >> endpoints should use the following URL: http://controller:8041 < > http://controller:8041> > >> > while in my case the > > 8041 port is not even > >> listening and the API indicated in the dashboard for Gnocchi is http:// > >> >x.x.x.x/metric. > >> > >> Thank you in advance. > >> > > > > OpenStack configuration is complex. The guide does not remotely touch on > > production environments. Please do not confuse devstack with production > > envionments. Also the doc you linked is not for devstack itself. > > > > The flow in general is: data is generated by the services like nova or > > glance. Ceilometer listens on the bus and also polls for information. > > The info is sent then over (as configured in publishers.yaml) to gnocchi. > > Gnocchi has an api, ceilometer has not. Reading your response, this > > makes me suspicious if there is gnocchi running. From the initial > > question it suggests it is, from your most recent response, it is not. > > > > > > Matthias > > > > > > -- > Rafael Weing?rtner > -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From Molka.Gharbaoui at santannapisa.it Fri Mar 4 16:58:46 2022 From: Molka.Gharbaoui at santannapisa.it (Molka Gharbaoui) Date: Fri, 4 Mar 2022 16:58:46 +0000 Subject: R: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: Yes, all ceilometer services are up and running: ceilometer-agent, ceilometer-notification. Actually, all devstack services are up and running. ________________________________ Da: Rafael Weing?rtner Inviato: venerd? 4 marzo 2022 17:56 A: Molka Gharbaoui Cc: Openstack Discuss List Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi I see. Is the Ceilometer compute agent up and running? On Fri, Mar 4, 2022 at 1:54 PM Molka Gharbaoui > wrote: I have a running instance and I would like to retrieve some statistics on it like cpu usage, memory usage, packets received and sent. ________________________________ Da: Rafael Weing?rtner > Inviato: venerd? 4 marzo 2022 17:28 A: Molka Gharbaoui > Cc: Openstack Discuss List > Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi What do you or what would you expect to see in Gnocchi being pushed by Ceilometer? What are the resources (volumes, VMs, and so) that you created in the OpenStack environment in DevStack? On Fri, Mar 4, 2022 at 1:26 PM Molka Gharbaoui > wrote: Unfortunately, the local.conf file shared by Matthias gives the same result: no metrics are pushed to gnocchi. Gnocchi database is empty: I can create metrics and resources manually but nothing is pushed from ceilometer to the database. So I guess the problem is at the ceilometer level. Here is an extract of the ceilometer-polling logs: 2022-03-04 16:13:07.904 679128 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of all_pollsters 2022-03-04 16:13:07.904 679128 DEBUG ceilometer.compute.pollsters [-] afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.usage volume: 5368709120 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/__init__.py:108 2022-03-04 16:13:07.907 679128 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of all_pollsters 2022-03-04 16:13:07.908 679128 DEBUG ceilometer.compute.pollsters [-] afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.read.requests volume: 6138 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/__init__.py:108 2022-03-04 16:13:08.038 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.038 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.pool, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.162 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.162 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.vip, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.163 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.outgoing.bytes, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] Skipping tenant, keystone issue: admin endpoint for identity service in RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: admin endpoint for identity service in RegionOne region not found 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster radosgw.containers.objects.size, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster port.receive.packets, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster hardware.network.ip.outgoing.datagrams, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.355 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.355 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.listener, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 And here I report the content of the ceilometer.conf file: [DEFAULT] debug = True transport_url = rabbit://stackrabbit:admin at 10.30.2.194:5672/ [oslo_messaging_notifications] topics = notifications [coordination] backend_url = redis://localhost:6379 [notification] workers = 2 [cache] backend_argument = url:redis://localhost:6379 backend_argument = distributed_lock:True backend_argument = db:0 backend_argument = redis_expiration_time:600 backend = dogpile.cache.redis enabled = True [service_credentials] auth_url = http://10.30.2.194/identity region_name = RegionOne password = admin username = ceilometer project_name = service project_domain_name = default user_domain_id = default auth_type = password [keystone_authtoken] memcached_servers = localhost:11211 cafile = /opt/stack/data/ca-bundle.pem project_domain_name = default project_name = admin user_domain_name = default password = admin username = ceilometer auth_url = http://10.30.2.194/identity interface = public auth_type = password I believe nothing should be changed in the configuration files since Devstack should configure every components automatically. I also checked pipeline.yaml and polling.yaml files and they seem correct. Do you have any idea on how to make ceilometer push metrics to the gnocchi database?? Thanks. ________________________________ Da: Matthias Runge > Inviato: gioved? 3 marzo 2022 13:58 A: Molka Gharbaoui > Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi On 3/3/22 12:31, Molka Gharbaoui wrote: > Hi Matthias, > > Sorry for the confusion. Gnocchi is running and I am able to execute > commands like "gnocchi metric list" and "gnocchi resource list" without > errors even though the response is empty. Also checking mysql, the > gnocchi database is correctly created. > > By the link I previously shared I wanted just to highlight that there > are some differences between what was written there and between the > environemnt I installed. I was searching there because I would have > expected that Devstack installs and configures every component in > automatic with no or minimum external intervention (in fact if I don't > consider the Telemetry aspect, the rest works perfectly fine). > > I will follow Rafael guidelines to better undrstand where is the issue. > > Thanks. Something I have used in the past is a local.conf like this: [[local|localrc]] ADMIN_PASSWORD=secret DATABASE_PASSWORD=$ADMIN_PASSWORD RABBIT_PASSWORD=$ADMIN_PASSWORD SERVICE_PASSWORD=$ADMIN_PASSWORD IMAGE_URL_SITE="https://download.fedoraproject.org" IMAGE_URL_PATH="/pub/fedora/linux/releases/33/Cloud/x86_64/images/" IMAGE_URL_FILE="Fedora-Cloud-Base-33-1.2.x86_64.qcow2" IMAGE_URLS+=","$IMAGE_URL_SITE$IMAGE_URL_PATH$IMAGE_URL_FILE CEILOMETER_BACKEND=gnocchi USE_PYTHON3="True" VERBOSE="True" VERBOSE_NO_TIMESTAMP="True" # LIBS_FROM_GIT=aodh,ceilometer,cinder,devstack,devstack-gate,dib-utils,diskimage-builder,glance,heat,keystone,neutron,nova,placement,requirements,swift enable_plugin aodh https://opendev.org/openstack/aodh enable_plugin ceilometer https://opendev.org/openstack/ceilometer enable_plugin heat https://opendev.org/openstack/heat INSTALL_TEMPEST=False Hope that helps. The IMAGE_URL_FILE needs probably adjustment, fedora 33 is not supported anymore. Matthias > > > ------------------------------------------------------------------------ > *Da:* Matthias Runge > > *Inviato:* gioved? 3 marzo 2022 11:59 > *A:* openstack-discuss at lists.openstack.org > > > *Oggetto:* Re: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > On 3/3/22 10:29, Molka Gharbaoui wrote: >> Thank you Matthias for your reply. >> >> Unfortunately, from the logs I still cannot understand what is the issue. >> >> Could someone kindly indicate me how ceilometer and gnocchi should be >> configured to get the data? >> >> For example in this link: >> https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > >> > the > >> endpoints should use the following URL: http://controller:8041 >> > while in my case the > 8041 port is not even >> listening and the API indicated in the dashboard for Gnocchi is http:// >> >x.x.x.x/metric. >> >> Thank you in advance. >> > > OpenStack configuration is complex. The guide does not remotely touch on > production environments. Please do not confuse devstack with production > envionments. Also the doc you linked is not for devstack itself. > > The flow in general is: data is generated by the services like nova or > glance. Ceilometer listens on the bus and also polls for information. > The info is sent then over (as configured in publishers.yaml) to gnocchi. > Gnocchi has an api, ceilometer has not. Reading your response, this > makes me suspicious if there is gnocchi running. From the initial > question it suggests it is, from your most recent response, it is not. > > > Matthias > -- Rafael Weing?rtner -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From arnaud.morin at gmail.com Fri Mar 4 17:07:51 2022 From: arnaud.morin at gmail.com (Arnaud Morin) Date: Fri, 4 Mar 2022 17:07:51 +0000 Subject: [nova][neutron] os-vif params Message-ID: Hey all, When I take a look at [1], I can see some parameters, such as network_device_mtu, isolate_vif, etc. As far as I can tell, nova is using these params when plugging the port in OVS (if using OVS as ml2 plugin of course). But I cannot find the config params in the documentation: [2] Am I missing something? Cheers, Arnaud. [1] https://github.com/openstack/os-vif/blob/master/vif_plug_ovs/ovs.py [2] https://docs.openstack.org/nova/latest/configuration/config.html From franck.vedel at univ-grenoble-alpes.fr Fri Mar 4 17:08:44 2022 From: franck.vedel at univ-grenoble-alpes.fr (Franck VEDEL) Date: Fri, 4 Mar 2022 18:08:44 +0100 Subject: [kolla-ansible][wallaby][centos][horizon-dashboard] Changing logo.svg Message-ID: <60FB6A20-20EA-4CF7-A962-3CF8AA9C88AA@univ-grenoble-alpes.fr> Hello, I come back to ask for some help with a new problem on my Openstack platform. For once, it's a small problem. I take this opportunity to thank those who helped me solve my problem which was the following: the instances were not created regularly on the 3 nova servers. Nova-scheduler was malfunctioning. I fixed the problem by setting build_failure_weight_multiplier=0 Since then, in DEBUG mode, we can clearly see the calculation of the weight and the distribution of the instances. It's beautiful. My new problem is: I'm using kolla-ansible, centos (stream) and Wallaby. In the directory /usr/local/lib/python3.6/site-packages/openstack_dashboard/static/dashboard/img/ I modified the 3 files logo.svg, logo-splash.svg, favicon.ico. I thought doing a "kolla-ansible -i multinode reconfigure horizon" would update the logos and favicon.ico....but no, it does nothing. How do I change these logos with kolla-ansible? Is it difficult? thank you in advance Franck VEDEL -------------- next part -------------- An HTML attachment was scrubbed... URL: From ralonsoh at redhat.com Fri Mar 4 17:33:52 2022 From: ralonsoh at redhat.com (Rodolfo Alonso Hernandez) Date: Fri, 4 Mar 2022 18:33:52 +0100 Subject: [nova][neutron] os-vif params In-Reply-To: References: Message-ID: Hi Arnaud: Those parameters should be set in nova.conf, in the "os_vif_XXX" section [1]. "network_device_mtu" is deprecated. The interface MTU is retrieved from the network MTU. "ovs_vsctl_timeout" is deprecated too, we don't use vsctl commands anymore. Maybe Nova folks can extend my answer. Regards. [1] https://github.com/openstack/nova/blob/50fdbc752a9ca9c31488140ef2997ed59d861a41/releasenotes/notes/libvirt-uses-os-vif-plugins-31a0617de0c248b9.yaml On Fri, Mar 4, 2022 at 6:18 PM Arnaud Morin wrote: > Hey all, > > When I take a look at [1], I can see some parameters, such as > network_device_mtu, isolate_vif, etc. > > As far as I can tell, nova is using these params when plugging the port > in OVS (if using OVS as ml2 plugin of course). > > But I cannot find the config params in the documentation: [2] > > Am I missing something? > > Cheers, > Arnaud. > > [1] https://github.com/openstack/os-vif/blob/master/vif_plug_ovs/ovs.py > [2] https://docs.openstack.org/nova/latest/configuration/config.html > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From smooney at redhat.com Fri Mar 4 17:37:05 2022 From: smooney at redhat.com (Sean Mooney) Date: Fri, 04 Mar 2022 17:37:05 +0000 Subject: [nova][neutron] os-vif params In-Reply-To: References: Message-ID: On Fri, 2022-03-04 at 17:07 +0000, Arnaud Morin wrote: > Hey all, > > When I take a look at [1], I can see some parameters, such as > network_device_mtu, isolate_vif, etc. > > As far as I can tell, nova is using these params when plugging the port > in OVS (if using OVS as ml2 plugin of course). > > But I cannot find the config params in the documentation: [2] > > Am I missing something? no we dont currently generate that documenation currntly https://docs.openstack.org/nova/latest/configuration/config.html should list it but we have not added the appropriate namespaces to the list of namespaces for nova to generate. to use these option today you create a [os_vif_ovs] section in your nova.conf we do that in devstack today https://github.com/openstack/devstack/blob/master/lib/os-vif so if you look at the nova-cpu.conf form one of our jobs you can see it set to false in this case since we are using ovn https://zuul.openstack.org/build/7ea68106815e492a8de73c28afffaa0a/log/controller/logs/etc/nova/nova-cpu_conf.txt#134-137 network_device_mtu is not used anymore that was for nova network and should be delete in the future. ovsdb_connection = tcp:[::1]:6640 isolate_vif = False ovsdb_interface = vsctl are all valid to configure still. i think we just need to add os-vif here https://github.com/openstack/nova/blob/a74bced0eb3232c0c3384e26db1838bbc26c2d82/doc/source/conf.py#L179-L197 although i coudl be wrong. i know there is something we need to tweek to include options form opther project likel the messaging option in the config ref the os-vif docs are located here but we do not generate a sperate config ref https://docs.openstack.org/os-vif/xena/index.html > > Cheers, > Arnaud. > > [1] https://github.com/openstack/os-vif/blob/master/vif_plug_ovs/ovs.py > [2] https://docs.openstack.org/nova/latest/configuration/config.html > From smooney at redhat.com Fri Mar 4 17:40:48 2022 From: smooney at redhat.com (Sean Mooney) Date: Fri, 04 Mar 2022 17:40:48 +0000 Subject: [nova][neutron] os-vif params In-Reply-To: References: Message-ID: <20490cbeda70bdd8515656934d4fa157589b1278.camel@redhat.com> On Fri, 2022-03-04 at 17:37 +0000, Sean Mooney wrote: > On Fri, 2022-03-04 at 17:07 +0000, Arnaud Morin wrote: > > Hey all, > > > > When I take a look at [1], I can see some parameters, such as > > network_device_mtu, isolate_vif, etc. > > > > As far as I can tell, nova is using these params when plugging the port > > in OVS (if using OVS as ml2 plugin of course). > > > > But I cannot find the config params in the documentation: [2] > > > > Am I missing something? > > no we dont currently generate that documenation currntly > https://docs.openstack.org/nova/latest/configuration/config.html should list it but we have not added > the appropriate namespaces to the list of namespaces for nova to generate. > > to use these option today you create a [os_vif_ovs] section in your nova.conf > we do that in devstack today > https://github.com/openstack/devstack/blob/master/lib/os-vif > > so if you look at the nova-cpu.conf form one of our jobs you can see it set to false in this case since we are using ovn > https://zuul.openstack.org/build/7ea68106815e492a8de73c28afffaa0a/log/controller/logs/etc/nova/nova-cpu_conf.txt#134-137 > > network_device_mtu is not used anymore that was for nova network and should be delete in the future. > > ovsdb_connection = tcp:[::1]:6640 > isolate_vif = False > ovsdb_interface = vsctl > > are all valid to configure still. > > i think we just need to add os-vif here > https://github.com/openstack/nova/blob/a74bced0eb3232c0c3384e26db1838bbc26c2d82/doc/source/conf.py#L179-L197 > although i coudl be wrong. actully i think we need to list the namespce here https://github.com/openstack/nova/blob/a74bced0eb3232c0c3384e26db1838bbc26c2d82/etc/nova/nova-config-generator.conf#L17 in the sample nova-config generator config i dont think its hard to fix that you just need to find the right place and added it but no one has done the work to figure it out and push a patch. > i know there is something we need to tweek to include options form opther project likel the messaging option in the config ref > the os-vif docs are located here but we do not generate a sperate config ref > https://docs.openstack.org/os-vif/xena/index.html > > > > Cheers, > > Arnaud. > > > > [1] https://github.com/openstack/os-vif/blob/master/vif_plug_ovs/ovs.py > > [2] https://docs.openstack.org/nova/latest/configuration/config.html > > > From mthode at domaintools.com Fri Mar 4 15:26:07 2022 From: mthode at domaintools.com (Matthew Thode) Date: Fri, 4 Mar 2022 09:26:07 -0600 Subject: [requirements][release][rfe] sushy 4.1.1 for Yoga In-Reply-To: References: Message-ID: <20220304152607.tkkvd7qhyl6o5ppu@domaintools.com> On 22-03-04 12:05:40, Iury Gregory wrote: > Hello Stackers, > > I would like to request a new release for sushy before creating the yoga > branch, It contains some important fixes (e.g it will fix a problem when > build sushy in Debian). The sushy lib is only used by ironic, so it > shouldn't cause any problems updating the version in requirements. > > Thanks! > > -- > > > *Att[]'sIury Gregory Melo Ferreira * > *MSc in Computer Science at UFCG* > *Part of the ironic-core and puppet-manager-core team in OpenStack* > *Software Engineer at Red Hat Czech* > *Social*: https://www.linkedin.com/in/iurygregory > *E-mail: iurygregory at gmail.com * Hi, can you link to the diff? -- Matthew Thode -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 833 bytes Desc: not available URL: From smooney at redhat.com Fri Mar 4 17:57:40 2022 From: smooney at redhat.com (Sean Mooney) Date: Fri, 04 Mar 2022 17:57:40 +0000 Subject: [nova][neutron] os-vif params In-Reply-To: References: Message-ID: <3457d3040994053d4ecec879ef33c9b09758774b.camel@redhat.com> On Fri, 2022-03-04 at 18:33 +0100, Rodolfo Alonso Hernandez wrote: > Hi Arnaud: > > Those parameters should be set in nova.conf, in the "os_vif_XXX" section > [1]. > > "network_device_mtu" is deprecated. The interface MTU is retrieved from the > network MTU. "ovs_vsctl_timeout" is deprecated too, we don't use vsctl > commands anymore. actully we do use vsctl at least in ci but it is deprecated. we are currently using it in ci to workaround a bug in the ovs python bindings. after the RC period is clear i might revert that since i think the ovs python bidnign are fixed but im not sure they are fixed in ubuntu 20.04 so i dont want to regress the ci stablity currently by reverting to the native bindings. https://review.opendev.org/c/openstack/devstack/+/796826 ill update https://github.com/openstack/devstack/blob/master/lib/os-vif#L6 to native once we are sure https://bugs.launchpad.net/os-vif/+bug/1929446 is fixed in the ovs python bindings. > > Maybe Nova folks can extend my answer. > > Regards. > > [1] > https://github.com/openstack/nova/blob/50fdbc752a9ca9c31488140ef2997ed59d861a41/releasenotes/notes/libvirt-uses-os-vif-plugins-31a0617de0c248b9.yaml > > > On Fri, Mar 4, 2022 at 6:18 PM Arnaud Morin wrote: > > > Hey all, > > > > When I take a look at [1], I can see some parameters, such as > > network_device_mtu, isolate_vif, etc. > > > > As far as I can tell, nova is using these params when plugging the port > > in OVS (if using OVS as ml2 plugin of course). > > > > But I cannot find the config params in the documentation: [2] > > > > Am I missing something? > > > > Cheers, > > Arnaud. > > > > [1] https://github.com/openstack/os-vif/blob/master/vif_plug_ovs/ovs.py > > [2] https://docs.openstack.org/nova/latest/configuration/config.html > > > > From gmann at ghanshyammann.com Fri Mar 4 18:32:50 2022 From: gmann at ghanshyammann.com (Ghanshyam Mann) Date: Fri, 04 Mar 2022 12:32:50 -0600 Subject: [tc][ptl] TC + Community Leaders interaction in Zed Virtual PTG In-Reply-To: <17f22bcdbf2.b05d02e2454544.7809265565474276350@ghanshyammann.com> References: <17f22bcdbf2.b05d02e2454544.7809265565474276350@ghanshyammann.com> Message-ID: <17f5634b96d.f2d285d0223727.6162405599807850471@ghanshyammann.com> ---- On Tue, 22 Feb 2022 12:41:16 -0600 Ghanshyam Mann wrote ---- > Hello Everyone, > > We had successful TC+Community leaders interaction sessions in the last PTG, and we will > continue the same for Zed cycle PTG also. > > I have created the below etherpad with the draft agenda, If you as PTL or project representatives > or SIG Chair and would like to attend this session, please add your name: > > - https://etherpad.opendev.org/p/tc-ptl-interaction-zed > > Also, I have created the doodle poll to select the best suitable time. It will be for two hours > (single or two sessions) either on Monday or Tuesday. Please add your preference in the below > doodle poll (including TC members): > > - https://doodle.com/poll/dcer6i3wk2b8eim3 Thanks Everyone for adding your availability in poll. I have booked the Monday 14-16 UTC slot for this session. - https://ethercalc.openstack.org/7yxdas7suqnd -gmann > > -gmann > > > From rafaelweingartner at gmail.com Fri Mar 4 18:48:47 2022 From: rafaelweingartner at gmail.com (=?UTF-8?Q?Rafael_Weing=C3=A4rtner?=) Date: Fri, 4 Mar 2022 15:48:47 -0300 Subject: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: And, what about ceilometer-compute? On Fri, Mar 4, 2022 at 2:04 PM Molka Gharbaoui < Molka.Gharbaoui at santannapisa.it> wrote: > Yes, all ceilometer services are up and running: ceilometer-agent, > ceilometer-notification. Actually, all devstack services are up and running. > ------------------------------ > *Da:* Rafael Weing?rtner > *Inviato:* venerd? 4 marzo 2022 17:56 > *A:* Molka Gharbaoui > *Cc:* Openstack Discuss List > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > I see. Is the Ceilometer compute agent up and running? > > On Fri, Mar 4, 2022 at 1:54 PM Molka Gharbaoui < > Molka.Gharbaoui at santannapisa.it> wrote: > > I have a running instance and I would like to retrieve some statistics on > it like cpu usage, memory usage, packets received and sent. > ------------------------------ > *Da:* Rafael Weing?rtner > *Inviato:* venerd? 4 marzo 2022 17:28 > *A:* Molka Gharbaoui > *Cc:* Openstack Discuss List > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > What do you or what would you expect to see in Gnocchi being pushed by > Ceilometer? > What are the resources (volumes, VMs, and so) that you created in the > OpenStack environment in DevStack? > > > On Fri, Mar 4, 2022 at 1:26 PM Molka Gharbaoui < > Molka.Gharbaoui at santannapisa.it> wrote: > > Unfortunately, the local.conf file shared by Matthias gives the same > result: no metrics are pushed to gnocchi. > > Gnocchi database is empty: I can create metrics and resources manually but > nothing is pushed from ceilometer to the database. So I guess the problem > is at the ceilometer level. > > Here is an extract of the ceilometer-polling logs: > > 2022-03-04 16:13:07.904 679128 INFO ceilometer.polling.manager [-] Polling > pollster disk.device.usage in the context of all_pollsters > 2022-03-04 16:13:07.904 679128 DEBUG ceilometer.compute.pollsters [-] > afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.usage volume: 5368709120 > _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/_ > _init__.py:108 > 2022-03-04 16:13:07.907 679128 INFO ceilometer.polling.manager [-] Polling > pollster disk.device.read.requests in the context of all_pollsters > 2022-03-04 16:13:07.908 679128 DEBUG ceilometer.compute.pollsters [-] > afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.read.requests volume: > 6138 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/_ > _init__.py:108 > 2022-03-04 16:13:08.038 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.038 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.pool, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.162 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.162 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.vip, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.163 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.outgoing.bytes, no resources found this > cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py: > 177 > 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] > Skipping tenant, keystone issue: admin endpoint for identity service in > RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: > admin endpoint for identity service in RegionOne region not found > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster radosgw.containers.objects.size, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster port.receive.packets, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster hardware.network.ip.outgoing.datagrams, no resources found this > cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py: > 177 > 2022-03-04 16:13:08.355 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.355 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.listener, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > > And here I report the content of the ceilometer.conf file: > [DEFAULT] > > debug = True > transport_url = rabbit://stackrabbit:admin at 10.30.2.194:5672/ > > [oslo_messaging_notifications] > topics = notifications > > [coordination] > backend_url = redis://localhost:6379 > > [notification] > workers = 2 > > [cache] > backend_argument = url:redis://localhost:6379 > backend_argument = distributed_lock:True > backend_argument = db:0 > backend_argument = redis_expiration_time:600 > backend = dogpile.cache.redis > enabled = True > > [service_credentials] > auth_url = http://10.30.2.194/identity > region_name = RegionOne > password = admin > username = ceilometer > project_name = service > project_domain_name = default > user_domain_id = default > auth_type = password > > [keystone_authtoken] > memcached_servers = localhost:11211 > cafile = /opt/stack/data/ca-bundle.pem > project_domain_name = default > project_name = admin > user_domain_name = default > password = admin > username = ceilometer > auth_url = http://10.30.2.194/identity > interface = public > auth_type = password > > I believe nothing should be changed in the configuration files since > Devstack should configure every components automatically. I also checked > pipeline.yaml and polling.yaml files and they seem correct. > > Do you have any idea on how to make ceilometer push metrics to the gnocchi > database?? > > Thanks. > > > ------------------------------ > *Da:* Matthias Runge > *Inviato:* gioved? 3 marzo 2022 13:58 > *A:* Molka Gharbaoui > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > On 3/3/22 12:31, Molka Gharbaoui wrote: > > Hi Matthias, > > > > Sorry for the confusion. Gnocchi is running and I am able to execute > > commands like "gnocchi metric list" and "gnocchi resource list" without > > errors even though the response is empty. Also checking mysql, the > > gnocchi database is correctly created. > > > > By the link I previously shared I wanted just to highlight that there > > are some differences between what was written there and between the > > environemnt I installed. I was searching there because I would have > > expected that Devstack installs and configures every component in > > automatic with no or minimum external intervention (in fact if I don't > > consider the Telemetry aspect, the rest works perfectly fine). > > > > I will follow Rafael guidelines to better undrstand where is the issue. > > > > Thanks. > > > Something I have used in the past is a local.conf like this: > > [[local|localrc]] > ADMIN_PASSWORD=secret > DATABASE_PASSWORD=$ADMIN_PASSWORD > RABBIT_PASSWORD=$ADMIN_PASSWORD > SERVICE_PASSWORD=$ADMIN_PASSWORD > > IMAGE_URL_SITE="https://download.fedoraproject.org" > IMAGE_URL_PATH="/pub/fedora/linux/releases/33/Cloud/x86_64/images/" > IMAGE_URL_FILE="Fedora-Cloud-Base-33-1.2.x86_64.qcow2" > IMAGE_URLS+=","$IMAGE_URL_SITE$IMAGE_URL_PATH$IMAGE_URL_FILE > > CEILOMETER_BACKEND=gnocchi > USE_PYTHON3="True" > > VERBOSE="True" > > VERBOSE_NO_TIMESTAMP="True" > > # > > LIBS_FROM_GIT=aodh,ceilometer,cinder,devstack,devstack-gate,dib-utils,diskimage-builder,glance,heat,keystone,neutron,nova,placement,requirements,swift > > enable_plugin aodh https://opendev.org/openstack/aodh > enable_plugin ceilometer https://opendev.org/openstack/ceilometer > enable_plugin heat https://opendev.org/openstack/heat > > > INSTALL_TEMPEST=False > > > Hope that helps. The IMAGE_URL_FILE needs probably adjustment, fedora 33 > is not supported anymore. > > Matthias > > > > > > > > > ------------------------------------------------------------------------ > > *Da:* Matthias Runge > > *Inviato:* gioved? 3 marzo 2022 11:59 > > *A:* openstack-discuss at lists.openstack.org > > > > *Oggetto:* Re: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > > metrics using ceilometer and gnocchi > > On 3/3/22 10:29, Molka Gharbaoui wrote: > >> Thank you Matthias for your reply. > >> > >> Unfortunately, from the logs I still cannot understand what is the > issue. > >> > >> Could someone kindly indicate me how ceilometer and gnocchi should be > >> configured to get the data? > >> > >> For example in this link: > >> > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html> > > > > >> < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html>> the > > > > >> endpoints should use the following URL: http://controller:8041 < > http://controller:8041> > >> > while in my case the > > 8041 port is not even > >> listening and the API indicated in the dashboard for Gnocchi is http:// > >> >x.x.x.x/metric. > >> > >> Thank you in advance. > >> > > > > OpenStack configuration is complex. The guide does not remotely touch on > > production environments. Please do not confuse devstack with production > > envionments. Also the doc you linked is not for devstack itself. > > > > The flow in general is: data is generated by the services like nova or > > glance. Ceilometer listens on the bus and also polls for information. > > The info is sent then over (as configured in publishers.yaml) to gnocchi. > > Gnocchi has an api, ceilometer has not. Reading your response, this > > makes me suspicious if there is gnocchi running. From the initial > > question it suggests it is, from your most recent response, it is not. > > > > > > Matthias > > > > > > -- > Rafael Weing?rtner > > > > -- > Rafael Weing?rtner > -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From arnaud.morin at gmail.com Fri Mar 4 18:57:02 2022 From: arnaud.morin at gmail.com (Arnaud) Date: Fri, 04 Mar 2022 19:57:02 +0100 Subject: [nova][neutron] os-vif params In-Reply-To: <3457d3040994053d4ecec879ef33c9b09758774b.camel@redhat.com> References: <3457d3040994053d4ecec879ef33c9b09758774b.camel@redhat.com> Message-ID: <45AC4BA4-DA72-42A1-8E7B-4DAD3870CF0B@gmail.com> Hi Sean, Rodolfo, Thanks for your answers. I will try to do the patch next week so these parameters are included in the nova config sample and documentation. About the isolated_vif params. It seems like a security enhancement to set it to True when using ovs? Do you confirm that? Regards, Arnaud Le 4 mars 2022 18:57:40 GMT+01:00, Sean Mooney a ?crit?: >On Fri, 2022-03-04 at 18:33 +0100, Rodolfo Alonso Hernandez wrote: >> Hi Arnaud: >> >> Those parameters should be set in nova.conf, in the "os_vif_XXX" section >> [1]. >> >> "network_device_mtu" is deprecated. The interface MTU is retrieved from the >> network MTU. "ovs_vsctl_timeout" is deprecated too, we don't use vsctl >> commands anymore. >actully we do use vsctl at least in ci but it is deprecated. >we are currently using it in ci to workaround a bug in the ovs python bindings. > >after the RC period is clear i might revert that since i think the ovs python bidnign are >fixed but im not sure they are fixed in ubuntu 20.04 so i dont want to regress the ci stablity >currently by reverting to the native bindings. > >https://review.opendev.org/c/openstack/devstack/+/796826 > >ill update https://github.com/openstack/devstack/blob/master/lib/os-vif#L6 to native >once we are sure https://bugs.launchpad.net/os-vif/+bug/1929446 is fixed in the ovs python bindings. > > >> >> Maybe Nova folks can extend my answer. >> >> Regards. >> >> [1] >> https://github.com/openstack/nova/blob/50fdbc752a9ca9c31488140ef2997ed59d861a41/releasenotes/notes/libvirt-uses-os-vif-plugins-31a0617de0c248b9.yaml >> >> >> On Fri, Mar 4, 2022 at 6:18 PM Arnaud Morin wrote: >> >> > Hey all, >> > >> > When I take a look at [1], I can see some parameters, such as >> > network_device_mtu, isolate_vif, etc. >> > >> > As far as I can tell, nova is using these params when plugging the port >> > in OVS (if using OVS as ml2 plugin of course). >> > >> > But I cannot find the config params in the documentation: [2] >> > >> > Am I missing something? >> > >> > Cheers, >> > Arnaud. >> > >> > [1] https://github.com/openstack/os-vif/blob/master/vif_plug_ovs/ovs.py >> > [2] https://docs.openstack.org/nova/latest/configuration/config.html >> > >> > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From Molka.Gharbaoui at santannapisa.it Fri Mar 4 18:57:30 2022 From: Molka.Gharbaoui at santannapisa.it (Molka Gharbaoui) Date: Fri, 4 Mar 2022 18:57:30 +0000 Subject: R: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: Yes sorry, I forgot to mention it but it is also running. ________________________________ Da: Rafael Weing?rtner Inviato: venerd? 4 marzo 2022 19:48 A: Molka Gharbaoui Cc: Openstack Discuss List Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi And, what about ceilometer-compute? On Fri, Mar 4, 2022 at 2:04 PM Molka Gharbaoui > wrote: Yes, all ceilometer services are up and running: ceilometer-agent, ceilometer-notification. Actually, all devstack services are up and running. ________________________________ Da: Rafael Weing?rtner > Inviato: venerd? 4 marzo 2022 17:56 A: Molka Gharbaoui > Cc: Openstack Discuss List > Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi I see. Is the Ceilometer compute agent up and running? On Fri, Mar 4, 2022 at 1:54 PM Molka Gharbaoui > wrote: I have a running instance and I would like to retrieve some statistics on it like cpu usage, memory usage, packets received and sent. ________________________________ Da: Rafael Weing?rtner > Inviato: venerd? 4 marzo 2022 17:28 A: Molka Gharbaoui > Cc: Openstack Discuss List > Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi What do you or what would you expect to see in Gnocchi being pushed by Ceilometer? What are the resources (volumes, VMs, and so) that you created in the OpenStack environment in DevStack? On Fri, Mar 4, 2022 at 1:26 PM Molka Gharbaoui > wrote: Unfortunately, the local.conf file shared by Matthias gives the same result: no metrics are pushed to gnocchi. Gnocchi database is empty: I can create metrics and resources manually but nothing is pushed from ceilometer to the database. So I guess the problem is at the ceilometer level. Here is an extract of the ceilometer-polling logs: 2022-03-04 16:13:07.904 679128 INFO ceilometer.polling.manager [-] Polling pollster disk.device.usage in the context of all_pollsters 2022-03-04 16:13:07.904 679128 DEBUG ceilometer.compute.pollsters [-] afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.usage volume: 5368709120 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/__init__.py:108 2022-03-04 16:13:07.907 679128 INFO ceilometer.polling.manager [-] Polling pollster disk.device.read.requests in the context of all_pollsters 2022-03-04 16:13:07.908 679128 DEBUG ceilometer.compute.pollsters [-] afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.read.requests volume: 6138 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/__init__.py:108 2022-03-04 16:13:08.038 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.038 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.pool, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.162 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.162 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.vip, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.163 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.outgoing.bytes, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] Skipping tenant, keystone issue: admin endpoint for identity service in RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: admin endpoint for identity service in RegionOne region not found 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster radosgw.containers.objects.size, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster port.receive.packets, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip pollster hardware.network.ip.outgoing.datagrams, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 2022-03-04 16:13:08.355 679128 WARNING ceilometer.neutron_client [-] The resource could not be found.: neutronclient.common.exceptions.NotFound: The resource could not be found. 2022-03-04 16:13:08.355 679128 DEBUG ceilometer.polling.manager [-] Skip pollster network.services.lb.listener, no resources found this cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 And here I report the content of the ceilometer.conf file: [DEFAULT] debug = True transport_url = rabbit://stackrabbit:admin at 10.30.2.194:5672/ [oslo_messaging_notifications] topics = notifications [coordination] backend_url = redis://localhost:6379 [notification] workers = 2 [cache] backend_argument = url:redis://localhost:6379 backend_argument = distributed_lock:True backend_argument = db:0 backend_argument = redis_expiration_time:600 backend = dogpile.cache.redis enabled = True [service_credentials] auth_url = http://10.30.2.194/identity region_name = RegionOne password = admin username = ceilometer project_name = service project_domain_name = default user_domain_id = default auth_type = password [keystone_authtoken] memcached_servers = localhost:11211 cafile = /opt/stack/data/ca-bundle.pem project_domain_name = default project_name = admin user_domain_name = default password = admin username = ceilometer auth_url = http://10.30.2.194/identity interface = public auth_type = password I believe nothing should be changed in the configuration files since Devstack should configure every components automatically. I also checked pipeline.yaml and polling.yaml files and they seem correct. Do you have any idea on how to make ceilometer push metrics to the gnocchi database?? Thanks. ________________________________ Da: Matthias Runge > Inviato: gioved? 3 marzo 2022 13:58 A: Molka Gharbaoui > Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi On 3/3/22 12:31, Molka Gharbaoui wrote: > Hi Matthias, > > Sorry for the confusion. Gnocchi is running and I am able to execute > commands like "gnocchi metric list" and "gnocchi resource list" without > errors even though the response is empty. Also checking mysql, the > gnocchi database is correctly created. > > By the link I previously shared I wanted just to highlight that there > are some differences between what was written there and between the > environemnt I installed. I was searching there because I would have > expected that Devstack installs and configures every component in > automatic with no or minimum external intervention (in fact if I don't > consider the Telemetry aspect, the rest works perfectly fine). > > I will follow Rafael guidelines to better undrstand where is the issue. > > Thanks. Something I have used in the past is a local.conf like this: [[local|localrc]] ADMIN_PASSWORD=secret DATABASE_PASSWORD=$ADMIN_PASSWORD RABBIT_PASSWORD=$ADMIN_PASSWORD SERVICE_PASSWORD=$ADMIN_PASSWORD IMAGE_URL_SITE="https://download.fedoraproject.org" IMAGE_URL_PATH="/pub/fedora/linux/releases/33/Cloud/x86_64/images/" IMAGE_URL_FILE="Fedora-Cloud-Base-33-1.2.x86_64.qcow2" IMAGE_URLS+=","$IMAGE_URL_SITE$IMAGE_URL_PATH$IMAGE_URL_FILE CEILOMETER_BACKEND=gnocchi USE_PYTHON3="True" VERBOSE="True" VERBOSE_NO_TIMESTAMP="True" # LIBS_FROM_GIT=aodh,ceilometer,cinder,devstack,devstack-gate,dib-utils,diskimage-builder,glance,heat,keystone,neutron,nova,placement,requirements,swift enable_plugin aodh https://opendev.org/openstack/aodh enable_plugin ceilometer https://opendev.org/openstack/ceilometer enable_plugin heat https://opendev.org/openstack/heat INSTALL_TEMPEST=False Hope that helps. The IMAGE_URL_FILE needs probably adjustment, fedora 33 is not supported anymore. Matthias > > > ------------------------------------------------------------------------ > *Da:* Matthias Runge > > *Inviato:* gioved? 3 marzo 2022 11:59 > *A:* openstack-discuss at lists.openstack.org > > > *Oggetto:* Re: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > On 3/3/22 10:29, Molka Gharbaoui wrote: >> Thank you Matthias for your reply. >> >> Unfortunately, from the logs I still cannot understand what is the issue. >> >> Could someone kindly indicate me how ceilometer and gnocchi should be >> configured to get the data? >> >> For example in this link: >> https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > >> > the > >> endpoints should use the following URL: http://controller:8041 >> > while in my case the > 8041 port is not even >> listening and the API indicated in the dashboard for Gnocchi is http:// >> >x.x.x.x/metric. >> >> Thank you in advance. >> > > OpenStack configuration is complex. The guide does not remotely touch on > production environments. Please do not confuse devstack with production > envionments. Also the doc you linked is not for devstack itself. > > The flow in general is: data is generated by the services like nova or > glance. Ceilometer listens on the bus and also polls for information. > The info is sent then over (as configured in publishers.yaml) to gnocchi. > Gnocchi has an api, ceilometer has not. Reading your response, this > makes me suspicious if there is gnocchi running. From the initial > question it suggests it is, from your most recent response, it is not. > > > Matthias > -- Rafael Weing?rtner -- Rafael Weing?rtner -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From senrique at redhat.com Fri Mar 4 19:47:39 2022 From: senrique at redhat.com (Sofia Enriquez) Date: Fri, 4 Mar 2022 16:47:39 -0300 Subject: Xena and CEPH RBD backend (show_image_direct_url status ) In-Reply-To: <20220228140433.Horde.HqHowF97yyXPEK-I97zzVWP@webmail.nde.ag> References: <20220228140433.Horde.HqHowF97yyXPEK-I97zzVWP@webmail.nde.ag> Message-ID: Greetings, Have you opened a bug report for this issue? If not, It would be a great help if you could file a bug: https://bugs.launchpad.net/cinder/. The cinder team holds a bug meeting every week. In order to reproduce and discuss this issue further we need a bug report with all the context of a bug. It's also really good to track fixes. Please at least provide the next information on it: description, version details, crystal clear details to reproduce the bug (steps), environment details and actual results like cinder logs if any. [1] Thanks, Sofia [1] https://wiki.openstack.org/wiki/BugFilingRecommendations On Mon, Feb 28, 2022 at 11:07 AM Eugen Block wrote: > Hi, > > it's disappointing that this is still an issue. > We're currently using OpenStack Ussuri with Ceph Nautilus (we plan the > Upgrade to Octopus soon) which works fine without enabling > show_image_direct_url. The same goes for Victoria and Octopus (one of > our customers uses this combination). > > > > How is the noted GRAVE Security RISK of enabling > > Show_image_direct_url mitigated ? (i.e I think , for CEPH RBD, it > > needs to be True to get cloning to work efficiently) > > I'm also wondering in which case the location contains credentials, I > haven't seen that yet. Depending on how your cloud is used (is it a > public or private cloud) maybe enabling the option is not that big of > a risk? > > Regards, > Eugen > > > Zitat von "west, andrew" : > > > Hello experts > > > > Currently using openstack Xena and Ceph backend (Pacific 16.2.7) > > > > It seems there is a bug (since Wallaby?) where the efficient use of > > a CEPH Pacific RBD backend (i.e with copy-on-write-cloning) is not > > working . > > Show_image_direct_url needs to be False to create volumes (or > > ephemeral volumes for nova) > > > > This can of course be tremendously slow (Nova , ephemeral root > > disk) without copy-on-write cloning feature of Ceph. > > > > As Ceph RBD is THE most favourite backend for block storage in > > openstack I am wondering how others are coping (or workarounds found > > ?) > > Which combinations of Openstack and Ceph are known to work well > > with copy-on-write-cloning? > > > > How is the noted GRAVE Security RISK of enabling > > Show_image_direct_url mitigated ? (i.e I think , for CEPH RBD, it > > needs to be True to get cloning to work efficiently) > > > > > > See another report of this issue here: > > Re: Ceph Pacifif and Openstack Wallaby - ERROR > > cinder.scheduler.flows.create_volume - CEPH Filesystem Users > > (spinics.net) > > > > Thanks for any help or pointers, > > > > Andrew West > > Openstack consulting > > CGG France > > > > > > ________________________________ > > "This e-mail and any accompanying attachments are confidential. The > > information is intended solely for the use of the individual to whom > > it is addressed. Any review, disclosure, copying, distribution, or > > use of the email by others is strictly prohibited. If you are not > > the intended recipient, you must not review, disclose, copy, > > distribute or use this e-mail; please delete it from your system and > > notify the sender immediately." > > > > > -- Sof?a Enriquez she/her Software Engineer Red Hat PnT IRC: @enriquetaso @RedHat Red Hat Red Hat -------------- next part -------------- An HTML attachment was scrubbed... URL: From rafaelweingartner at gmail.com Fri Mar 4 20:54:28 2022 From: rafaelweingartner at gmail.com (=?UTF-8?Q?Rafael_Weing=C3=A4rtner?=) Date: Fri, 4 Mar 2022 17:54:28 -0300 Subject: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: Did you check Gnocchi log files? Is there some error or something happening there? On Fri, Mar 4, 2022 at 4:02 PM Molka Gharbaoui < Molka.Gharbaoui at santannapisa.it> wrote: > Yes sorry, I forgot to mention it but it is also running. > ------------------------------ > *Da:* Rafael Weing?rtner > *Inviato:* venerd? 4 marzo 2022 19:48 > *A:* Molka Gharbaoui > *Cc:* Openstack Discuss List > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > And, what about ceilometer-compute? > > On Fri, Mar 4, 2022 at 2:04 PM Molka Gharbaoui < > Molka.Gharbaoui at santannapisa.it> wrote: > > Yes, all ceilometer services are up and running: ceilometer-agent, > ceilometer-notification. Actually, all devstack services are up and running. > ------------------------------ > *Da:* Rafael Weing?rtner > *Inviato:* venerd? 4 marzo 2022 17:56 > *A:* Molka Gharbaoui > *Cc:* Openstack Discuss List > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > I see. Is the Ceilometer compute agent up and running? > > On Fri, Mar 4, 2022 at 1:54 PM Molka Gharbaoui < > Molka.Gharbaoui at santannapisa.it> wrote: > > I have a running instance and I would like to retrieve some statistics on > it like cpu usage, memory usage, packets received and sent. > ------------------------------ > *Da:* Rafael Weing?rtner > *Inviato:* venerd? 4 marzo 2022 17:28 > *A:* Molka Gharbaoui > *Cc:* Openstack Discuss List > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > What do you or what would you expect to see in Gnocchi being pushed by > Ceilometer? > What are the resources (volumes, VMs, and so) that you created in the > OpenStack environment in DevStack? > > > On Fri, Mar 4, 2022 at 1:26 PM Molka Gharbaoui < > Molka.Gharbaoui at santannapisa.it> wrote: > > Unfortunately, the local.conf file shared by Matthias gives the same > result: no metrics are pushed to gnocchi. > > Gnocchi database is empty: I can create metrics and resources manually but > nothing is pushed from ceilometer to the database. So I guess the problem > is at the ceilometer level. > > Here is an extract of the ceilometer-polling logs: > > 2022-03-04 16:13:07.904 679128 INFO ceilometer.polling.manager [-] Polling > pollster disk.device.usage in the context of all_pollsters > 2022-03-04 16:13:07.904 679128 DEBUG ceilometer.compute.pollsters [-] > afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.usage volume: 5368709120 > _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/_ > _init__.py:108 > 2022-03-04 16:13:07.907 679128 INFO ceilometer.polling.manager [-] Polling > pollster disk.device.read.requests in the context of all_pollsters > 2022-03-04 16:13:07.908 679128 DEBUG ceilometer.compute.pollsters [-] > afcf32c4-e7c2-40fd-8f75-fdf0bfdb5ef9/disk.device.read.requests volume: > 6138 _stats_to_sample /opt/stack/ceilometer/ceilometer/compute/pollsters/_ > _init__.py:108 > 2022-03-04 16:13:08.038 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.038 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.pool, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.162 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.162 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.vip, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.163 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.outgoing.bytes, no resources found this > cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py: > 177 > 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] > Skipping tenant, keystone issue: admin endpoint for identity service in > RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: > admin endpoint for identity service in RegionOne region not found > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster radosgw.containers.objects.size, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster port.receive.packets, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > 2022-03-04 16:13:08.164 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster hardware.network.ip.outgoing.datagrams, no resources found this > cycle poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py: > 177 > 2022-03-04 16:13:08.355 679128 WARNING ceilometer.neutron_client [-] The > resource could not be found.: neutronclient.common.exceptions.NotFound: > The resource could not be found. > 2022-03-04 16:13:08.355 679128 DEBUG ceilometer.polling.manager [-] Skip > pollster network.services.lb.listener, no resources found this cycle > poll_and_notify /opt/stack/ceilometer/ceilometer/polling/manager.py:177 > > And here I report the content of the ceilometer.conf file: > [DEFAULT] > > debug = True > transport_url = rabbit://stackrabbit:admin at 10.30.2.194:5672/ > > [oslo_messaging_notifications] > topics = notifications > > [coordination] > backend_url = redis://localhost:6379 > > [notification] > workers = 2 > > [cache] > backend_argument = url:redis://localhost:6379 > backend_argument = distributed_lock:True > backend_argument = db:0 > backend_argument = redis_expiration_time:600 > backend = dogpile.cache.redis > enabled = True > > [service_credentials] > auth_url = http://10.30.2.194/identity > region_name = RegionOne > password = admin > username = ceilometer > project_name = service > project_domain_name = default > user_domain_id = default > auth_type = password > > [keystone_authtoken] > memcached_servers = localhost:11211 > cafile = /opt/stack/data/ca-bundle.pem > project_domain_name = default > project_name = admin > user_domain_name = default > password = admin > username = ceilometer > auth_url = http://10.30.2.194/identity > interface = public > auth_type = password > > I believe nothing should be changed in the configuration files since > Devstack should configure every components automatically. I also checked > pipeline.yaml and polling.yaml files and they seem correct. > > Do you have any idea on how to make ceilometer push metrics to the gnocchi > database?? > > Thanks. > > > ------------------------------ > *Da:* Matthias Runge > *Inviato:* gioved? 3 marzo 2022 13:58 > *A:* Molka Gharbaoui > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > On 3/3/22 12:31, Molka Gharbaoui wrote: > > Hi Matthias, > > > > Sorry for the confusion. Gnocchi is running and I am able to execute > > commands like "gnocchi metric list" and "gnocchi resource list" without > > errors even though the response is empty. Also checking mysql, the > > gnocchi database is correctly created. > > > > By the link I previously shared I wanted just to highlight that there > > are some differences between what was written there and between the > > environemnt I installed. I was searching there because I would have > > expected that Devstack installs and configures every component in > > automatic with no or minimum external intervention (in fact if I don't > > consider the Telemetry aspect, the rest works perfectly fine). > > > > I will follow Rafael guidelines to better undrstand where is the issue. > > > > Thanks. > > > Something I have used in the past is a local.conf like this: > > [[local|localrc]] > ADMIN_PASSWORD=secret > DATABASE_PASSWORD=$ADMIN_PASSWORD > RABBIT_PASSWORD=$ADMIN_PASSWORD > SERVICE_PASSWORD=$ADMIN_PASSWORD > > IMAGE_URL_SITE="https://download.fedoraproject.org" > IMAGE_URL_PATH="/pub/fedora/linux/releases/33/Cloud/x86_64/images/" > IMAGE_URL_FILE="Fedora-Cloud-Base-33-1.2.x86_64.qcow2" > IMAGE_URLS+=","$IMAGE_URL_SITE$IMAGE_URL_PATH$IMAGE_URL_FILE > > CEILOMETER_BACKEND=gnocchi > USE_PYTHON3="True" > > VERBOSE="True" > > VERBOSE_NO_TIMESTAMP="True" > > # > > LIBS_FROM_GIT=aodh,ceilometer,cinder,devstack,devstack-gate,dib-utils,diskimage-builder,glance,heat,keystone,neutron,nova,placement,requirements,swift > > enable_plugin aodh https://opendev.org/openstack/aodh > enable_plugin ceilometer https://opendev.org/openstack/ceilometer > enable_plugin heat https://opendev.org/openstack/heat > > > INSTALL_TEMPEST=False > > > Hope that helps. The IMAGE_URL_FILE needs probably adjustment, fedora 33 > is not supported anymore. > > Matthias > > > > > > > > > ------------------------------------------------------------------------ > > *Da:* Matthias Runge > > *Inviato:* gioved? 3 marzo 2022 11:59 > > *A:* openstack-discuss at lists.openstack.org > > > > *Oggetto:* Re: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > > metrics using ceilometer and gnocchi > > On 3/3/22 10:29, Molka Gharbaoui wrote: > >> Thank you Matthias for your reply. > >> > >> Unfortunately, from the logs I still cannot understand what is the > issue. > >> > >> Could someone kindly indicate me how ceilometer and gnocchi should be > >> configured to get the data? > >> > >> For example in this link: > >> > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html> > > > > >> < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html > > < > https://docs.openstack.org/ceilometer/latest/install/install-base-ubuntu.html>> the > > > > >> endpoints should use the following URL: http://controller:8041 < > http://controller:8041> > >> > while in my case the > > 8041 port is not even > >> listening and the API indicated in the dashboard for Gnocchi is http:// > >> >x.x.x.x/metric. > >> > >> Thank you in advance. > >> > > > > OpenStack configuration is complex. The guide does not remotely touch on > > production environments. Please do not confuse devstack with production > > envionments. Also the doc you linked is not for devstack itself. > > > > The flow in general is: data is generated by the services like nova or > > glance. Ceilometer listens on the bus and also polls for information. > > The info is sent then over (as configured in publishers.yaml) to gnocchi. > > Gnocchi has an api, ceilometer has not. Reading your response, this > > makes me suspicious if there is gnocchi running. From the initial > > question it suggests it is, from your most recent response, it is not. > > > > > > Matthias > > > > > > -- > Rafael Weing?rtner > > > > -- > Rafael Weing?rtner > > > > -- > Rafael Weing?rtner > -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From Molka.Gharbaoui at santannapisa.it Sat Mar 5 14:12:53 2022 From: Molka.Gharbaoui at santannapisa.it (Molka Gharbaoui) Date: Sat, 5 Mar 2022 14:12:53 +0000 Subject: R: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: No errors in the gnocchi log files. Here is part of the output of the "gnocchi metric list --debug": http://10.30.2.194:80 "GET /metric/v1/metric HTTP/1.1" 200 2 RESP: [200] Connection: close Content-Length: 2 Content-Type: application/json Date: Sat, 05 Mar 2022 09:33:24 GMT Server: Apache/2.4.41 (Ubuntu) RESP BODY: [] The response body is empty since the database is empty. Other output from the log files: [697660] DEBUG gnocchi.chef: Processing measures for sack incoming128-123 2022-03-05 10:17:59,042 [697663] DEBUG gnocchi.chef: Processing measures for sack incoming128-23 [697660] DEBUG gnocchi.cli.metricd: Full scan of sacks has been done There is only one error in all the log files and it regards ceilometer logs. I report it here: 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] Skipping tenant, keystone issue: admin endpoint for identity service in RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: admin endpoint for identity service in RegionOne region not found Thank you for your help! ________________________________ Da: Rafael Weing?rtner Inviato: venerd? 4 marzo 2022 21:54 A: Molka Gharbaoui Cc: Openstack Discuss List Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi Did you check Gnocchi log files? Is there some error or something happening there? -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From rafaelweingartner at gmail.com Sat Mar 5 15:00:19 2022 From: rafaelweingartner at gmail.com (=?UTF-8?Q?Rafael_Weing=C3=A4rtner?=) Date: Sat, 5 Mar 2022 12:00:19 -0300 Subject: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: The messages: "%d measures posted against %d metrics through %d resources" appear in the log files of Ceilometer? On Sat, Mar 5, 2022 at 11:12 AM Molka Gharbaoui < Molka.Gharbaoui at santannapisa.it> wrote: > No errors in the gnocchi log files. > > Here is part of the output of the "gnocchi metric list --debug": > > http://10.30.2.194:80 "GET /metric/v1/metric HTTP/1.1" 200 2 > RESP: [200] Connection: close Content-Length: 2 Content-Type: > application/json Date: Sat, 05 Mar 2022 09:33:24 GMT Server: Apache/2.4.41 > (Ubuntu) > RESP BODY: [] > > The response body is empty since the database is empty. Other output from > the log files: > > [697660] DEBUG gnocchi.chef: Processing measures for sack > incoming128-123 > 2022-03-05 10:17:59,042 [697663] DEBUG gnocchi.chef: Processing > measures for sack incoming128-23 > [697660] DEBUG gnocchi.cli.metricd: Full scan of sacks has been done > > There is only one error in all the log files and it regards ceilometer > logs. I report it here: > > 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] > Skipping tenant, keystone issue: admin endpoint for identity service in > RegionOne region not found: > keystoneauth1.exceptions.catalog.EndpointNotFound: admin endpoint for > identity service in RegionOne region not found > > Thank you for your help! > > ------------------------------ > *Da:* Rafael Weing?rtner > *Inviato:* venerd? 4 marzo 2022 21:54 > *A:* Molka Gharbaoui > *Cc:* Openstack Discuss List > *Oggetto:* Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve > metrics using ceilometer and gnocchi > > Did you check Gnocchi log files? Is there some error or something > happening there? > > -- > Rafael Weing?rtner > -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From Molka.Gharbaoui at santannapisa.it Sat Mar 5 15:19:24 2022 From: Molka.Gharbaoui at santannapisa.it (Molka Gharbaoui) Date: Sat, 5 Mar 2022 15:19:24 +0000 Subject: R: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi In-Reply-To: References: <2a04ab06-1651-eb25-f0f7-29b3a4805541@matthias-runge.de> <414fa552-988a-48a7-c331-c58df442617f@matthias-runge.de> Message-ID: No, the message: "%d measures posted against %d metrics through %d resources" does not appear in ceilometer logs. I executed: ceilometer-upgrade, ceilometer-polling and ceilometer-agent-notification commands. ________________________________ Da: Rafael Weing?rtner Inviato: sabato 5 marzo 2022 16:00 A: Molka Gharbaoui Cc: Openstack Discuss List Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi The messages: "%d measures posted against %d metrics through %d resources" appear in the log files of Ceilometer? On Sat, Mar 5, 2022 at 11:12 AM Molka Gharbaoui > wrote: No errors in the gnocchi log files. Here is part of the output of the "gnocchi metric list --debug": http://10.30.2.194:80 "GET /metric/v1/metric HTTP/1.1" 200 2 RESP: [200] Connection: close Content-Length: 2 Content-Type: application/json Date: Sat, 05 Mar 2022 09:33:24 GMT Server: Apache/2.4.41 (Ubuntu) RESP BODY: [] The response body is empty since the database is empty. Other output from the log files: [697660] DEBUG gnocchi.chef: Processing measures for sack incoming128-123 2022-03-05 10:17:59,042 [697663] DEBUG gnocchi.chef: Processing measures for sack incoming128-23 [697660] DEBUG gnocchi.cli.metricd: Full scan of sacks has been done There is only one error in all the log files and it regards ceilometer logs. I report it here: 2022-03-04 16:13:08.164 679128 ERROR ceilometer.polling.manager [-] Skipping tenant, keystone issue: admin endpoint for identity service in RegionOne region not found: keystoneauth1.exceptions.catalog.EndpointNotFound: admin endpoint for identity service in RegionOne region not found Thank you for your help! ________________________________ Da: Rafael Weing?rtner > Inviato: venerd? 4 marzo 2022 21:54 A: Molka Gharbaoui > Cc: Openstack Discuss List > Oggetto: Re: R: R: [Devstack][Gnocchi][Ceilometer] Not able to retrieve metrics using ceilometer and gnocchi Did you check Gnocchi log files? Is there some error or something happening there? -- Rafael Weing?rtner -- Rafael Weing?rtner -------------- next part -------------- An HTML attachment was scrubbed... URL: From skaplons at redhat.com Sat Mar 5 17:20:18 2022 From: skaplons at redhat.com (Slawek Kaplonski) Date: Sat, 05 Mar 2022 18:20:18 +0100 Subject: [adjutant][tc][all] Call for volunteers to be a PTL and maintainers Message-ID: <4381995.LvFx2qVVIh@p1> Hi, After last PTL elections [1] Adjutant project don't have any PTL. It also didn't had PTL in the Yoga cycle already. So this is call for maintainters for Adjutant. If You are using it or interested in it, and if You are willing to help maintaining this project, please contact TC members through this mailing list or directly on the #openstack-tc channel @OFTC. We can talk possibilities to make someone a PTL of the project or going with this project to the Distributed Project Leadership [2] model. [1] http://lists.openstack.org/pipermail/openstack-discuss/2022-February/027411.html [2] https://governance.openstack.org/tc/resolutions/20200803-distributed-project-leadership.html -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From skaplons at redhat.com Sat Mar 5 17:25:42 2022 From: skaplons at redhat.com (Slawek Kaplonski) Date: Sat, 05 Mar 2022 18:25:42 +0100 Subject: [trove][tc][all] Call for volunteers to be a PTL and maintainers Message-ID: <2095127.irdbgypaU6@p1> Hi, After last PTL elections [1] Trove project still don't have any PTL elected. So this is call for maintainers for Trove. If You are using it or interested in it, and if You are willing to help maintaining this project, please contact TC members through this mailing list or directly on the #openstack-tc channel @OFTC. We can talk possibilities to make someone a PTL of the project or going with this project to the Distributed Project Leadership [2] model. [1] http://lists.openstack.org/pipermail/openstack-discuss/2022-February/027411.html [2] https://governance.openstack.org/tc/resolutions/20200803-distributed-project-leadership.html -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From lokendrarathour at gmail.com Sun Mar 6 04:20:26 2022 From: lokendrarathour at gmail.com (Lokendra Rathour) Date: Sun, 6 Mar 2022 09:50:26 +0530 Subject: Fwd: [Triple0] Upgrade Error In-Reply-To: References: <87333433-63c3-6b52-ab36-8de16dff277d@redhat.com> Message-ID: Hi Team/Harald, Any input possible on this please? Thanks. On Fri, 4 Mar 2022, 15:10 Lokendra Rathour, wrote: > Hi Harald / Team, > Thanks for the support here. > It did help in the classification of the issue and we saw that we need to > re-render the templates for an upgrade to proceed. > After re-rendering we saw the diff of DNS related in the templates. > > Now we have completed the next few steps and are facing issues when > running the below commands: > > openstack overcloud upgrade run --limit Controller --playbook upgrade_steps_playbook.yaml > > > It gives the error as : > > *2022-03-04 14:29:56.571690 | 525400cb-2e90-4086-e760-00000000009b | > TASK | Ensure DNF modules have the right stream* > *2022-03-04 14:29:57.723661 | 525400cb-2e90-4086-e760-00000000009b | > FATAL | Ensure DNF modules have the right stream | overcloud-novacompute-0 > | item={'module': 'container-tools', 'stream': '3.0'} | > error={"ansible_loop_var": "item", "changed": false, "item": {"module": > "container-tools", "stream": "3.0"}, "module_stderr": "Traceback (most > recent call last):\n File \"\", line 102, in \n File > \"\", line 94, in _ansiballz_main\n File \"\", line 40, in > invoke_module\n File \"/usr/lib64/python3.6/runpy.py\", line 205, in > run_module\n return _run_module_code(code, init_globals, run_name, > mod_spec)\n File \"/usr/lib64/python3.6/runpy.py\", line 96, in > _run_module_code\n mod_name, mod_spec, pkg_name, script_name)\n File > \"/usr/lib64/python3.6/runpy.py\", line 85, in _run_code\n exec(code, > run_globals)\n File > \"/tmp/ansible_dnf_payload_3lb54omj/ansible_dnf_payload.zip/ansible/modules/packaging/os/dnf.py\", > line 1370, in \n File > \"/tmp/ansible_dnf_payload_3lb54omj/ansible_dnf_payload.zip/ansible/modules/packaging/os/dnf.py\", > line 1359, in main\n File > \"/tmp/ansible_dnf_payload_3lb54omj/ansible_dnf_payload.zip/ansible/modules/packaging/os/dnf.py\", > line 1338, in run\n File > \"/tmp/ansible_dnf_payload_3lb54omj/ansible_dnf_payload.zip/ansible/modules/packaging/os/dnf.py\", > line 996, in ensure\n File > \"/usr/lib/python3.6/site-packages/dnf/module/module_base.py\", line 144, > in install\n \"Installing module from Fail-Safe repository is not > allowed\"))\ndnf.exceptions.Error: Installing module from Fail-Safe > repository is not allowed\n", "module_stdout": "", "msg": "MODULE > FAILURE\nSee stdout/stderr for the exact error", "rc": 1}* > > *Few things to note:* > > - *the document followed * Upgrading to a Next Major Release ? TripleO > 3.0.0 documentation (openstack.org) > > - The command which was executed to run upgrade prepare: > - > > > > > > > > > > > > *openstack overcloud upgrade prepare --templates \ -n > /home/stack/templates/network_data.yaml \ -r > /home/stack/templates/roles_data.yaml \ -e > /home/stack/templates/environment.yaml \ -e > /home/stack/templates/environments/network-isolation.yaml \ -e > /home/stack/templates/environments/network-environment.yaml \ -e > /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-conductor.yaml > \ -e > /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml > \ -e > /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-overcloud.yaml > \ -e /home/stack/templates/ironic-config.yaml \ -e > /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml \ > -e /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ > -e /home/stack/containers-prepare-parameter.yaml* > > Please advice over the same issue. > > Best Regards, > Lokendra > > > On Sun, Feb 27, 2022 at 6:50 AM Harald Jensas wrote: > >> On 2/25/22 12:21, Lokendra Rathour wrote: >> > Hi Team, >> > >> > we are trying to upgrade *Triple0 Train to Ussuri.* >> > Undercloud upgrade is done successfully. >> > *_OverCloud Upgrade:_* >> > As a first step in overcloud upgrade, we are running " *openstack >> > overcloud upgrade prepare* " command. >> > which is throwing following error: >> > >> >> > overcloud: ERROR: Property error: >> > resources.InternalApiVirtualIP.properties: Unknown Property DnsName\n >> > File \"/usr/lib/python3.6/site-packages/heat/common/wsgi.py\", line >> > 891, in __call__\n request, **action_args)\n File >> > \"/usr/lib/python3.6/site-packages/heat/common/wsgi.py\", line 964, in >> > dispatch\n return method(*args, **kwargs)\n File >> > \"/usr/lib/python3.6/site-packages/heat/api/openstack/v1/util.py\", >> line >> > 56, in handle_stack_method\n return handler(controller, req, >> > **kwargs)\n File >> > \"/usr/lib/python3.6/site-packages/heat/api/openstack/v1/stacks.py\", >> > line 633, in validate_template\n raise >> > exc.HTTPBadRequest(result['Error'])\n", "success": false}* >> > >> >> The DnsName property was added to templats in Ussuri, this change: >> https://review.opendev.org/c/openstack/tripleo-heat-templates/+/715883 >> >> >> Somewhere in you environment files you are mapping >> `OS::TripleO::Network::Ports::InternalApiVipPort` to a local template >> file instead of using the in-tree template. >> >> You either need to add DnsName paramter to your local VipPort template, >> or change the resource_registry so that you map to the in-tree port >> template. >> `/usr/share/openstack-tripleo-heat-template/network/ports/internal_api.yaml` >> >> is the in-tree template you want. >> >> >> > *_Command used to run the overcloud upgrade prepare:_* >> > >> > *(undercloud) [stack at undercloud ~]$ cat upgrade_overcloud_prepare.sh >> > openstack overcloud upgrade prepare --templates \ >> > -r /home/stack/templates/roles_data.yaml \ >> > -n /home/stack/templates/network_data.yaml \ >> > -e /home/stack/templates/environment.yaml \ >> > -e /home/stack/templates/environments/network-isolation.yaml \ >> >> Most likely the mapping is in this network-isolation.yaml. >> If you use the in-tree environment >> `/usr/share/openstack-tripleo-heat-tempaltes/envioronments/network-isolation.yaml` >> >> instead you should not run into these type of issues on upgrade. >> >> > -e /home/stack/templates/environments/network-environment.yaml \ >> >> I would also recommend to use the in-tree network-environment.yaml >> -e >> >> /usr/share/openstack-tripleo-heat-tempaltes/envioronments/network-environment.yaml >> >> Then add another file to override anything in the in-tree >> network-environment.yaml. >> -e /home/stack/templates/environments/network-environment-overrides.yaml >> >> > -e >> > >> /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-conductor.yaml >> >> > \ >> > -e >> > >> /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-inspector.yaml >> >> > \ >> > -e >> > >> /usr/share/openstack-tripleo-heat-templates/environments/services/ironic-overcloud.yaml >> >> > \ >> > -e /home/stack/templates/ironic-config.yaml \ >> > -e >> > /usr/share/openstack-tripleo-heat-templates/environments/docker-ha.yaml >> \ >> > -e >> > /usr/share/openstack-tripleo-heat-templates/environments/podman.yaml \ >> > -e /home/stack/templates/init-repo.yaml \ >> > -e /home/stack/containers-prepare-parameter.yaml >> >> >> -- >> Harald >> >> > > -- > ~ Lokendra > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From radoslaw.piliszek at gmail.com Sun Mar 6 08:46:48 2022 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Sun, 6 Mar 2022 09:46:48 +0100 Subject: [ironic][tripleo][kayobe][tenks][vbmc] Do you also experience random issues with vbmc? Message-ID: Dear fellow upstream OpenStackers, I tagged the main projects that use vbmc intensively because I have a question regarding the stability of vbmc. I have been working on adding baremetal testing in Kolla Ansible using Tenks and thus also VirtualBMC. I am encountering random failures that I am at the moment unable to debug. It seems vbmc likes to reject the "chassis bootdev pxe" ipmitool command from time to time (the command is issued when trying to deploy a host). This seems permanent per case as ironic retries this call several times only to fail. The error message is "Error setting Chassis Boot Parameter 5". I can't see any trace in vbmc logs (with DEBUG enabled). Other ipmitool commands seem to work fine (like periodic power status). I was wondering if any of you also experienced this issue. I am providing a link to an excerpt from the logs. [1] It also shows a raw ipmitool command succeeding only to let the next one fail. [1] https://paste.ubuntu.com/p/MgHHpDvVGf/ -yoctozepto From marc-antoine.godde at viarezo.fr Mon Mar 7 01:36:24 2022 From: marc-antoine.godde at viarezo.fr (Marc-Antoine Godde) Date: Mon, 7 Mar 2022 02:36:24 +0100 Subject: IPv6 deployment on OpenStack Message-ID: Hello. We are progressively adding support for IPv6 in my company. We decided to use SLAAC only for laptops, phones, ? since DHCPv6 isn?t supported on Android. RDNSS support will also increase. We are now planning our deployment on OpenStack. We already know that we'll rely only on neutron but we are not yet fixed between DHCPv6 and SLAAC ? Do you have any arguments for one these for VMs ? Thanks, Marc-Antoine. -------------- next part -------------- An HTML attachment was scrubbed... URL: From skaplons at redhat.com Mon Mar 7 08:12:27 2022 From: skaplons at redhat.com (Slawek Kaplonski) Date: Mon, 07 Mar 2022 09:12:27 +0100 Subject: IPv6 deployment on OpenStack In-Reply-To: References: Message-ID: <14575046.tv2OnDr8pf@p1> Hi, On poniedzia?ek, 7 marca 2022 02:36:24 CET Marc-Antoine Godde wrote: > Hello. > > We are progressively adding support for IPv6 in my company. We decided to use SLAAC only for laptops, phones, ? since DHCPv6 isn?t supported on Android. RDNSS support will also increase. We are now planning our deployment on OpenStack. We already know that we'll rely only on neutron but we are not yet fixed between DHCPv6 and SLAAC ? Do you have any arguments for one these for VMs ? > > Thanks, > Marc-Antoine. > With SLAAC You need to have Your network connected to the router in Neutron and You can only configure IP address on the VM. With DHCPv6 You can configure other things, like some static-routes, etc. Neutron supports DHCPv6 in the stateful and stateless variants. With stateless, You are using RA for address configuration and DHCP server for other configation. Please see [1] for more details. [1] https://docs.openstack.org/neutron/latest/admin/config-ipv6.html#address-modes-for-ports -- Slawek Kaplonski Principal Software Engineer Red Hat -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: This is a digitally signed message part. URL: From mark at stackhpc.com Mon Mar 7 08:51:12 2022 From: mark at stackhpc.com (Mark Goddard) Date: Mon, 7 Mar 2022 08:51:12 +0000 Subject: [ironic][tripleo][kayobe][tenks][vbmc] Do you also experience random issues with vbmc? In-Reply-To: References: Message-ID: On Sun, 6 Mar 2022 at 08:51, Rados?aw Piliszek wrote: > > Dear fellow upstream OpenStackers, > > I tagged the main projects that use vbmc intensively because I have a > question regarding the stability of vbmc. > I have been working on adding baremetal testing in Kolla Ansible using > Tenks and thus also VirtualBMC. > I am encountering random failures that I am at the moment unable to debug. > It seems vbmc likes to reject the "chassis bootdev pxe" ipmitool > command from time to time (the command is issued when trying to deploy > a host). > This seems permanent per case as ironic retries this call several > times only to fail. > The error message is "Error setting Chassis Boot Parameter 5". > I can't see any trace in vbmc logs (with DEBUG enabled). > Other ipmitool commands seem to work fine (like periodic power status). > I was wondering if any of you also experienced this issue. > I am providing a link to an excerpt from the logs. [1] > It also shows a raw ipmitool command succeeding only to let the next one fail. > > [1] https://paste.ubuntu.com/p/MgHHpDvVGf/ Radek, Try this on for size: https://review.opendev.org/c/openstack/kayobe-config-dev/+/829006 Mark > > -yoctozepto > From mark at stackhpc.com Mon Mar 7 08:56:34 2022 From: mark at stackhpc.com (Mark Goddard) Date: Mon, 7 Mar 2022 08:56:34 +0000 Subject: [kolla-ansible][wallaby][centos][horizon-dashboard] Changing logo.svg In-Reply-To: <60FB6A20-20EA-4CF7-A962-3CF8AA9C88AA@univ-grenoble-alpes.fr> References: <60FB6A20-20EA-4CF7-A962-3CF8AA9C88AA@univ-grenoble-alpes.fr> Message-ID: Hi Franck, Are you sure you're using the new image? If you don't change the tag, you'll need to explicitly pull the image from your registry. Mark On Fri, 4 Mar 2022 at 17:14, Franck VEDEL wrote: > > Hello, > I come back to ask for some help with a new problem on my Openstack platform. For once, it's a small problem. > I take this opportunity to thank those who helped me solve my problem which was the following: the instances were not created regularly on the 3 nova servers. Nova-scheduler was malfunctioning. I fixed the problem by setting build_failure_weight_multiplier=0 > Since then, in DEBUG mode, we can clearly see the calculation of the weight and the distribution of the instances. It's beautiful. > > My new problem is: I'm using kolla-ansible, centos (stream) and Wallaby. > In the directory /usr/local/lib/python3.6/site-packages/openstack_dashboard/static/dashboard/img/ I modified the 3 files logo.svg, logo-splash.svg, favicon.ico. > > I thought doing a "kolla-ansible -i multinode reconfigure horizon" would update the logos and favicon.ico....but no, it does nothing. > How do I change these logos with kolla-ansible? Is it difficult? > > thank you in advance > > > Franck VEDEL > > From radoslaw.piliszek at gmail.com Mon Mar 7 09:24:48 2022 From: radoslaw.piliszek at gmail.com (=?UTF-8?Q?Rados=C5=82aw_Piliszek?=) Date: Mon, 7 Mar 2022 10:24:48 +0100 Subject: [ironic][tripleo][kayobe][tenks][vbmc] Do you also experience random issues with vbmc? In-Reply-To: References: Message-ID: On Mon, 7 Mar 2022 at 09:51, Mark Goddard wrote: > > On Sun, 6 Mar 2022 at 08:51, Rados?aw Piliszek > wrote: > > > > Dear fellow upstream OpenStackers, > > > > I tagged the main projects that use vbmc intensively because I have a > > question regarding the stability of vbmc. > > I have been working on adding baremetal testing in Kolla Ansible using > > Tenks and thus also VirtualBMC. > > I am encountering random failures that I am at the moment unable to debug. > > It seems vbmc likes to reject the "chassis bootdev pxe" ipmitool > > command from time to time (the command is issued when trying to deploy > > a host). > > This seems permanent per case as ironic retries this call several > > times only to fail. > > The error message is "Error setting Chassis Boot Parameter 5". > > I can't see any trace in vbmc logs (with DEBUG enabled). > > Other ipmitool commands seem to work fine (like periodic power status). > > I was wondering if any of you also experienced this issue. > > I am providing a link to an excerpt from the logs. [1] > > It also shows a raw ipmitool command succeeding only to let the next one fail. > > > > [1] https://paste.ubuntu.com/p/MgHHpDvVGf/ > > Radek, > > Try this on for size: > https://review.opendev.org/c/openstack/kayobe-config-dev/+/829006 Thanks, Mark! Trying it out now in [2]. [2] https://review.opendev.org/c/openstack/kolla-ansible/+/832204 -yoctozepto From marc-antoine.godde at viarezo.fr Mon Mar 7 09:36:30 2022 From: marc-antoine.godde at viarezo.fr (Marc-Antoine Godde) Date: Mon, 7 Mar 2022 10:36:30 +0100 Subject: IPv6 deployment on OpenStack In-Reply-To: <14575046.tv2OnDr8pf@p1> References: <14575046.tv2OnDr8pf@p1> Message-ID: <8195F0AD-5FE8-432C-A1A7-8EB287B13C64@viarezo.fr> Hello, Thanks for your answer. If I?m correct, we can just use a virtual router with SLAAC since RADVD can deal with RS and emit RA (with support for RFC6106), right ? More generally, aren?t we suppose to have a virtual router every time, even in DHCPv6 (stateless and statefull), to answer RS ? I have to admit that I?m not very familiar at the moment with the implementations of RFCs in OpenStack. Currently, we prefer the idea of adding IPv6 through SLAAC to have a uniform network. If we do so, we?d like to avoid sending RA from our physical router to limit its load. Yet, we do not any other arguments to support this choice. Do you have any recommendations on what to do in latest versions of OpenStack ? What is usually done ? Thanks, Marc-Antoine > Le 7 mars 2022 ? 09:12, Slawek Kaplonski a ?crit : > > Hi, > > On poniedzia?ek, 7 marca 2022 02:36:24 CET Marc-Antoine Godde wrote: >> Hello. >> >> We are progressively adding support for IPv6 in my company. We decided to use SLAAC only for laptops, phones, ? since DHCPv6 isn?t supported on Android. RDNSS support will also increase. We are now planning our deployment on OpenStack. We already know that we'll rely only on neutron but we are not yet fixed between DHCPv6 and SLAAC ? Do you have any arguments for one these for VMs ? >> >> Thanks, >> Marc-Antoine. >> > > With SLAAC You need to have Your network connected to the router in Neutron and You can only configure IP address on the VM. With DHCPv6 You can configure other things, like some static-routes, etc. > Neutron supports DHCPv6 in the stateful and stateless variants. With stateless, You are using RA for address configuration and DHCP server for other configation. Please see [1] for more details. > > [1] https://docs.openstack.org/neutron/latest/admin/config-ipv6.html#address-modes-for-ports > > -- > Slawek Kaplonski > Principal Software Engineer > Red Hat From ignaziocassano at gmail.com Mon Mar 7 09:47:19 2022 From: ignaziocassano at gmail.com (Ignazio Cassano) Date: Mon, 7 Mar 2022 10:47:19 +0100 Subject: [openstack][queens][cinder] volume retype errors Message-ID: Hello, sometimes the volume retype from a netapp nfs storage to another netapp nfs storage does not work. I do not know the reason but I think something is going wrong in nova: /var/log/nova/nova-compute.log:2022-02-02 14:23:39.992 194288 ERROR root [req-ebbf749f-359f-4282-94c5-0913cffe97a3 66adb965bef64eaaab2af93ade87e2ca 68feefdf7c5847388534e9214dc13366 - default default] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 5655, in _swap_volume\n mountpoint, resize_to)\n', ' File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 1647, in swap_volume\n self._disconnect_volume(context, new_connection_info, instance)\n', ' File "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 220, in __exit__\n self.force_reraise()\n', ' File "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 196, in force_reraise\n six.reraise(self.type_, self.value, self.tb)\n', ' File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 1644, in swap_volume\n self._swap_volume(guest, disk_dev, conf, resize_to)\n', ' File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 1593, in _swap_volume\n raise exception.VolumeRebaseFailed(reason=six.text_type(exc))\n', "VolumeRebaseFailed: Volume rebase failed: Requested operation is not valid: pivot of disk 'vda' requires an active copy job\n"]: VolumeAttachmentNotFound: Volume attachment 2cd820e0-85e8-498d-a62a-800260d0cf31 could not be found Must I try to change some parameters in nova.conf on compute nodes about volume attachment ? Thanks Ignazio -------------- next part -------------- An HTML attachment was scrubbed... URL: From amonster369 at gmail.com Mon Mar 7 11:06:31 2022 From: amonster369 at gmail.com (A Monster) Date: Mon, 7 Mar 2022 12:06:31 +0100 Subject: Use idle backup servers Message-ID: Is there a way to use multiple servers as backup but these servers stay idle until the main server fails for example two controller servers but only one of them which is running services, and in case this main server fails, the cloud switches to the other server. -------------- next part -------------- An HTML attachment was scrubbed... URL: From mdemaced at redhat.com Mon Mar 7 11:07:09 2022 From: mdemaced at redhat.com (Maysa De Macedo Souza) Date: Mon, 7 Mar 2022 12:07:09 +0100 Subject: [kuryr] Zed PTG planning Message-ID: Hello everyone, The virtual PTG will happen between 4th-8th April. I've created a doodle[1] to try finding a time slot that works for most of us. There is also an etherpad[2] to propose topics to be discussed, feel free to fill it if you have one. Thank you. Kind regards, Maysa Macedo. [1] https://doodle.com/meeting/participate/id/0dN1Movb [2] https://etherpad.opendev.org/p/kuryr-zed-ptg -------------- next part -------------- An HTML attachment was scrubbed... URL: From smooney at redhat.com Mon Mar 7 11:09:48 2022 From: smooney at redhat.com (Sean Mooney) Date: Mon, 07 Mar 2022 11:09:48 +0000 Subject: [nova][neutron] os-vif params In-Reply-To: <45AC4BA4-DA72-42A1-8E7B-4DAD3870CF0B@gmail.com> References: <3457d3040994053d4ecec879ef33c9b09758774b.camel@redhat.com> <45AC4BA4-DA72-42A1-8E7B-4DAD3870CF0B@gmail.com> Message-ID: <4010df1ac7cdbae603e767aaf6d6c1d3cc26a29b.camel@redhat.com> On Fri, 2022-03-04 at 19:57 +0100, Arnaud wrote: > Hi Sean, Rodolfo, > > Thanks for your answers. > I will try to do the patch next week so these parameters are included in the nova config sample and documentation. > > About the isolated_vif params. It seems like a security enhancement to set it to True when using ovs? Do you confirm that? yes if you are usging ovs with ml2/ovs and the l2 agent then you shoudl enable isolated_vif to enaure that the port are added to the "dead" vlan 4095 which will drop all packets until the l2 agent moves the port onto the correct local vlan for the neutron netowrk. this impoves tenant isolation during move operations. it is not requried/tested with ovn, odl or other sdn contolers that use ovs as a switch. they do not use vlans for local tenant isolation or the normal action. they use ovs in secure mode which will drop packets by default if there are no openflow rules to process the packet. as such they shoudl not have isolated_vif enabled. > > Regards, > Arnaud > > Le 4 mars 2022 18:57:40 GMT+01:00, Sean Mooney a ?crit?: > > On Fri, 2022-03-04 at 18:33 +0100, Rodolfo Alonso Hernandez wrote: > > > Hi Arnaud: > > > > > > Those parameters should be set in nova.conf, in the "os_vif_XXX" section > > > [1]. > > > > > > "network_device_mtu" is deprecated. The interface MTU is retrieved from the > > > network MTU. "ovs_vsctl_timeout" is deprecated too, we don't use vsctl > > > commands anymore. > > actully we do use vsctl at least in ci but it is deprecated. > > we are currently using it in ci to workaround a bug in the ovs python bindings. > > > > after the RC period is clear i might revert that since i think the ovs python bidnign are > > fixed but im not sure they are fixed in ubuntu 20.04 so i dont want to regress the ci stablity > > currently by reverting to the native bindings. > > > > https://review.opendev.org/c/openstack/devstack/+/796826 > > > > ill update https://github.com/openstack/devstack/blob/master/lib/os-vif#L6 to native > > once we are sure https://bugs.launchpad.net/os-vif/+bug/1929446 is fixed in the ovs python bindings. > > > > > > > > > > Maybe Nova folks can extend my answer. > > > > > > Regards. > > > > > > [1] > > > https://github.com/openstack/nova/blob/50fdbc752a9ca9c31488140ef2997ed59d861a41/releasenotes/notes/libvirt-uses-os-vif-plugins-31a0617de0c248b9.yaml > > > > > > > > > On Fri, Mar 4, 2022 at 6:18 PM Arnaud Morin wrote: > > > > > > > Hey all, > > > > > > > > When I take a look at [1], I can see some parameters, such as > > > > network_device_mtu, isolate_vif, etc. > > > > > > > > As far as I can tell, nova is using these params when plugging the port > > > > in OVS (if using OVS as ml2 plugin of course). > > > > > > > > But I cannot find the config params in the documentation: [2] > > > > > > > > Am I missing something? > > > > > > > > Cheers, > > > > Arnaud. > > > > > > > > [1] https://github.com/openstack/os-vif/blob/master/vif_plug_ovs/ovs.py > > > > [2] https://docs.openstack.org/nova/latest/configuration/config.html > > > > > > > > > > From janders at redhat.com Mon Mar 7 12:13:16 2022 From: janders at redhat.com (Jacob Anders) Date: Mon, 7 Mar 2022 22:13:16 +1000 Subject: [ironic] Zed PTG slots In-Reply-To: References: Message-ID: Thank you Iury and Riccardo. I also vote for 1 and 3. Thank you, Jacob On Fri, Mar 4, 2022 at 8:39 PM Riccardo Pittau wrote: > Thanks Iury! > > My choices are for options #1 and #3 > > Ciao! > Riccardo > > On Fri, Mar 4, 2022 at 4:28 AM Iury Gregory wrote: > >> Hello Ironicers! >> >> For the Zed PTG I'm planning to have some slots that will allow APAC >> contributors to participate in the discussions. >> >> Keep in mind that we have some guidelines we should try to follow: >> 1. Cross project discussions (like SIGs or support project teams) should >> be scheduled towards the start of the week so that any discussions that >> might shape those of other teams happen first. >> 2. No team should sign up for more than 4 hours per UTC day to help keep >> participants actively engaged. >> 3. No team should sign up for more than 16 hours across all time slots to >> avoid burning out our contributors and to enable participation in multiple >> teams discussions. >> >> Based on this I have the following proposals: (Feel free to provide more >> ideas also) =) >> *Option #1* *16hrs* Monday till Thursday >> 1h (APAC night + EMEA morning) >> 2hrs (EMEA afternoon + NA morning) >> 1hr (NA afternoon/night + APAC morning) >> >> *Option #2* *15hrs* Tuesday till Friday >> 1h (APAC night + EMEA morning) >> 2hrs (EMEA afternoon + NA morning) >> 1hr (NA afternoon/night + APAC morning) >> * On Friday the NA + APAC session won't happen because is saturday for >> APAC contributors * >> >> *Option #3* *16hrs* Monday till Thursday >> 3hrs (EMEA afternoon + NA morning) >> 1hr (NA afternoon/night + APAC morning) >> >> *Option #4* *15hrs* Tuesday till Friday >> 3hrs (EMEA afternoon + NA morning) >> 1hr (NA afternoon/night + APAC morning) >> * On Friday the NA + APAC session won't happen because is saturday for >> APAC contributors * >> >> Please choose two options and we will vote during our upstream meeting on >> March 7th. >> >> Thank you! >> >> -- >> >> >> *Att[]'sIury Gregory Melo Ferreira * >> *MSc in Computer Science at UFCG* >> *Part of the ironic-core and puppet-manager-core team in OpenStack* >> *Software Engineer at Red Hat Czech* >> *Social*: https://www.linkedin.com/in/iurygregory >> *E-mail: iurygregory at gmail.com * >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: From sshnaidm at redhat.com Mon Mar 7 14:38:49 2022 From: sshnaidm at redhat.com (Sagi Shnaidman) Date: Mon, 7 Mar 2022 16:38:49 +0200 Subject: [ansible openstack collections/modules] Core team clean up Message-ID: Hello, I took a look at our core team ansible-collections-openstack-core in Gerrit.[1] There are people that might be not participating in the development of Ansible Openstack collections anymore:[2] Jesse Pretorius Julia Kreger Kevin Carter Mohammed Naser Monty Taylor If there are no objections from them, I'd like to propose we remove these folks from our core team, while thanking them for their contributions. If anyone still wants to resume their work on collections or just want to help in reviews and merges, please let me know. Same if you are sure you don't plan to participate in collections development in the next future and would like to be removed from the list for now. Please reply and let me know any agreements or concerns. [1] https://review.opendev.org/admin/groups/0e01228e912733e8b9a8d957631e41665aa0ffbd,members [2] https://www.stackalytics.io/report/contribution?module=opendev.org/openstack/ansible-collections-openstack&project_type=openstack-others&days=180 -- Best regards Sagi Shnaidman -------------- next part -------------- An HTML attachment was scrubbed... URL: From andrew.west-contractor at CGG.COM Mon Mar 7 14:53:54 2022 From: andrew.west-contractor at CGG.COM (west, andrew) Date: Mon, 7 Mar 2022 14:53:54 +0000 Subject: [openstack-ansible] bootstap-aio.sh is failing with error :jinja2.exceptions.TemplateRuntimeError: No filter named 'ipaddr' found Message-ID: Am trying to install openstack-ansible AIO (24.0.0.0rc1) on a recent ubuntu focal (20.04) cloud image VM kernel 5.4.0-99-generic ansible core 2.12.2 jinja 3.0.3 python 3.8.10 the bootstap-aio.sh script is failing with error jinja2.exceptions.TemplateRuntimeError: No filter named 'ipaddr' found (I see the ipaddr filter moved (big changes separating modules > ansible2.10 I think) to the ansible.netcommon collection So following google: ansible galaxy collection install ansible.netcommon Also needs pip install netaddr However this still doesn't solve it, add traceback to playbook: Seems some of core ansible_collections plugins need updating ? (sorry, I'm fairly new to ansible) EXEC /bin/sh -c 'echo ~root && sleep 0' EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1646414479.2525613-38683-223155148734113 `" && echo ansible-tmp-1646414479.2525613-38683-223155148734113="` echo /root/.ansible/tmp/ansible-tmp-1646414479.2525613-38683-223155148734113 `" ) && sleep 0' redirecting filter ansible.builtin.ipaddr to ansible.netcommon.ipaddr redirecting filter ansible.builtin.ipaddr to ansible.netcommon.ipaddr redirecting filter ansible.builtin.ipaddr to ansible.netcommon.ipaddr EXEC /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1646414479.2525613-38683-223155148734113/ > /dev/null 2>&1 && sleep 0' The full traceback is: Traceback (most recent call last): File "/opt/ansible-runtime/lib/python3.8/site-packages/ansible/executor/task_executor.py", line 119, in run item_results = self._run_loop(items) File "/opt/ansible-runtime/lib/python3.8/site-packages/ansible/executor/task_executor.py", line 354, in _run_loop res = self._execute(variables=task_vars) File "/opt/ansible-runtime/lib/python3.8/site-packages/ansible/executor/task_executor.py", line 589, in _execute result = self._handler.run(task_vars=variables) File "/etc/ansible/ansible_collections/openstack/config_template/plugins/action/config_template.py", line 869, in run resultant = self._check_templar(data=template_data) File "/etc/ansible/ansible_collections/openstack/config_template/plugins/action/config_template.py", line 768, in _check_templar return self._templar.template( File "/opt/ansible-runtime/lib/python3.8/site-packages/ansible/template/__init__.py", line 869, in template result = self.do_template( File "/opt/ansible-runtime/lib/python3.8/site-packages/ansible/template/__init__.py", line 1139, in do_template res = j2_concat(rf) File "