owned this note changed 3 years ago
Published Linked with GitHub

Ruck Rover - 27th May 2022 - 02nd June 2022

tags: ruck_rover
Previous RR notes: https://hackmd.io/2hB-P772SqyqDs0KKZzZEQ?view

Cockpit

Downstream cockpit

Thursday 02 June

new/transient/no bug yet

Bugs:

Wednesday 01 June (and earlier ongoing things tracked here)

Bugs:


STOP (all tracked bugs duplicated above stop scrolling) STOP



STOP (all tracked bugs duplicated above stop scrolling) STOP


Tuesday 31 May

Bugs:


Promotions upstream handoff:

1 TRAIN

2 WALLABY 9 promoted

3 WALLABY/8

4 MASTER

5 VICTORIA

new/transient/no bug yet

Monday 30 May

Bugs:

  • https://bugs.launchpad.net/tripleo/+bug/1973223 Master Sc010-kvm job is failing on octavia related tempest test: octavia_tempest_plugin.tests.scenario.v2.test_traffic_ops.TrafficOperationsScenarioTest

  • https://bugs.launchpad.net/tripleo/+bug/1964940 Compute tests are failing with failed to reach ACTIVE status and task state "None" within the required time.

  • https://bugs.launchpad.net/tripleo/+bug/1972163 cinder tempest.api.compute.admin.test_volumes_negative* tempest tests failing randomly in multiple branches.

  • https://bugs.launchpad.net/bugs/1971465 fs001 and fs035 OVB jobs failing tempest - identity/haproxy connection errors

  • https://bugzilla.redhat.com/show_bug.cgi?id=2089304 fs020 and full-tempest-scenario job failing on tempest test neutron_tempest_plugin.scenario.test_trunk.TrunkTest.test_trunk_subport_lifecycle (edit)

  • https://bugs.launchpad.net/tripleo/+bug/1973568 Master Scenario002 is failing on Barbican related tempest test - tempest.lib.exceptions.UnexpectedResponseCode: Unexpected response code received , Details: 503

  • https://bugzilla.redhat.com/show_bug.cgi?id=2091502 ERROR: Cannot install stackviz because these package versions have conflicting dependencies

  • https://bugs.launchpad.net/tripleo/+bug/1976251 [CI] tox-ansible-test-sanity doesn't take the "ignore" anymore

  • https://bugs.launchpad.net/tripleo/+bug/1975917 AttributeError: 'Service' object has no attribute 'enabled'

  • https://code.engineering.redhat.com/gerrit/c/networking-ovn/+/411213

    ​​​​<bhagyashris> slaweq, hey can you help us to merge this one  https://code.engineering.redhat.com/gerrit/c/networking-ovn/+/411213
    ​​​​<bhagyashris> we are blocked due to this ^
    ​​​​<slaweq> bhagyashris: sure, looking
    ​​​​<slaweq> bhagyashris: done
    ​​​​<bhagyashris> slaweq, thanks
    
  • regarding the curl error:

    • pinged migarcia
      ​​​​​​​<bhagyashris> migarcia, hey
      ​​​​​​​ <bhagyashris> around?
      ​​​​​​​ <migarcia> bhagyashris: I am, what's up?
      ​​​​​​​ <bhagyashris> we are facing one issue build push upload image 
      ​​​​​​​ <bhagyashris> https://sf.hosted.upshift.rdu2.redhat.com/logs/openstack-periodic-integration-rhos-17-rhel9/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-build-containers-ubi-9-internal-rhel-9-build-push-upload-rhos-17/1c70911/logs/container-builds/d200f2a7-c38e-49ad-b9ef-e83cabfa5fc0/base/base-build.log
      ​​​​​​​ <bhagyashris> lon is out today 
      ​​​​​​​ <bhagyashris> issue is: Errors during downloading metadata for repository 'osptrunk-candidate-deps':
      ​​​​​​​ <bhagyashris>   - Curl error (28): Timeout was reached for http://download.eng.bos.redhat.com/brewroot/repos/rhos-17.0-rhel-9-trunk-candidate/latest/x86_64/repodata/f6d120c5ebe86676cd598a6c179f7bef99e4aa3fc54f9291e27708b502d1f7fc-primary.xml.gz
      ​​​​​​​ <migarcia> bhagyashris: I would rekick, could be a network blip or that the repo was regenerated while the job was running
      ​​​​​​​ <migarcia> rhos-17.0-rhel-9-trunk-candidate/latest symlink gets updated regularly as new builds are tagged in
      ​​​​​​​ <bhagyashris> migarcia, ack thanks ! will check the result in the rekick 
      ​​​​​​​ <migarcia> cool, let me know
      ​​​​​​​ <bhagyashris> migarcia, thanks 
      ​​​​​​​ <bhagyashris> let me know once you re kicked 
      ​​​​​​​ <bhagyashris> migarcia, hey you are rekicking or should i rekicked 
      ​​​​​​​ <migarcia> bhagyashris: please do
      ​​​​​​​ <bhagyashris> migarcia, ack
      ​​​​​​​ <bhagyashris> migarcia, hey we are still with same issue on the recent run 
      ​​​​​​​ <bhagyashris>  https://sf.hosted.upshift.rdu2.redhat.com/logs/94/947e8a93a865e16481d14a1dd9fe1f91216e1a8d/openstack-periodic-integration-rhos-17-rhel9/periodic-tripleo-build-containers-ubi-9-internal-rhel-9-build-push-upload-rhos-17/45338ea/logs/container-builds/a3e81dbf-42ba-42e9-bf72-d5aeb0e65b4f/base/base-build.log
      ​​​​​​​ <migarcia> bhagyashris: huh, I can download that file just fine.
      ​​​​​​​ <migarcia> and it looks like the job was also downloading it fine, but very slow for some reason
      ​​​​​​​ <migarcia> osptrunk-candidate-deps                         506  B/s |  64 kB     02:10
      ​​​​​​​ <ysandeep> bhagyashris: could you hold a node and check mtu on cni-podman bridge
      
      ​​​​​​​ <ysandeep> bhagyashris, sounds similiar to https://bugzilla.redhat.com/show_bug.cgi?id=2060932 
      
      ​​​​​​​ <bhagyashris> ysandeep, let me hit the testproject patch
      
      Update: recently podman version updated from 2:4.0.2-6.el9_0 to 2:4.0.2-7.el9_0 which basically includes the new dependencies "netavark" and that creates the "podman0" bridget from here https://github.com/containers/netavark/blob/02e031fdd9f7cd849c4fdd18cdd1ecb1a135485f/src/test/config/setupopts2.test.json#L14-L22 and takes mtu value as 1500 which basically take more time to download metadata for repository 'osptrunk-candidate-deps' and get timed and failed. Will debug more tomorrow on that and will file bug if required

new/transient/no bug yet

Friday 27 May

Bugs:

<bhagyashris> #rhos-ops Hi, we are faing the node_failure on rhos17-rhel9
<bhagyashris> facing*
<bhagyashris> currently running integration line - rhos17 on rhel9
<bhagyashris> https://sf.hosted.upshift.rdu2.redhat.com/zuul/t/tripleo-ci-internal/status
<bhjf> Title: Zuul (at sf.hosted.upshift.rdu2.redhat.com)
<bhagyashris> psedlak|ruck, ^

<bhagyashris> dpawlik, ^
<bhagyashris> facing node failure on downstream 

<dpawlik> kforde: hey, all is fine with the infra?
<dpawlik> kforde: ah, just horizon does not work. Was thinking that something happend with one vm

<dpawlik> bhagyashris: can I deque your job and recheck?
<bhagyashris> yeah

<bhagyashris> dpawlik, yeah 

<dpawlik> bhagyashris: "Global Service Outage Ongoing: RDU2 DC Impact"
<dpawlik> it can be related
<bhagyashris> dpawlik, ack
<dpawlik> we got network flappings between services like zookeeper, DNS does not work...

Thursday 26

(previous ruck|rover pad: https://hackmd.io/uiv6iiN5QR-Z3mfFyKWeqA)

Select a repo