all: children: zuul_unreachable: hosts: {} hosts: controller: ansible_connection: ssh ansible_host: 199.204.45.95 ansible_port: 22 ansible_python_interpreter: auto ansible_user: zuul nodepool: az: nova cloud: ansible-vexxhost external_id: fabb1c08-e604-4673-ac95-fabf0e63b3b7 host_id: 6b8d58a114f50c983ebf6931b748184c76907eec51ac86ada773f491 interface_ip: 199.204.45.95 label: ansible-fedora-37-1vcpu private_ipv4: 192.168.0.55 private_ipv6: null provider: ansible-vexxhost-ca-ymq-1 public_ipv4: 199.204.45.95 public_ipv6: 2604:e100:1:0:f816:3eff:feaf:e00 region: ca-ymq-1 slot: null zuul_use_fetch_output: true vars: zuul: _inheritance_path: - '' - '' - '' ansible_version: '8' artifacts: - branch: main change: '1034' job: build-ansible-collection metadata: type: zuul_manifest name: Zuul Manifest patchset: 0abf97a588d34d60db823b14a653479c63e480fd project: ansible-collections/kubernetes.core url: https://be606fbd26f408804df8-ba59a7ac35915695a373252d89ec730f.ssl.cf1.rackcdn.com/ansible/f981b872d79d4ff0b8836aae284981a8/zuul-manifest.json - branch: main change: '1034' job: build-ansible-collection metadata: type: ansible_collection version: 7.0.0 name: kubernetes.core patchset: 0abf97a588d34d60db823b14a653479c63e480fd project: ansible-collections/kubernetes.core url: https://be606fbd26f408804df8-ba59a7ac35915695a373252d89ec730f.ssl.cf1.rackcdn.com/ansible/f981b872d79d4ff0b8836aae284981a8/artifacts/kubernetes-core-7.0.0.tar.gz attempts: 1 branch: main build: ef2ea97eeda94cdbb68de8b2d1b8d8a9 build_refs: - branch: main change: '1034' change_message: "Support take_ownership parameter in helm installation\n\n##### SUMMARY\r\nAdds support for the take_ownership for initial release installation operations.\r\n\r\n##### ISSUE TYPE\r\n- Bugfix Pull Request\r\n\r\n##### COMPONENT NAME\r\nplugins/modules/helm.py\r\n\r\n##### ADDITIONAL INFORMATION\r\nI recently had to migrate a namespace k8s from flat manifest installation into helm a release. \r\nI was so glad to see the `take_ownership` feature but realized that it work only after first installation of the release.\r\nSeeing no reason to denied this use case i suggest this very simple changes.\r\n\r\nTo reproduce it:\r\n1) Create a new namespace in any cluster.\r\n2) Create a secret\r\n3) Install any helm chart that deploy the same secret using `take_ownership: true`.\r\n" change_url: https://github.com/ansible-collections/kubernetes.core/pull/1034 commit_id: 0abf97a588d34d60db823b14a653479c63e480fd patchset: 0abf97a588d34d60db823b14a653479c63e480fd project: canonical_hostname: github.com canonical_name: github.com/ansible-collections/kubernetes.core name: ansible-collections/kubernetes.core short_name: kubernetes.core src_dir: src/github.com/ansible-collections/kubernetes.core topic: null buildset: 2ce18ee7ae4547519059d42878663a9e buildset_refs: - branch: main change: '1034' change_message: "Support take_ownership parameter in helm installation\n\n##### SUMMARY\r\nAdds support for the take_ownership for initial release installation operations.\r\n\r\n##### ISSUE TYPE\r\n- Bugfix Pull Request\r\n\r\n##### COMPONENT NAME\r\nplugins/modules/helm.py\r\n\r\n##### ADDITIONAL INFORMATION\r\nI recently had to migrate a namespace k8s from flat manifest installation into helm a release. \r\nI was so glad to see the `take_ownership` feature but realized that it work only after first installation of the release.\r\nSeeing no reason to denied this use case i suggest this very simple changes.\r\n\r\nTo reproduce it:\r\n1) Create a new namespace in any cluster.\r\n2) Create a secret\r\n3) Install any helm chart that deploy the same secret using `take_ownership: true`.\r\n" change_url: https://github.com/ansible-collections/kubernetes.core/pull/1034 commit_id: 0abf97a588d34d60db823b14a653479c63e480fd patchset: 0abf97a588d34d60db823b14a653479c63e480fd project: canonical_hostname: github.com canonical_name: github.com/ansible-collections/kubernetes.core name: ansible-collections/kubernetes.core short_name: kubernetes.core src_dir: src/github.com/ansible-collections/kubernetes.core topic: null change: '1034' change_message: "Support take_ownership parameter in helm installation\n\n##### SUMMARY\r\nAdds support for the take_ownership for initial release installation operations.\r\n\r\n##### ISSUE TYPE\r\n- Bugfix Pull Request\r\n\r\n##### COMPONENT NAME\r\nplugins/modules/helm.py\r\n\r\n##### ADDITIONAL INFORMATION\r\nI recently had to migrate a namespace k8s from flat manifest installation into helm a release. \r\nI was so glad to see the `take_ownership` feature but realized that it work only after first installation of the release.\r\nSeeing no reason to denied this use case i suggest this very simple changes.\r\n\r\nTo reproduce it:\r\n1) Create a new namespace in any cluster.\r\n2) Create a secret\r\n3) Install any helm chart that deploy the same secret using `take_ownership: true`.\r\n" change_url: https://github.com/ansible-collections/kubernetes.core/pull/1034 child_jobs: [] commit_id: 0abf97a588d34d60db823b14a653479c63e480fd event_id: 99fab8b6-0d97-11f1-879d-f9cb395be0c9 executor: hostname: ze01.softwarefactory-project.io inventory_file: /var/lib/zuul/builds/ef2ea97eeda94cdbb68de8b2d1b8d8a9/ansible/inventory.yaml log_root: /var/lib/zuul/builds/ef2ea97eeda94cdbb68de8b2d1b8d8a9/work/logs result_data_file: /var/lib/zuul/builds/ef2ea97eeda94cdbb68de8b2d1b8d8a9/work/results.json src_root: /var/lib/zuul/builds/ef2ea97eeda94cdbb68de8b2d1b8d8a9/work/src work_root: /var/lib/zuul/builds/ef2ea97eeda94cdbb68de8b2d1b8d8a9/work items: - branch: main change: '1034' change_message: "Support take_ownership parameter in helm installation\n\n##### SUMMARY\r\nAdds support for the take_ownership for initial release installation operations.\r\n\r\n##### ISSUE TYPE\r\n- Bugfix Pull Request\r\n\r\n##### COMPONENT NAME\r\nplugins/modules/helm.py\r\n\r\n##### ADDITIONAL INFORMATION\r\nI recently had to migrate a namespace k8s from flat manifest installation into helm a release. \r\nI was so glad to see the `take_ownership` feature but realized that it work only after first installation of the release.\r\nSeeing no reason to denied this use case i suggest this very simple changes.\r\n\r\nTo reproduce it:\r\n1) Create a new namespace in any cluster.\r\n2) Create a secret\r\n3) Install any helm chart that deploy the same secret using `take_ownership: true`.\r\n" change_url: https://github.com/ansible-collections/kubernetes.core/pull/1034 commit_id: 0abf97a588d34d60db823b14a653479c63e480fd patchset: 0abf97a588d34d60db823b14a653479c63e480fd project: canonical_hostname: github.com canonical_name: github.com/ansible-collections/kubernetes.core name: ansible-collections/kubernetes.core short_name: kubernetes.core src_dir: src/github.com/ansible-collections/kubernetes.core topic: null job: ansible-galaxy-importer jobtags: [] max_attempts: 3 message: U3VwcG9ydCB0YWtlX293bmVyc2hpcCBwYXJhbWV0ZXIgaW4gaGVsbSBpbnN0YWxsYXRpb24KCiMjIyMjIFNVTU1BUlkNCkFkZHMgc3VwcG9ydCBmb3IgdGhlIHRha2Vfb3duZXJzaGlwIGZvciBpbml0aWFsIHJlbGVhc2UgaW5zdGFsbGF0aW9uIG9wZXJhdGlvbnMuDQoNCiMjIyMjIElTU1VFIFRZUEUNCi0gQnVnZml4IFB1bGwgUmVxdWVzdA0KDQojIyMjIyBDT01QT05FTlQgTkFNRQ0KcGx1Z2lucy9tb2R1bGVzL2hlbG0ucHkNCg0KIyMjIyMgQURESVRJT05BTCBJTkZPUk1BVElPTg0KSSByZWNlbnRseSBoYWQgdG8gbWlncmF0ZSBhIG5hbWVzcGFjZSBrOHMgZnJvbSBmbGF0IG1hbmlmZXN0IGluc3RhbGxhdGlvbiBpbnRvIGhlbG0gYSByZWxlYXNlLiANCkkgd2FzIHNvIGdsYWQgdG8gc2VlIHRoZSBgdGFrZV9vd25lcnNoaXBgIGZlYXR1cmUgYnV0IHJlYWxpemVkIHRoYXQgaXQgd29yayBvbmx5IGFmdGVyIGZpcnN0IGluc3RhbGxhdGlvbiBvZiB0aGUgcmVsZWFzZS4NClNlZWluZyBubyByZWFzb24gdG8gZGVuaWVkIHRoaXMgdXNlIGNhc2UgaSBzdWdnZXN0IHRoaXMgdmVyeSBzaW1wbGUgY2hhbmdlcy4NCg0KVG8gcmVwcm9kdWNlIGl0Og0KMSkgQ3JlYXRlIGEgbmV3IG5hbWVzcGFjZSBpbiBhbnkgY2x1c3Rlci4NCjIpIENyZWF0ZSBhIHNlY3JldA0KMykgSW5zdGFsbCBhbnkgaGVsbSBjaGFydCB0aGF0IGRlcGxveSB0aGUgc2FtZSBzZWNyZXQgdXNpbmcgYHRha2Vfb3duZXJzaGlwOiB0cnVlYC4NCg== patchset: 0abf97a588d34d60db823b14a653479c63e480fd pipeline: gate playbook_context: playbook_projects: trusted/project_0/github.com/ansible/zuul-config: canonical_name: github.com/ansible/zuul-config checkout: master commit: bd3b9c37c47d555493fc4feb3ab88d70887ad6fd trusted/project_1/opendev.org/zuul/zuul-jobs: canonical_name: opendev.org/zuul/zuul-jobs checkout: master commit: 571c0efa3491d12ecb8fc1169c510716d55c0fc2 untrusted/project_0/github.com/ansible/ansible-zuul-jobs: canonical_name: github.com/ansible/ansible-zuul-jobs checkout: master commit: 192320b9d41936ac6065fcaf6e286bf4dca783a5 untrusted/project_1/github.com/ansible/zuul-config: canonical_name: github.com/ansible/zuul-config checkout: master commit: bd3b9c37c47d555493fc4feb3ab88d70887ad6fd untrusted/project_2/opendev.org/zuul/zuul-jobs: canonical_name: opendev.org/zuul/zuul-jobs checkout: master commit: 571c0efa3491d12ecb8fc1169c510716d55c0fc2 playbooks: - path: untrusted/project_0/github.com/ansible/ansible-zuul-jobs/playbooks/ansible-galaxy-importer/run.yaml roles: - checkout: master checkout_description: playbook branch link_name: ansible/playbook_0/role_0/zuul-jobs link_target: untrusted/project_0/github.com/ansible/ansible-zuul-jobs role_path: ansible/playbook_0/role_0/zuul-jobs/roles - checkout: master checkout_description: project default branch link_name: ansible/playbook_0/role_1/zuul-config link_target: untrusted/project_1/github.com/ansible/zuul-config role_path: ansible/playbook_0/role_1/zuul-config/roles - checkout: master checkout_description: project default branch link_name: ansible/playbook_0/role_2/zuul-jobs link_target: untrusted/project_2/opendev.org/zuul/zuul-jobs role_path: ansible/playbook_0/role_2/zuul-jobs/roles post_review: false project: canonical_hostname: github.com canonical_name: github.com/ansible-collections/kubernetes.core name: ansible-collections/kubernetes.core short_name: kubernetes.core src_dir: src/github.com/ansible-collections/kubernetes.core projects: github.com/ansible-collections/kubernetes.core: canonical_hostname: github.com canonical_name: github.com/ansible-collections/kubernetes.core checkout: main checkout_description: zuul branch commit: ec88e75f41cb885c329102fecf57396da4b50e2c name: ansible-collections/kubernetes.core required: false short_name: kubernetes.core src_dir: src/github.com/ansible-collections/kubernetes.core github.com/ansible-network/releases: canonical_hostname: github.com canonical_name: github.com/ansible-network/releases checkout: master checkout_description: project default branch commit: 646b310655c531e4904be07f4ff8fc3a29addd09 name: ansible-network/releases required: true short_name: releases src_dir: src/github.com/ansible-network/releases ref: refs/pull/1034/head resources: {} tenant: ansible timeout: 1800 topic: null voting: true zuul_use_fetch_output: true