Bug 1647227
| Summary: | problems in manifest | ||
|---|---|---|---|
| Product: | Container Native Virtualization (CNV) | Reporter: | Steve Reichard <sreichar> |
| Component: | Storage | Assignee: | Adam Litke <alitke> |
| Status: | CLOSED CURRENTRELEASE | QA Contact: | shiyang.wang <shiywang> |
| Severity: | high | Docs Contact: | |
| Priority: | high | ||
| Version: | 1.3 | CC: | cnv-qe-bugs, fsimonce, lbednar, sgordon, vkhanna |
| Target Milestone: | --- | ||
| Target Release: | 1.3 | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | 1.3.0-4 | Doc Type: | If docs needed, set a value |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2019-01-08 14:27:03 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
Confirmed that if I modified the manifest, I am able to deploy sucessfully: [root@spr-master ~]# sed -e 's#registry.access.redhat.com/cnv13#cnv#' /usr/share/kubevirt-cdi/manifests/cdi-controller.yaml > cdi-controller.yaml [root@spr-master ~]# oc delete -f /usr/share/kubevirt-cdi/manifests/cdi-controller.yaml serviceaccount "cdi-sa" deleted clusterrole.rbac.authorization.k8s.io "cdi" deleted clusterrolebinding.rbac.authorization.k8s.io "cdi-sa" deleted deployment.apps "cdi-deployment" deleted customresourcedefinition.apiextensions.k8s.io "datavolumes.cdi.kubevirt.io" deleted serviceaccount "cdi-apiserver" deleted rolebinding.rbac.authorization.k8s.io "cdi-apiserver" deleted role.rbac.authorization.k8s.io "cdi-apiserver" deleted clusterrolebinding.rbac.authorization.k8s.io "cdi-apiserver" deleted clusterrole.rbac.authorization.k8s.io "cdi-apiserver" deleted clusterrolebinding.rbac.authorization.k8s.io "cdi-apiserver-auth-delegator" deleted service "cdi-api" deleted deployment.apps "cdi-api" deleted service "cdi-uploadproxy" deleted deployment.apps "cdi-uploadproxy" deleted [root@spr-master ~]# oc get pods NAME READY STATUS RESTARTS AGE master-api-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 3 6d master-controllers-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 2 6d master-etcd-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 2 6d virt-api-bc78946cb-6ndpm 1/1 Running 0 19m virt-api-bc78946cb-qsgq4 1/1 Running 0 19m virt-controller-794776c85f-drxfq 1/1 Running 0 19m virt-controller-794776c85f-pq69k 1/1 Running 0 19m virt-handler-2ntdf 1/1 Running 0 19m virt-handler-lw4gl 1/1 Running 0 19m virt-handler-m7pvf 1/1 Running 0 19m [root@spr-master ~]# oc apply -f cdi-controller.yaml serviceaccount/cdi-sa created clusterrole.rbac.authorization.k8s.io/cdi created clusterrolebinding.rbac.authorization.k8s.io/cdi-sa created deployment.apps/cdi-deployment created customresourcedefinition.apiextensions.k8s.io/datavolumes.cdi.kubevirt.io created serviceaccount/cdi-apiserver created rolebinding.rbac.authorization.k8s.io/cdi-apiserver created role.rbac.authorization.k8s.io/cdi-apiserver created clusterrolebinding.rbac.authorization.k8s.io/cdi-apiserver created clusterrole.rbac.authorization.k8s.io/cdi-apiserver created clusterrolebinding.rbac.authorization.k8s.io/cdi-apiserver-auth-delegator created service/cdi-api created deployment.apps/cdi-api created service/cdi-uploadproxy created deployment.apps/cdi-uploadproxy created [root@spr-master ~]# oc get pods NAME READY STATUS RESTARTS AGE cdi-api-556469bfbc-jf5mz 1/1 Running 0 45s cdi-deployment-6c49474554-5q2n2 1/1 Running 0 45s cdi-uploadproxy-784878c6cc-6t8kw 1/1 Running 0 44s master-api-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 3 6d master-controllers-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 2 6d master-etcd-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 2 6d virt-api-bc78946cb-6ndpm 1/1 Running 0 20m virt-api-bc78946cb-qsgq4 1/1 Running 0 20m virt-controller-794776c85f-drxfq 1/1 Running 0 20m virt-controller-794776c85f-pq69k 1/1 Running 0 20m virt-handler-2ntdf 1/1 Running 0 20m virt-handler-lw4gl 1/1 Running 0 20m virt-handler-m7pvf 1/1 Running 0 20m [root@spr-master ~]# Updated d/s spec file to fix the environment variables. Please check the latest build: https://brewweb.engineering.redhat.com/brew/buildinfo?buildID=796662 *** Bug 1647217 has been marked as a duplicate of this bug. *** verified with kubevirt-cdi-manifests-1.3.0-4.baac3e8.noarch.rpm |
Description of problem: I attempted to use the build from today and I believe there are a could issues with the manifest. I was my understanding that the registy was not going to be specified. Also, the domain is cnv13-tech-preview, I believe this is supposed to be cnv-tech-preview. [root@spr-master ~]# oc apply -f /usr/share/kubevirt-cdi/manifests/cdi-controller.yaml serviceaccount/cdi-sa created clusterrole.rbac.authorization.k8s.io/cdi created clusterrolebinding.rbac.authorization.k8s.io/cdi-sa created deployment.apps/cdi-deployment created customresourcedefinition.apiextensions.k8s.io/datavolumes.cdi.kubevirt.io created serviceaccount/cdi-apiserver created rolebinding.rbac.authorization.k8s.io/cdi-apiserver created role.rbac.authorization.k8s.io/cdi-apiserver created clusterrolebinding.rbac.authorization.k8s.io/cdi-apiserver created clusterrole.rbac.authorization.k8s.io/cdi-apiserver created clusterrolebinding.rbac.authorization.k8s.io/cdi-apiserver-auth-delegator created service/cdi-api created deployment.apps/cdi-api created service/cdi-uploadproxy created deployment.apps/cdi-uploadproxy created [root@spr-master ~]# oc get pods NAME READY STATUS RESTARTS AGE cdi-api-68c644f4c6-bgvgn 0/1 ErrImagePull 0 9s cdi-deployment-7d85475676-bmmwq 0/1 ErrImagePull 0 9s cdi-uploadproxy-6dcf86d94-45zl7 0/1 ErrImagePull 0 9s master-api-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 3 6d master-controllers-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 2 6d master-etcd-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 2 6d virt-api-bc78946cb-6ndpm 1/1 Running 0 7m virt-api-bc78946cb-qsgq4 1/1 Running 0 7m virt-controller-794776c85f-drxfq 1/1 Running 0 7m virt-controller-794776c85f-pq69k 1/1 Running 0 7m virt-handler-2ntdf 1/1 Running 0 7m virt-handler-lw4gl 1/1 Running 0 7m virt-handler-m7pvf 1/1 Running 0 7m [root@spr-master ~]# oc get pods NAME READY STATUS RESTARTS AGE cdi-api-68c644f4c6-bgvgn 0/1 ImagePullBackOff 0 20s cdi-deployment-7d85475676-bmmwq 0/1 ImagePullBackOff 0 20s cdi-uploadproxy-6dcf86d94-45zl7 0/1 ImagePullBackOff 0 20s master-api-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 3 6d master-controllers-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 2 6d master-etcd-spr-master.cloud.lab.eng.bos.redhat.com 1/1 Running 2 6d virt-api-bc78946cb-6ndpm 1/1 Running 0 7m virt-api-bc78946cb-qsgq4 1/1 Running 0 7m virt-controller-794776c85f-drxfq 1/1 Running 0 7m virt-controller-794776c85f-pq69k 1/1 Running 0 7m virt-handler-2ntdf 1/1 Running 0 7m virt-handler-lw4gl 1/1 Running 0 7m virt-handler-m7pvf 1/1 Running 0 7m [root@spr-master ~]# yum list installed | grep cdi kubevirt-cdi-manifests.noarch 1.3.0-3.baac3e8 installed [root@spr-master ~]# rpm -ql kubevirt-cdi-manifests.noarch /usr/share/kubevirt-cdi/manifests /usr/share/kubevirt-cdi/manifests/cdi-controller.yaml [root@spr-master ~]# grep -i cnv /usr/share/kubevirt-cdi/manifests/cdi-controller.yaml image: registry.access.redhat.com/cnv13-tech-preview/virt-cdi-controller:v1.3.0 value: registry.access.redhat.com/cnv13-tech-preview/virt-cdi-importer:v1.3.0 value: registry.access.redhat.com/cnv13-tech-preview/virt-cdi-cloner:v1.3.0 value: registry.access.redhat.com/cnv13-tech-preview/virt-cdi-uploadserver:v1.3.0 image: registry.access.redhat.com/cnv13-tech-preview/virt-cdi-apiserver:v1.3.0 image: registry.access.redhat.com/cnv13-tech-preview/virt-cdi-uploadproxy:v1.3.0 [root@spr-master ~]# Version-Release number of selected component (if applicable): 1.3.0-3.baac3e8 How reproducible: Only tried once. Steps to Reproduce: 1. 2. 3. Actual results: Expected results: Additional info: