Bug 2210378

Summary: Cannot create external storage when using OVNKubernetes
Product: [Red Hat Storage] Red Hat OpenShift Data Foundation Reporter: joherr
Component: odf-operatorAssignee: Nitin Goyal <nigoyal>
Status: CLOSED NOTABUG QA Contact: Elad <ebenahar>
Severity: unspecified Docs Contact:
Priority: unspecified    
Version: 4.12CC: bkunal, ebenahar, muagarwa, odf-bz-bot, vfarias
Target Milestone: ---   
Target Release: ---   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2023-08-08 08:50:15 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 2186659    

Description joherr 2023-05-26 23:04:21 UTC
Description of problem (please be detailed as possible and provide log
snippests):
When trying to create a storage system using the ODF Operator and the OVNKubernetes network stack, the ODF pods cannot reach the external ceph cluster.


Version of all relevant components (if applicable):
OpenShift 4.13.0
OpenShift Data Foundation 4.12.3-rhodf


Does this issue impact your ability to continue to work with the product
(please explain in detail what is the user impact)?
It does not impact because there is a solution


Is there any workaround available to the best of your knowledge?
Enabling routingViaHost in the clusters network operator allows pods to have egress traffic. This fixes the issue.
It would be great to have a note in the ODF Operators documentation when connecting to an external ceph cluster that the patch is needed.

Patch:
oc patch network.operator cluster -p '{"spec": {"defaultNetwork": {"ovnKubernetesConfig": {"gatewayConfig": {"routingViaHost": true}}}}}' --type merge


Rate from 1 - 5 the complexity of the scenario you performed that caused this
bug (1 - very simple, 5 - very complex)?
1

Can this issue reproducible?
yes

Can this issue reproduce from the UI?
yes

If this is a regression, please provide more details to justify this:


Steps to Reproduce:
1.Install NMNState and configure a second network for storage traffic
2.install the ODF operator
3.Create an external storage system that connects to an existing Ceph cluster.


Actual results:
External storage system never completes. Looking at logs shows pods cannot communicate to the external cluster even though the cluster nodes can.

Expected results:
A usable storage system to be created.


Additional info:
This issue does not occur with the OpenShiftSDN network stack.

Comment 2 Nitin Goyal 2023-05-29 04:46:39 UTC
I think this should be a documentation bug.

Comment 9 joherr 2023-08-07 15:28:06 UTC
I did already open a KCS article on this.

https://access.redhat.com/solutions/7017082

The title is specific to "Cannot create an external Ceph storage system in Red Hat OpenShift Data Foundation when using OVNKubernetes".

But I do comment in the KCS the following:
~~~
The same issue is seen when creating a backend for the Trident Operator when trying to connect to a NetApp device.
The backend cannot get created because the pods used for the Operator cannot communicate with the NetApp device.

This also resolves an issue when trying to ssh into virtual machines when using the Red Hat OpenShift Virtualization Operator and the OVNKubernetes network stack.
~~~

I can edit or create another if desired. Please let me know.

Comment 10 Nitin Goyal 2023-08-08 08:50:15 UTC
Closing this bug as KCS is already created and no further help is required from the engineering team. The other discussion about the KCS can still happen on it.