Bug 2175612 - noobaa-core-0 crashing and storagecluster not getting to ready state during ODF deployment with FIPS enabled in 4.13cluster
Summary: noobaa-core-0 crashing and storagecluster not getting to ready state during O...
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat OpenShift Data Foundation
Classification: Red Hat Storage
Component: Multi-Cloud Object Gateway
Version: 4.13
Hardware: ppc64le
OS: Unspecified
unspecified
medium
Target Milestone: ---
: ODF 4.13.0
Assignee: Liran Mauda
QA Contact: Sagi Hirshfeld
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2023-03-06 06:32 UTC by narayanspg
Modified: 2023-08-09 16:49 UTC (History)
14 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Cause: isa-l.gyp sources were not handled properly across the different archs supported Fix: Fix fetching and handling the isa-l.gyp sources to be as expected in all archs
Clone Of:
Environment:
Last Closed: 2023-06-21 15:24:25 UTC
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Github noobaa noobaa-core pull 7243 0 None Merged Fix isa-l.gyp to source the common sources explicitly in all arches 2023-04-10 06:09:48 UTC
Github noobaa noobaa-core pull 7251 0 None Merged [Backport into 5.13] backporting some fixes into 5.13 2023-04-10 06:09:48 UTC
Red Hat Product Errata RHBA-2023:3742 0 None None None 2023-06-21 15:24:52 UTC

Description narayanspg 2023-03-06 06:32:59 UTC
Description of problem (please be detailed as possible and provide log
snippests):

Created OCP cluster with FIPS enabled. Deployed ODF through UI and noobaa-core-0 is getting into crashloopbackoff error.

When we are create OCP cluster with FIPS disabled then ODF deployment is getting successful. 


Version of all relevant components (if applicable):
[root@nara5-cicd-odf-913b-syd05-bastion-0 ~]# oc get clusterversion
NAME      VERSION                                      AVAILABLE   PROGRESSING   SINCE   STATUS
version   4.13.0-0.nightly-ppc64le-2023-02-17-084453   True        False         3h49m   Cluster version is 4.13.0-0.nightly-ppc64le-2023-02-17-084453
[root@nara5-cicd-odf-913b-syd05-bastion-0 ~]# oc get storagecluster
NAME                 AGE   PHASE         EXTERNAL   CREATED AT             VERSION
ocs-storagecluster   61m   Progressing              2023-03-06T05:29:21Z   4.13.0
[root@nara5-cicd-odf-913b-syd05-bastion-0 ~]# oc describe csv odf-operator.v4.13.0 -n openshift-storage | grep full
Labels:       full_version=4.13.0-96

Does this issue impact your ability to continue to work with the product
(please explain in detail what is the user impact)?
With FIPS enabled we are not able to get ODF working

Is there any workaround available to the best of your knowledge?
NA

Rate from 1 - 5 the complexity of the scenario you performed that caused this
bug (1 - very simple, 5 - very complex)?


Can this issue reproducible?
Yes

Can this issue reproduce from the UI?
yes

If this is a regression, please provide more details to justify this:


Steps to Reproduce:
1. create OCP cluster with FIPS enabled
2. deploy ODF
3. noobaa pod is not getting to running state and storagecluster is stuck in progressing state.


Actual results:
ODF deployment is not successfull with FIPS enabled. 

Expected results:
ODF should be deployed successfully with FIPS enabled. 

Additional info:

Below are the noobaa logs:

[root@nara5-cicd-odf-913b-syd05-bastion-0 ~]# oc logs -f noobaa-core-0
Version is: 5.13.0-ccc7ff9
calling noobaa_init.sh
Running /usr/local/bin/node src/upgrade/upgrade_manager.js --upgrade_scripts_dir /root/node_modules/noobaa-core/src/upgrade/upgrade_scripts
OpenSSL 3.0.1 14 Dec 2021 setting up
init_rand_seed: starting ...
read_rand_seed: opening /dev/urandom ...
Mar-6 5:34:32.562 [/20]   [LOG] CONSOLE:: load_config_local: LOADED { DEFAULT_ACCOUNT_PREFERENCES: { ui_theme: 'LIGHT' }, REMOTE_NOOAA_NAMESPACE: 'openshift-storage', ALLOW_BUCKET_CREATE_ON_INTERNAL: false }
Mar-6 5:34:37.628 [/20]    [L0] core.rpc.rpc:: RPC register_n2n_proxy
Mar-6 5:34:37.891 [/20]   [LOG] CONSOLE:: loading .env file...
Mar-6 5:34:37.896 [/20]   [LOG] CONSOLE:: detect_fips_mode: found /proc/sys/crypto/fips_enabled with value 1
/usr/local/bin/node: symbol lookup error: /root/node_modules/noobaa-core/build/Release/nb_native.node: undefined symbol: md5_ctx_mgr_init_base
upgrade_manager failed with exit code 127
noobaa_init.sh finished
noobaa_init failed with exit code 127. aborting
[root@nara5-cicd-odf-913b-syd05-bastion-0 ~]#

Comment 1 narayanspg 2023-03-06 06:59:49 UTC
must-gather logs are available at https://drive.google.com/file/d/12KJMx_Q9loBQif6dxAh7gtd6i-RYkr_F/view?usp=share_link

Comment 3 narayanspg 2023-03-13 06:54:11 UTC
Hi ,

I have been asked needinfo but am not able to see them as comments are private. 
please let me know the required info.

Thanks,
Narayan

Comment 4 Liran Mauda 2023-03-14 12:02:56 UTC
Updated in DM.

Comment 5 narayanspg 2023-03-16 01:58:15 UTC
Hi Liran,

tested on ODF build 4.13.0-90 and  4.13.0-92. both build environments have same issue(noobaa-core-0 crashing and storagecluster not getting to ready state).

have retained cluster with build version 4.13.0-92 if you would like to access it. also have must gather logs for build 4.13.0-90 build.

Comment 10 narayanspg 2023-03-21 06:42:52 UTC
Hi Liran,

tested on build 4.13.0-89 and still see the noobaa-core crashing due to which storagecluster is not getting to ready state.

shared the cluster details over IM if you would like verify.

[root@fips3-cicd-odf-f92c-sao01-bastion-0 ~]# oc describe csv odf-operator.v4.13.0 -n openshift-storage | grep full
Labels:       full_version=4.13.0-89
          f:full_version:
[root@fips3-cicd-odf-f92c-sao01-bastion-0 ~]# oc get pods
NAME                                                              READY   STATUS                 RESTARTS        AGE
csi-addons-controller-manager-78f5bcc9fb-lw7jj                    2/2     Running                0               43m
csi-cephfsplugin-6kqpk                                            2/2     Running                0               25m
csi-cephfsplugin-hq84z                                            2/2     Running                0               25m
csi-cephfsplugin-jrd4r                                            2/2     Running                0               25m
csi-cephfsplugin-provisioner-68b9b6dd87-99cwv                     5/5     Running                0               25m
csi-cephfsplugin-provisioner-68b9b6dd87-hd44m                     5/5     Running                0               25m
csi-rbdplugin-hcph7                                               3/3     Running                0               25m
csi-rbdplugin-provisioner-5d56c8bc84-6xwl4                        6/6     Running                0               25m
csi-rbdplugin-provisioner-5d56c8bc84-8cslz                        6/6     Running                0               25m
csi-rbdplugin-vwm8n                                               3/3     Running                0               25m
csi-rbdplugin-zcsw9                                               3/3     Running                0               25m
noobaa-core-0                                                     0/1     CrashLoopBackOff       8 (4m49s ago)   22m
noobaa-db-pg-0                                                    1/1     Running                0               22m
noobaa-operator-645fdb94b8-gvs9n                                  1/1     Running                0               44m
ocs-metrics-exporter-57895c9ccf-trbbh                             1/1     Running                0               43m
ocs-operator-585f588f7b-zhr7t                                     1/1     Running                0               43m
odf-console-6d98c4c849-z889p                                      1/1     Running                0               44m
odf-operator-controller-manager-75b9bb8675-mhjl5                  2/2     Running                0               44m
rook-ceph-crashcollector-79d743de54eb92f2153e337679efd005-d8m4p   1/1     Running                0               22m
rook-ceph-crashcollector-89626277bd0a03dff0a95de9f03e3802-zcw4f   1/1     Running                0               23m
rook-ceph-crashcollector-e89289c97a67939b890cfbc685c93798-qjckh   1/1     Running                0               23m
rook-ceph-exporter-79d743de54eb92f2153e337679efd005-55f76bq7rzf   0/1     CreateContainerError   0               23m
rook-ceph-exporter-79d743de54eb92f2153e337679efd005-5c994cq89dd   0/1     CreateContainerError   0               22m
rook-ceph-exporter-89626277bd0a03dff0a95de9f03e3802-7499bb8t69j   0/1     CreateContainerError   0               23m
rook-ceph-exporter-89626277bd0a03dff0a95de9f03e3802-8568f9l5txk   0/1     CreateContainerError   0               23m
rook-ceph-exporter-e89289c97a67939b890cfbc685c93798-6ffb64qr72t   0/1     CreateContainerError   0               23m
rook-ceph-mds-ocs-storagecluster-cephfilesystem-a-687695d4m9m8q   2/2     Running                0               23m
rook-ceph-mds-ocs-storagecluster-cephfilesystem-b-5b9fbf69v8ccq   2/2     Running                0               22m
rook-ceph-mgr-a-5d445988bb-jnd4j                                  2/2     Running                0               23m
rook-ceph-mon-a-cd5f74545-dcx6s                                   2/2     Running                0               24m
rook-ceph-mon-b-54b46b7c68-wsm6t                                  2/2     Running                0               24m
rook-ceph-mon-c-97cfcf46b-6tcg4                                   2/2     Running                0               24m
rook-ceph-operator-5679dd6894-5ff5k                               1/1     Running                0               25m
rook-ceph-osd-0-86c679fb8c-grmgz                                  2/2     Running                0               23m
rook-ceph-osd-1-6996bf55c6-5pnh8                                  2/2     Running                0               23m
rook-ceph-osd-2-7d8656c8f7-pbd7j                                  2/2     Running                0               23m
rook-ceph-osd-prepare-3f5c0325f7bf6544a089bebc082f7032-wjrhq      0/1     Completed              0               23m
rook-ceph-osd-prepare-769ce41a3c76ea9bd790b0c372e65411-8thr8      0/1     Completed              0               23m
rook-ceph-osd-prepare-8ea6c20b536c1991820a71bbd9f05119-x6627      0/1     Completed              0               23m
rook-ceph-rgw-ocs-storagecluster-cephobjectstore-a-64556bdjw24t   2/2     Running                0               22m
[root@fips3-cicd-odf-f92c-sao01-bastion-0 ~]#

Comment 12 narayanspg 2023-03-30 06:08:55 UTC
Hi Liran, Let me know which build the fix will be available. Thanks.

Comment 17 narayanspg 2023-04-10 13:05:50 UTC
Tried on the latest build with fips enabled. storagecluster is not getting to ready state though all pods are up and running with reason - "Waiting on Nooba instance to finish initialization"

you can access the cluster with below details:
web_console_url = "https://console-openshift-console.apps.fips5-cicd-odf-e564.redhat.com"
kubeadm/opyKm-ysFdo-uFjDA-A3eaM

etc_hosts_entries = <<EOT

169.57.180.66 api.fips5-cicd-odf-e564.redhat.com console-openshift-console.apps.fips5-cicd-odf-e564.redhat.com integrated-oauth-server-openshift-authentication.apps.fips5-cicd-odf-e564.redhat.com oauth-openshift.apps.fips5-cicd-odf-e564.redhat.com prometheus-k8s-openshift-monitoring.apps.fips5-cicd-odf-e564.redhat.com grafana-openshift-monitoring.apps.fips5-cicd-odf-e564.redhat.com example.apps.fips5-cicd-odf-e564.redhat.com

EOT


[root@fips5-cicd-odf-e564-sao01-bastion-0 ~]# oc describe csv odf-operator.v4.13.0 -n openshift-storage | grep full
Labels:       full_version=4.13.0-130
          f:full_version:
[root@fips5-cicd-odf-e564-sao01-bastion-0 ~]#
[root@fips5-cicd-odf-e564-sao01-bastion-0 ~]#
[root@fips5-cicd-odf-e564-sao01-bastion-0 ~]# oc logs  noobaa-core-0 | grep fips
Apr-10 10:07:56.195 [/20]   [LOG] CONSOLE:: detect_fips_mode: found /proc/sys/crypto/fips_enabled with value 1
detect_fips_mode: found /proc/sys/crypto/fips_enabled with value 1
detect_fips_mode: found /proc/sys/crypto/fips_enabled with value 1
detect_fips_mode: found /proc/sys/crypto/fips_enabled with value 1
[root@fips5-cicd-odf-e564-sao01-bastion-0 ~]# oc get storagecluster
NAME                 AGE    PHASE         EXTERNAL   CREATED AT             VERSION
ocs-storagecluster   179m   Progressing              2023-04-10T10:03:25Z   4.13.0
[root@fips5-cicd-odf-e564-sao01-bastion-0 ~]# oc describe storagecluster ocs-storagecluster | grep Waiting
    Message:               Waiting on Nooba instance to finish initialization
[root@fips5-cicd-odf-e564-sao01-bastion-0 ~]#

Comment 18 narayanspg 2023-04-11 05:59:31 UTC
The above Cluster will be deleted EOD today. Let us know if it is getting used.

Comment 20 krishnaram Karthick 2023-04-12 07:21:24 UTC
Moving the bug to assigned based on the above comment.

Comment 23 narayanspg 2023-04-18 06:30:22 UTC
Hi Liran, as discussed you can close this one as noobaa pods are not crashing and original error is not seen now. created new BZ 2187602.
Thank you.

Comment 29 errata-xmlrpc 2023-06-21 15:24:25 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Red Hat OpenShift Data Foundation 4.13.0 enhancement and bug fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2023:3742


Note You need to log in before you can comment on or make changes to this bug.