Bug 1710124

Summary: [CRI-O] error creating read-write layer with ID no such file or directory
Product: OpenShift Container Platform Reporter: Steven Walter <stwalter>
Component: NodeAssignee: Peter Hunt <pehunt>
Status: CLOSED CURRENTRELEASE QA Contact: weiwei jiang <wjiang>
Severity: high Docs Contact:
Priority: high    
Version: 3.11.0CC: aos-bugs, bleanhar, dornelas, dwalsh, jokerman, kechung, knakai, ktadimar, mmccomas, mpatel, nagrawal, openshift-bugs-escalate, pehunt, rekhan, sponnaga, umohnani, wjiang
Target Milestone: ---   
Target Release: 3.11.z   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2020-05-13 20:35:34 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 1186913    

Description Steven Walter 2019-05-14 23:05:57 UTC
Description of problem:

Customer notes sometimes when application teams try to schedule their pods they see:
Failed create pod sandbox: rpc error: code = Unknown desc = error creating pod sandbox with name "k8s_test-exxample-15-xwsdw_example--example_3610-11e9-afa1-001a4a195f6a_0": error creating read-write layer with ID "23f96769bbe5xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxfa2ff4a7b3e7e": symlink ../23f96769bbe58288bae1bc6641bced0e55881dc0bd04046e15cfa2ff4a7b3e7e/diff /var/lib/containers/storage/overlay/l/4M53TMQASZFSDKR7R65KP4U4DF: no such file or directory"

Version-Release number of selected component (if applicable):
3.11

Working on getting crio and node logs

Comment 2 Mrunal Patel 2019-05-15 00:19:56 UTC
The fix for this went into cri-o 1.11.14.

Comment 13 Urvashi Mohnani 2019-06-06 13:08:06 UTC
Hi, could we get the cri-o logs or access to the cluster?

Comment 16 Urvashi Mohnani 2019-06-11 19:24:36 UTC
The original issue where the symlinks were not found has been fixed. But looks like another issue is occurring now. I will need access to the cluster where this is happening to further debug what is going wrong.