diff --git a/.openshift-tests-extension/openshift_payload_hyperkube.json b/.openshift-tests-extension/openshift_payload_hyperkube.json new file mode 100644 index 0000000000000..23e86487fa5f1 --- /dev/null +++ b/.openshift-tests-extension/openshift_payload_hyperkube.json @@ -0,0 +1,80079 @@ +[ + { + "name": "[sig-cluster-lifecycle] [Feature:BootstrapTokens] should sign the new added bootstrap tokens [Disabled:Unimplemented] [Suite:k8s]", + "labels": { + "Feature:BootstrapTokens": {}, + "sig-cluster-lifecycle": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cluster-lifecycle] [Feature:BootstrapTokens] should resign the bootstrap tokens when the clusterInfo ConfigMap updated [Serial] [Disruptive] [Disabled:Unimplemented] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:BootstrapTokens": {}, + "Serial": {}, + "sig-cluster-lifecycle": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cluster-lifecycle] [Feature:BootstrapTokens] should delete the signed bootstrap tokens from clusterInfo ConfigMap when bootstrap token is deleted [Disabled:Unimplemented] [Suite:k8s]", + "labels": { + "Feature:BootstrapTokens": {}, + "sig-cluster-lifecycle": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cluster-lifecycle] [Feature:BootstrapTokens] should delete the token secret when the secret expired [Disabled:Unimplemented] [Suite:k8s]", + "labels": { + "Feature:BootstrapTokens": {}, + "sig-cluster-lifecycle": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cluster-lifecycle] [Feature:BootstrapTokens] should not delete the token secret when the secret is not expired [Disabled:Unimplemented] [Suite:k8s]", + "labels": { + "Feature:BootstrapTokens": {}, + "sig-cluster-lifecycle": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-architecture] Conformance Tests should have at least two untainted nodes [Conformance] [Skipped:SingleReplicaTopology] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-architecture": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] Certificates API [Privileged:ClusterAdmin] should support building a client with a CSR [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] Certificates API [Privileged:ClusterAdmin] should support CSR API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:NodeAuthenticator] The kubelet's main port 10250 should reject requests with no credentials [Skipped:ibmroks] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NodeAuthenticator": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:NodeAuthenticator] The kubelet can delegate ServiceAccount tokens to the API server [Skipped:ibmroks] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NodeAuthenticator": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:NodeAuthorizer] Getting a non-existent secret should exit with the Forbidden error, not a NotFound error [Skipped:ibmroks] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NodeAuthorizer": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:NodeAuthorizer] Getting an existing secret should exit with the Forbidden error [Skipped:ibmroks] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NodeAuthorizer": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:NodeAuthorizer] Getting a non-existent configmap should exit with the Forbidden error, not a NotFound error [Skipped:ibmroks] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NodeAuthorizer": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:NodeAuthorizer] Getting an existing configmap should exit with the Forbidden error [Skipped:ibmroks] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NodeAuthorizer": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:NodeAuthorizer] Getting a secret for a workload the node has access to should succeed [Skipped:ibmroks] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NodeAuthorizer": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:NodeAuthorizer] A node shouldn't be able to create another node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NodeAuthorizer": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:NodeAuthorizer] A node shouldn't be able to delete another node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NodeAuthorizer": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] [Feature:ClusterTrustBundle] [Feature:ClusterTrustBundleProjection] should be able to mount a single ClusterTrustBundle by name [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:ClusterTrustBundle": {}, + "Feature:ClusterTrustBundleProjection": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] SelfSubjectReview testing SSR in different API groups authentication/v1beta1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] SelfSubjectReview testing SSR in different API groups authentication/v1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] SelfSubjectReview should support SelfSubjectReview API operations [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts no secret-based service account token should be auto-generated [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts should mount an API token into pods [Conformance] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts should mount projected service account token [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts should set ownership and permission when RunAsUser or FsGroup is present [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts should support InClusterConfig with token rotation [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts should run through the lifecycle of a ServiceAccount [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts should guarantee kube-root-ca.crt exist in any namespace [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccounts should update a ServiceAccount [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] SubjectReview should support SubjectReview API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Pods should function for intra-pod communication: sctp [LinuxOnly] [Feature:SCTPConnectivity] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SCTPConnectivity": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Pods should function for node-pod communication: sctp [LinuxOnly] [Feature:SCTPConnectivity] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SCTPConnectivity": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API [Serial] [Disruptive] [Feature:EphemeralStorage] Downward API tests for local ephemeral storage should provide container's limits.ephemeral-storage and requests.ephemeral-storage as env vars [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:EphemeralStorage": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API [Serial] [Disruptive] [Feature:EphemeralStorage] Downward API tests for local ephemeral storage should provide default limits.ephemeral-storage from node allocatable [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:EphemeralStorage": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] new files should be created with FSGroup ownership when container is root [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] new files should be created with FSGroup ownership when container is non-root [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] nonexistent volume subPath should have the correct mode and owner using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] files with FSGroup ownership should support (root,0644,tmpfs) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] volume on default medium should have the correct mode using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] [NodeFeature:FSGroup] volume on tmpfs should have the correct mode using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir volumes pod should support memory backed volumes of specified size [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance] [Skipped:ibmroks] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPath should support r/w [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPath should support subPath [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected combined should project all components that make up the projection API [Projection] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable from pods in volume as non-root with FSGroup [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root with FSGroup [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap should be immutable if `immutable` field is set [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap Should fail non-optional pod creation due to configMap object does not exist [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] ConfigMap Should fail non-optional pod creation due to the key in the configMap object does not exist [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should provide podname as non-root with fsgroup [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should provide podname as non-root with fsgroup and defaultMode [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable from pods in volume as non-root with FSGroup [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root with FSGroup [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap Should fail non-optional pod creation due to configMap object does not exist [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected configMap Should fail non-optional pod creation due to the key in the configMap object does not exist [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should provide podname as non-root with fsgroup [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should provide podname as non-root with fsgroup and defaultMode [LinuxOnly] [NodeFeature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:FSGroup": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret Should fail non-optional pod creation due to secret object does not exist [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Projected secret Should fail non-optional pod creation due to the key in the secret object does not exist [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets should be immutable if `immutable` field is set [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets Should fail non-optional pod creation due to secret object does not exist [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Secrets Should fail non-optional pod creation due to the key in the secret object does not exist [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Volumes NFSv4 should be mountable for NFSv4 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Volumes NFSv3 should be mountable for NFSv3 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] Logging soak [Performance] [Slow] [Disruptive] should survive logging 1KB every 1s seconds, for a duration of 2m0s [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] MetricsGrabber should grab all metrics from API server. [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] MetricsGrabber should grab all metrics from a Kubelet. [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] MetricsGrabber should grab all metrics from a Scheduler. [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] MetricsGrabber should grab all metrics from a ControllerManager. [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] MetricsGrabber should grab all metrics slis from API server. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] Events should manage the lifecycle of an event [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] Events should delete a collection of events [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] Events API should ensure that an event can be fetched, patched, deleted, and listed [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] Events API should delete a collection of events [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-instrumentation] Metrics should grab all metrics from kubelet /metrics/resource endpoint [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-instrumentation": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should support a 'default-deny-ingress' policy [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should support a 'default-deny-all' policy [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy to allow traffic from pods within server namespace based on PodSelector [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy to allow ingress traffic for a target [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy to allow ingress traffic from pods in all namespaces [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy to allow traffic only from a different namespace, based on NamespaceSelector [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on PodSelector with MatchExpressions [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on NamespaceSelector with MatchExpressions [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on PodSelector or NamespaceSelector [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on PodSelector and NamespaceSelector [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on Multiple PodSelectors and NamespaceSelectors [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on any PodSelectors [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy to allow traffic only from a pod in a different namespace based on PodSelector and NamespaceSelector [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on Ports [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce multiple, stacked policies with overlapping podSelectors [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should support allow-all policy [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should allow ingress access on one named port [Feature:NetworkPolicy] [Skipped:Network/OVNKubernetes] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should allow ingress access from namespace on one named port [Feature:NetworkPolicy] [Skipped:Network/OVNKubernetes] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should allow egress access on one named port [Feature:NetworkPolicy] [Skipped:Network/OVNKubernetes] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce updated policy [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should allow ingress access from updated namespace [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should allow ingress access from updated pod [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should deny ingress from pods on other namespaces [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should deny ingress access to updated pod [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should deny egress from pods based on PodSelector [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should deny egress from all pods in a namespace [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should work with Ingress, Egress specified together [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should support denying of egress traffic on the client side (even if the server explicitly allows this traffic) [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce egress policy allowing traffic to a server in a different namespace based on PodSelector and NamespaceSelector [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce ingress policy allowing any port traffic to a server on a specific protocol [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce multiple ingress policies with ingress allow-all policy taking precedence [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce multiple egress policies with egress allow-all policy taking precedence [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should stop enforcing policies after they are deleted [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should allow egress access to server in CIDR block [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce except clause while egress access to server in CIDR block [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should ensure an IP overlapping both IPBlock.CIDR and IPBlock.Except is allowed [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policies to check ingress and egress policies can be controlled independently based on PodSelector [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Skipped:Network/OpenShiftSDN] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should not mistakenly treat 'protocol: SCTP' as 'protocol: TCP', even if the plugin doesn't support SCTP [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should properly isolate pods that are selected by a policy allowing SCTP, even if the plugin doesn't support SCTP [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should not allow access by TCP when a policy specifies only UDP [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy to allow traffic based on NamespaceSelector with MatchLabels using default ns label [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on NamespaceSelector with MatchExpressions using default ns label [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol [LinuxOnly] NetworkPolicy between server and client using UDP should support a 'default-deny-ingress' policy [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol [LinuxOnly] NetworkPolicy between server and client using UDP should enforce policy based on Ports [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol [LinuxOnly] NetworkPolicy between server and client using UDP should enforce policy to allow traffic only from a pod in a different namespace based on PodSelector and NamespaceSelector [Feature:NetworkPolicy] [Disabled:Broken] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol [Feature:SCTPConnectivity] [LinuxOnly] NetworkPolicy between server and client using SCTP should support a 'default-deny-ingress' policy [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "Feature:SCTPConnectivity": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol [Feature:SCTPConnectivity] [LinuxOnly] NetworkPolicy between server and client using SCTP should enforce policy based on Ports [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "Feature:SCTPConnectivity": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol [Feature:SCTPConnectivity] [LinuxOnly] NetworkPolicy between server and client using SCTP should enforce policy to allow traffic only from a pod in a different namespace based on PodSelector and NamespaceSelector [Feature:NetworkPolicy] [Skipped:Network/OpenShiftSDN/Multitenant] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:NetworkPolicy": {}, + "Feature:SCTPConnectivity": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol API should support creating NetworkPolicy API operations [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Netpol API should support creating NetworkPolicy API with endport field [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] kubelet registers plugin [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] kubelet must retry NodePrepareResources [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] kubelet must not run a pod if a claim is not ready [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] kubelet must unprepare resources for force-deleted pod [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] kubelet must call NodePrepareResources even if not used by any container [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] kubelet must map configs and devices to the right containers [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports claim and class parameters [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports reusing resources [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports sharing a claim concurrently [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports sharing a claim sequentially [Slow] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node retries pod scheduling after creating device class [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node retries pod scheduling after updating device class [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node runs a pod without a generated resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports simple pod referencing inline resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports inline claim referenced by multiple containers [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports simple pod referencing external resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports external claim referenced by multiple pods [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports external claim referenced by multiple containers of multiple pods [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node supports init containers [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node removes reservation from claim when pod is done [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node deletes generated claims when pod is done [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node does not delete generated claims when pod is restarting [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on single node must deallocate after use [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on multiple nodes with network-attached resources schedules onto different nodes [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on multiple nodes with network-attached resources [Serial] [Disruptive] [Slow] must deallocate on non graceful node shutdown [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "Serial": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on multiple nodes reallocation works [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with classic DRA [Feature:DRAControlPlaneController] on multiple nodes with node-local resources uses all resources [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports claim and class parameters [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports reusing resources [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports sharing a claim concurrently [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports sharing a claim sequentially [Slow] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node retries pod scheduling after creating device class [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node retries pod scheduling after updating device class [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node runs a pod without a generated resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports simple pod referencing inline resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports inline claim referenced by multiple containers [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports simple pod referencing external resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports external claim referenced by multiple pods [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports external claim referenced by multiple containers of multiple pods [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node supports init containers [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node removes reservation from claim when pod is done [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node deletes generated claims when pod is done [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node does not delete generated claims when pod is restarting [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on single node must deallocate after use [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on multiple nodes with different ResourceSlices keeps pod pending because of CEL runtime errors [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] with structured parameters on multiple nodes with node-local resources uses all resources [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster support validating admission policy for admin access [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster truncates the name of a generated resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster supports count/resourceclaims.resource.k8s.io ResourceQuota [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor supports scheduled pod referencing inline resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor supports scheduled pod referencing external resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor supports simple pod referencing inline resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor supports inline claim referenced by multiple containers [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor supports simple pod referencing external resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor supports external claim referenced by multiple pods [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor supports external claim referenced by multiple containers of multiple pods [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor supports init containers [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor removes reservation from claim when pod is done [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor deletes generated claims when pod is done [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor does not delete generated claims when pod is restarting [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] with setting ReservedFor must deallocate after use [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor supports scheduled pod referencing inline resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor supports scheduled pod referencing external resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor supports simple pod referencing inline resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor supports inline claim referenced by multiple containers [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor supports simple pod referencing external resource claim [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor supports external claim referenced by multiple pods [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor supports external claim referenced by multiple containers of multiple pods [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor supports init containers [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor removes reservation from claim when pod is done [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor deletes generated claims when pod is done [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor does not delete generated claims when pod is restarting [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with classic DRA [Feature:DRAControlPlaneController] without setting ReservedFor must deallocate after use [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DRAControlPlaneController": {}, + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with structured parameters must apply per-node permission checks [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] cluster with structured parameters must manage ResourceSlices [Slow] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] DRA [Feature:DynamicResourceAllocation] multiple drivers using only drapbv1alpha3 work [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] ConfigMap should update ConfigMap successfully [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] ConfigMap should run through a ConfigMap lifecycle [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] ConfigMap should be consumable as environment variable names when configmap keys start with a digit [Feature:RelaxedEnvironmentVariableValidation] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:RelaxedEnvironmentVariableValidation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should be restarted with an exec liveness probe with timeout [MinimumKubeletVersion:1.20] [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should not be ready with an exec readiness probe timeout [MinimumKubeletVersion:1.20] [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should be restarted with a failing exec liveness probe that took longer than the timeout [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should be restarted with a local redirect http liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should *not* be restarted with a non-local redirect http liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should be restarted startup probe fails [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should *not* be restarted by liveness probe because startup probe delays it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should be restarted by liveness probe after startup probe enables it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should be ready immediately after startupProbe succeeds [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should override timeoutGracePeriodSeconds when LivenessProbe field is set [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should override timeoutGracePeriodSeconds when StartupProbe field is set [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should *not* be restarted with a GRPC liveness probe [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should be restarted with a GRPC liveness probe [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should mark readiness on pods to false while pod is in progress of terminating when a pod has a readiness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Probing container should mark readiness on pods to false and disable liveness probes while pod is in progress of terminating [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container with readiness probe should not be ready before initial delay and never restart [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container with readiness probe that fails should never be ready and never restart [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should be restarted with a exec \"cat /tmp/health\" liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should be restarted with a /healthz http liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should *not* be restarted with a tcp:8080 liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should have monotonically increasing restart count [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should *not* be restarted with a /healthz http liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should be restarted with an exec liveness probe with timeout [MinimumKubeletVersion:1.20] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should not be ready with an exec readiness probe timeout [MinimumKubeletVersion:1.20] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should be restarted with a failing exec liveness probe that took longer than the timeout [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should be restarted with a local redirect http liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should *not* be restarted with a non-local redirect http liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should be restarted startup probe fails [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should *not* be restarted by liveness probe because startup probe delays it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should be restarted by liveness probe after startup probe enables it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should be ready immediately after startupProbe succeeds [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should override timeoutGracePeriodSeconds when LivenessProbe field is set [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should override timeoutGracePeriodSeconds when StartupProbe field is set [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should *not* be restarted with a GRPC liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should be restarted with a GRPC liveness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should mark readiness on pods to false while pod is in progress of terminating when a pod has a readiness probe [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Probing restartable init container should mark readiness on pods to false and disable liveness probes while pod is in progress of terminating [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Containers should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Containers should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Downward API should provide host IP and pod IP as an env var if pod uses host network [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Downward API [Serial] [Disruptive] [NodeFeature:DownwardAPIHugePages] Downward API tests for hugepages should provide container's limits.hugepages-\u003cpagesize\u003e and requests.hugepages-\u003cpagesize\u003e as env vars [Suite:k8s]", + "labels": { + "Disruptive": {}, + "NodeFeature:DownwardAPIHugePages": {}, + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Downward API [Serial] [Disruptive] [NodeFeature:DownwardAPIHugePages] Downward API tests for hugepages should provide default limits.hugepages-\u003cpagesize\u003e from node allocatable [Suite:k8s]", + "labels": { + "Disruptive": {}, + "NodeFeature:DownwardAPIHugePages": {}, + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Ephemeral Containers [NodeConformance] will start an ephemeral container in an existing pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Ephemeral Containers [NodeConformance] should update the ephemeral containers in an existing pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Variable Expansion should allow substituting values in a volume subpath [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Variable Expansion should fail substituting values in a volume subpath with backticks [Slow] [Conformance] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Variable Expansion should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Variable Expansion should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Variable Expansion should succeed in writing subpaths in container [Slow] [Conformance] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Variable Expansion allow almost all printable ASCII characters as environment variable names [Feature:RelaxedEnvironmentVariableValidation] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:RelaxedEnvironmentVariableValidation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] ImageCredentialProvider [Feature:KubeletCredentialProviders] should be able to create pod with image credentials fetched from external credential provider [Disabled:RebaseInProgress] [Suite:k8s]", + "labels": { + "Feature:KubeletCredentialProviders": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Kubelet when scheduling an agnhost Pod with hostAliases should write entries to /etc/hosts [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Kubelet with pods in a privileged namespace when scheduling an agnhost Pod with hostAliases and hostNetwork should write entries to /etc/hosts when hostNetwork is enabled [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Lease lease API should be available [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart https hook properly [MinimumKubeletVersion:1.23] [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop https hook properly [MinimumKubeletVersion:1.23] [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart https hook properly [MinimumKubeletVersion:1.23] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [NodeFeature:SidecarContainers] [Feature:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop https hook properly [MinimumKubeletVersion:1.23] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SidecarContainers": {}, + "NodeFeature:SidecarContainers": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [Feature:PodLifecycleSleepAction] when create a pod with lifecycle hook using sleep action valid prestop hook using sleep action [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:PodLifecycleSleepAction": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [Feature:PodLifecycleSleepAction] when create a pod with lifecycle hook using sleep action reduce GracePeriodSeconds during runtime [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:PodLifecycleSleepAction": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [Feature:PodLifecycleSleepAction] when create a pod with lifecycle hook using sleep action ignore terminated container [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:PodLifecycleSleepAction": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NodeLease NodeLease the kubelet should create and update a lease in the kube-node-lease namespace [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NodeLease NodeLease should have OwnerReferences set [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NodeLease NodeLease the kubelet should report node status infrequently [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] PodOSRejection [NodeConformance] Kubelet should reject pod when the node OS doesn't match pod's OS [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should get a host IP [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should be submitted and removed [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should be updated [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should contain environment variables for services [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should support remote command execution over websockets [NodeConformance] [Conformance] [Skipped:Proxy] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance] [Skipped:Proxy] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should have their auto-restart back-off timer reset on image update [Slow] [NodeConformance] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should cap back-off at MaxContainerBackOff [Slow] [NodeConformance] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should support pod readiness gates [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should delete a collection of pods [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should run through the lifecycle of Pods and PodStatus [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods should patch a pod status [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] PodTemplates should run the lifecycle of PodTemplates [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] PodTemplates should delete a collection of pod templates [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] PodTemplates should replace a pod template [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] PrivilegedPod [NodeConformance] should enable privileged commands [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test on terminated container should report termination message if TerminationMessagePath is set [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test on terminated container should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test on terminated container should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test on terminated container should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test on terminated container should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test when running a container with a new image should not be able to pull image from invalid registry [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test when running a container with a new image should be able to pull image [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test when running a container with a new image should not be able to pull from private registry without secret [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Container Runtime blackbox test when running a container with a new image should be able to pull from private registry with secret [NodeConformance] [Disabled:Broken] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should reject a Pod requesting a RuntimeClass with an unconfigured handler [NodeFeature:RuntimeHandler] [Disabled:Broken] [Suite:k8s]", + "labels": { + "NodeFeature:RuntimeHandler": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should run a Pod requesting a RuntimeClass with a configured handler [NodeFeature:RuntimeHandler] [Disabled:Broken] [Suite:k8s]", + "labels": { + "NodeFeature:RuntimeHandler": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should support RuntimeClasses API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Secrets should be consumable via the environment [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Secrets should fail to create secret due to empty secret key [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Secrets should patch a secret [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Secrets should be consumable as environment variable names when secret keys start with a digit [Feature:RelaxedEnvironmentVariableValidation] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:RelaxedEnvironmentVariableValidation": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a pod with HostUsers must create the user namespace if set to false [LinuxOnly] [Feature:UserNamespacesSupport] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:UserNamespacesSupport": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a pod with HostUsers must not create the user namespace if set to true [LinuxOnly] [Feature:UserNamespacesSupport] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:UserNamespacesSupport": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a pod with HostUsers should mount all volumes with proper permissions with hostUsers=false [LinuxOnly] [Feature:UserNamespacesSupport] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:UserNamespacesSupport": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a pod with HostUsers should set FSGroup to user inside the container with hostUsers=false [LinuxOnly] [Feature:UserNamespacesSupport] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:UserNamespacesSupport": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a container with runAsUser should run the container with uid 0 [LinuxOnly] [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a container with runAsNonRoot should run with an explicit non-root user ID [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a container with runAsNonRoot should not run with an explicit root user ID [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a container with runAsNonRoot should run with an image specified user ID [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a container with runAsNonRoot should not run without a specified user ID [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a pod with readOnlyRootFilesystem should run the container with readonly rootfs when readOnlyRootFilesystem=true [LinuxOnly] [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context When creating a pod with privileged should run the container as privileged when true [LinuxOnly] [NodeFeature:HostAccess] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeFeature:HostAccess": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context when creating containers with AllowPrivilegeEscalation should allow privilege escalation when not explicitly set and uid != 0 [LinuxOnly] [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context when creating containers with AllowPrivilegeEscalation should allow privilege escalation when true [LinuxOnly] [NodeConformance] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] User Namespaces for Pod Security Standards [LinuxOnly] with UserNamespacesSupport and UserNamespacesPodSecurityStandards enabled should allow pod [Feature:UserNamespacesPodSecurityStandards] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:UserNamespacesPodSecurityStandards": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Sysctls [LinuxOnly] [NodeConformance] should support sysctls [MinimumKubeletVersion:1.21] [Environment:NotInUserNS] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Environment:NotInUserNS": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Sysctls [LinuxOnly] [NodeConformance] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Sysctls [LinuxOnly] [NodeConformance] should not launch unsafe, but not explicitly enabled sysctls on the node [MinimumKubeletVersion:1.21] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Sysctls [LinuxOnly] [NodeConformance] should support sysctls with slashes as separator [MinimumKubeletVersion:1.23] [Environment:NotInUserNS] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Environment:NotInUserNS": {}, + "NodeConformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] LimitRange should list, patch and delete a LimitRange by collection [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates local ephemeral storage resource limits of pods that are allowed to run [Feature:LocalStorageCapacityIsolation] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:LocalStorageCapacityIsolation": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates pod overhead is considered along with resource limits of pods that are allowed to run verify pod overhead is accounted for [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates that NodeAffinity is respected if not matching [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates that required NodeAffinity setting is respected if matching [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates that taints-tolerations is respected if matching [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates that taints-tolerations is respected if not matching [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] [Slow] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] PodTopologySpread Filtering validates 4 pods with MaxSkew=1 are evenly distributed into 2 nodes [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] validates Pods with non-empty schedulingGates are blocked on scheduling [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPredicates [Serial] when PVC has node-affinity to non-existent/illegal nodes, the pod should be scheduled normally if suitable nodes exist [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPreemption [Serial] validates basic preemption works [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPreemption [Serial] validates lower priority pod preemption by critical pod [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPreemption [Serial] validates pod disruption condition is added to the preempted pod [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPreemption [Serial] PodTopologySpread Preemption validates proper pods are preempted [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPreemption [Serial] PriorityClass endpoints verify PriorityClass endpoints can be operated with different HTTP methods [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPriorities [Serial] Pod should be scheduled to node that don't match the PodAntiAffinity terms [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPriorities [Serial] Pod should be preferably scheduled to nodes pod can tolerate [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] SchedulerPriorities [Serial] PodTopologySpread Scoring validates pod should be preferably scheduled to node which makes the matching pods more evenly distributed [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] Multi-AZ Clusters should spread the pods of a service across zones [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-scheduling] Multi-AZ Clusters should spread the pods of a replication controller across zones [Serial] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-scheduling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should schedule multiple jobs concurrently [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should not schedule jobs when suspended [Slow] [Conformance] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Slow": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Slow": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should replace jobs when ReplaceConcurrent [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should be able to schedule after more than 100 missed schedule [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should not emit unexpected warnings [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should remove from active list jobs that have been deleted [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should delete successful finished jobs with limit of one successful job [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should delete failed finished jobs with limit of one job [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should support timezone [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] CronJob should support CronJob API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DaemonRestart [Disruptive] Controller Manager should not create/delete replicas across restart [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DaemonRestart [Disruptive] Scheduler should continue assigning pods to nodes across restart [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DaemonRestart [Disruptive] Kubelet should not restart containers across restart [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DaemonRestart [Disruptive] Kube-proxy should recover after being killed accidentally [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should run a job to completion when tasks succeed [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should allow to use the pod failure policy on exit code to fail the job early [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should allow to use the pod failure policy to not count the failure towards the backoffLimit [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should allow to use a pod failure policy to ignore failure for an evicted pod; matching on the exit code [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should allow to use a pod failure policy to ignore failure for an evicted pod; matching on the DisruptionTarget condition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should not create pods when created in suspend state [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should delete pods when suspended [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should recreate pods only after they have failed if pod replacement policy is set to Failed [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should create pods for an Indexed job with completion indexes and specified hostname [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job with successPolicy should succeeded when all indexes succeeded [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job with successPolicy succeededIndexes rule should succeeded even when some indexes remain pending [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job with successPolicy succeededCount rule should succeeded even when some indexes remain pending [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should execute all indexes despite some failing when using backoffLimitPerIndex [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should terminate job execution when the number of failed indexes exceeds maxFailedIndexes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should mark indexes as failed when the FailIndex action is matched in podFailurePolicy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should remove pods when job is deleted [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should run a job to completion when tasks sometimes fail and are not locally restarted [Flaky] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should fail when exceeds active deadline [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should delete a job [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should fail to exceed backoffLimit [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should run a job to completion with CPU requests [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should apply changes to a job status [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should manage the lifecycle of a job [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Job should update the status ready field [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] TTLAfterFinished job should be deleted once it finishes after TTL seconds [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ControllerRevision [Serial] should manage the lifecycle of a ControllerRevision [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController Listing PodDisruptionBudgets for all namespaces should list and delete a collection of PodDisruptionBudgets [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should create a PodDisruptionBudget [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should observe PodDisruptionBudget status updated [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should update/patch PodDisruptionBudget status [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should observe that the PodDisruptionBudget status is not updated for unmanaged pods [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController evictions: no PDB =\u003e should allow an eviction [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController evictions: too few pods, absolute =\u003e should not allow an eviction [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController evictions: enough pods, absolute =\u003e should allow an eviction [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController evictions: enough pods, replicaSet, percentage =\u003e should allow an eviction [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController evictions: too few pods, replicaSet, percentage =\u003e should not allow an eviction [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController evictions: maxUnavailable allow single eviction, percentage =\u003e should allow an eviction [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController evictions: maxUnavailable deny evictions, integer =\u003e should not allow an eviction [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should block an eviction until the PDB is updated to allow it [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should evict ready pods with Default UnhealthyPodEvictionPolicy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should evict ready pods with IfHealthyBudget UnhealthyPodEvictionPolicy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should evict ready pods with AlwaysAllow UnhealthyPodEvictionPolicy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should not evict unready pods with Default UnhealthyPodEvictionPolicy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should not evict unready pods with IfHealthyBudget UnhealthyPodEvictionPolicy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] DisruptionController should evict unready pods with AlwaysAllow UnhealthyPodEvictionPolicy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicationController should serve a basic image on each replica with a private image [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicationController should adopt matching pods on creation [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicationController should release no longer matching pods [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicationController should test the lifecycle of a ReplicationController [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicationController should get and update a ReplicationController scale [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicaSet should serve a basic image on each replica with a private image [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicaSet should surface a failure condition on a common issue like exceeded quota [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicaSet Replicaset should have a working scale subresource [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicaSet Replace and Patch tests [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicaSet should list and delete a collection of ReplicaSets [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] ReplicaSet should validate Replicaset Status endpoints [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should provide basic identity [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should adopt matching orphans and release non-matching pods [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should not deadlock when a pod's predecessor fails [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications with PVCs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications for partiton1 and delete pod-0 without failing container [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications for partiton1 and delete pod-0 with failing container [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should implement legacy replacement when the update strategy is OnDelete [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Slow": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Slow": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should list, patch and delete a collection of StatefulSets [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should validate Statefulset Status endpoints [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Deploy clustered applications [Feature:StatefulSet] [Slow] should creating a working zookeeper cluster [Suite:k8s]", + "labels": { + "Feature:StatefulSet": {}, + "Slow": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Deploy clustered applications [Feature:StatefulSet] [Slow] should creating a working redis cluster [Suite:k8s]", + "labels": { + "Feature:StatefulSet": {}, + "Slow": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Deploy clustered applications [Feature:StatefulSet] [Slow] should creating a working mysql cluster [Suite:k8s]", + "labels": { + "Feature:StatefulSet": {}, + "Slow": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Deploy clustered applications [Feature:StatefulSet] [Slow] should creating a working CockroachDB cluster [Suite:k8s]", + "labels": { + "Feature:StatefulSet": {}, + "Slow": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet MinReadySeconds should be honored when enabled [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet AvailableReplicas should get updated accordingly when MinReadySeconds is enabled [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should delete PVCs with a WhenDeleted policy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should delete PVCs with a OnScaledown policy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should not delete PVC with OnScaledown policy if another controller owns the PVC [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should delete PVCs after adopting pod (WhenDeleted) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should delete PVCs after adopting pod (WhenScaled) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should not delete PVCs when there is another controller [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Automatically recreate PVC for pending pod when PVC is missing PVC should be recreated when pod is pending due to missing PVC [Disruptive] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Scaling StatefulSetStartOrdinal Setting .start.ordinal [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Scaling StatefulSetStartOrdinal Increasing .start.ordinal [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Scaling StatefulSetStartOrdinal Decreasing .start.ordinal [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] StatefulSet Scaling StatefulSetStartOrdinal Removing .start.ordinal [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should run and stop complex daemon with node affinity [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should not update pod when spec was updated and update strategy is OnDelete [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance] [Skipped:SingleReplicaTopology] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should surge pods onto nodes when spec was updated and update strategy is RollingUpdate [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should list and delete a collection of DaemonSets [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Daemon set [Serial] should verify changes to a daemon set status [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment deployment reaping should cascade to its replica sets and pods [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment deployment should delete old replica sets [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment deployment should support rollover [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment iterative rollouts should eventually progress [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment test Deployment ReplicaSet orphaning and adoption regarding controllerRef [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment Deployment should have a working scale subresource [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment deployment should support proportional scaling [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment should not disrupt a cloud load-balancer's connectivity during rollout [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment should run the lifecycle of a Deployment [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] Deployment should validate Deployment Status endpoints [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:ClusterSizeAutoscalingScaleUp] [Slow] Autoscaling Autoscaling a service from 1 pod and 3 nodes to 8 pods and \u003e=4 nodes takes less than 15 minutes [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaler scalability [Slow] should scale up at all [Feature:ClusterAutoscalerScalability1] [Suite:k8s]", + "labels": { + "Feature:ClusterAutoscalerScalability1": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaler scalability [Slow] should scale up twice [Feature:ClusterAutoscalerScalability2] [Suite:k8s]", + "labels": { + "Feature:ClusterAutoscalerScalability2": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaler scalability [Slow] should scale down empty nodes [Feature:ClusterAutoscalerScalability3] [Suite:k8s]", + "labels": { + "Feature:ClusterAutoscalerScalability3": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaler scalability [Slow] should scale down underutilized nodes [Feature:ClusterAutoscalerScalability4] [Suite:k8s]", + "labels": { + "Feature:ClusterAutoscalerScalability4": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaler scalability [Slow] shouldn't scale down with underutilized nodes due to host port conflicts [Feature:ClusterAutoscalerScalability5] [Suite:k8s]", + "labels": { + "Feature:ClusterAutoscalerScalability5": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaler scalability [Slow] CA ignores unschedulable pods while scheduling schedulable pods [Feature:ClusterAutoscalerScalability6] [Suite:k8s]", + "labels": { + "Feature:ClusterAutoscalerScalability6": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] shouldn't increase cluster size if pending pod is too large [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pending pods are small [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] Should scale up GPU pool from 0 [GpuType:] [Feature:ClusterSizeAutoscalingGpu] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingGpu": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] Should scale up GPU pool from 1 [GpuType:] [Feature:ClusterSizeAutoscalingGpu] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingGpu": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] Should not scale GPU pool up if pod does not require GPUs [GpuType:] [Feature:ClusterSizeAutoscalingGpu] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingGpu": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] Should scale down GPU pool from 1 [GpuType:] [Feature:ClusterSizeAutoscalingGpu] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingGpu": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pending pods are small and one node is broken [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] shouldn't trigger additional scale-ups during processing scale-up [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pending pods are small and there is another node pool that is not autoscaled [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should disable node pool autoscaling [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pods are pending due to host port conflict [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pods are pending due to pod anti-affinity [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pod requesting EmptyDir volume is pending [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pod requesting volume is pending [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should add node to the particular mig [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should scale up correct target pool [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should correctly scale down after a node is not needed [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should correctly scale down after a node is not needed and one node is broken [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should correctly scale down after a node is not needed when there is non autoscaled pool [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should be able to scale down when rescheduling a pod is required and pdb allows for it [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] shouldn't be able to scale down when rescheduling a pod is required, but pdb doesn't allow drain [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should be able to scale down by draining multiple pods one by one as dictated by pdb [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should be able to scale down by draining system pods with pdb [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] Should be able to scale a node group up from 0 [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] Should be able to scale a node group down to 0 [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] Shouldn't perform scale up operation and should list unhealthy status if most of the cluster is broken [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] shouldn't scale up when expendable pod is created [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should scale up when non expendable pod is created [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] shouldn't scale up when expendable pod is preempted [Feature:ClusterSizeAutoscalingScaleUp] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleUp": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should scale down when expendable pod is running [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] shouldn't scale down when non expendable pod is running [Feature:ClusterSizeAutoscalingScaleDown] [Suite:k8s]", + "labels": { + "Feature:ClusterSizeAutoscalingScaleDown": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] should scale up when unprocessed pod is created and is going to be unschedulable [Feature:ClusterScaleUpBypassScheduler] [Suite:k8s]", + "labels": { + "Feature:ClusterScaleUpBypassScheduler": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] shouldn't scale up when unprocessed pod is created and is going to be schedulable [Feature:ClusterScaleUpBypassScheduler] [Suite:k8s]", + "labels": { + "Feature:ClusterScaleUpBypassScheduler": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] Cluster size autoscaling [Slow] shouldn't scale up when unprocessed pod is created and scheduler is not specified to be bypassed [Feature:ClusterScaleUpBypassScheduler] [Suite:k8s]", + "labels": { + "Feature:ClusterScaleUpBypassScheduler": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] DNS horizontal autoscaling [Serial] [Slow] [KubeUp] [sig-cloud-provider-gcp] kube-dns-autoscaler should scale kube-dns pods when cluster size changed [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "KubeUp": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] DNS horizontal autoscaling kube-dns-autoscaler should scale kube-dns pods in both nonfaulty and faulty scenarios [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] Deployment (Pod Resource) Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods using Average Utilization for aggregation [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] Deployment (Pod Resource) Should scale from 5 pods to 3 pods and then from 3 pods to 1 pod using Average Utilization for aggregation [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] Deployment (Pod Resource) Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods using Average Value for aggregation [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] Deployment (Container Resource) Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods using Average Utilization for aggregation [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] Deployment (Container Resource) Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods using Average Value for aggregation [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] ReplicaSet Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] ReplicaSet Should scale from 5 pods to 3 pods and then from 3 pods to 1 pod [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] ReplicationController Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods and verify decision stability [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] ReplicationController Should scale from 5 pods to 3 pods and then from 3 pods to 1 pod and verify decision stability [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) ReplicationController light Should scale from 1 pod to 2 pods [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) ReplicationController light [Slow] Should scale from 2 pods to 1 pod [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] ReplicaSet with idle sidecar (ContainerResource use case) Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods on a busy application with an idle sidecar container [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) [Serial] [Slow] ReplicaSet with idle sidecar (ContainerResource use case) Should not scale up on a busy sidecar with an idle application [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: CPU) CustomResourceDefinition Should scale with a CRD targetRef [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: Memory) [Serial] [Slow] Deployment (Pod Resource) Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods using Average Utilization for aggregation [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: Memory) [Serial] [Slow] Deployment (Pod Resource) Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods using Average Value for aggregation [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: Memory) [Serial] [Slow] Deployment (Container Resource) Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods using Average Utilization for aggregation [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] Horizontal pod autoscaling (scale resource: Memory) [Serial] [Slow] Deployment (Container Resource) Should scale from 1 pod to 3 pods and then from 3 pods to 5 pods using Average Value for aggregation [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with short downscale stabilization window should scale down soon after the stabilization period [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with long upscale stabilization window should scale up only after the stabilization period [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with autoscaling disabled shouldn't scale up [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with autoscaling disabled shouldn't scale down [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with scale limited by number of Pods rate should scale up no more than given number of Pods per minute [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with scale limited by number of Pods rate should scale down no more than given number of Pods per minute [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with scale limited by percentage should scale up no more than given percentage of current Pods per minute [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with scale limited by percentage should scale down no more than given percentage of current Pods per minute [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with both scale up and down controls configured should keep recommendation within the range over two stabilization windows [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-autoscaling] [Feature:HPA] [Serial] [Slow] Horizontal pod autoscaling (non-default behavior) with both scale up and down controls configured should keep recommendation within the range with stabilization window and pod limit rate [Suite:k8s]", + "labels": { + "Feature:HPA": {}, + "Serial": {}, + "Slow": {}, + "sig-autoscaling": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Conntrack should be able to preserve UDP traffic when server pod cycles for a NodePort service [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Conntrack should be able to preserve UDP traffic when server pod cycles for a ClusterIP service [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Conntrack should be able to preserve UDP traffic when server pod cycles for a ClusterIP service and client is hostNetwork [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Conntrack should be able to preserve UDP traffic when initial unready endpoints get ready [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Conntrack proxy implementation should not be vulnerable to the invalid conntrack state bug [Privileged] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should provide DNS for the cluster [Conformance] [Skipped:Proxy] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should provide DNS for the cluster [Provider:GCE] [Skipped:Proxy] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should resolve DNS of partial qualified names for the cluster [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should provide /etc/hosts entries for the cluster [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should provide DNS for services [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] [Skipped:Proxy] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should provide DNS for pods for Hostname [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should provide DNS for pods for Subdomain [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should provide DNS for ExternalName services [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should support configurable pod DNS nameservers [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should support configurable pod resolv.conf [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS should work with the pod containing more than 6 DNS search paths and longer than 256 search list characters [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS HostNetwork should resolve DNS of partial qualified names for services on hostNetwork pods with dnsPolicy: ClusterFirstWithHostNet [LinuxOnly] [Disabled:RebaseInProgress] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS HostNetwork spec.Hostname field is not silently ignored and is used for hostname for a Pod [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS HostNetwork spec.Hostname field is silently ignored and the node hostname is used when hostNetwork is set to true for a Pod [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS configMap nameserver Change stubDomain should be able to change stubDomain configuration [Slow] [Serial] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS configMap nameserver Forward PTR lookup should forward PTR records lookup to upstream nameserver [Slow] [Serial] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] DNS configMap nameserver Forward external name lookup should forward externalname lookup to upstream nameserver [Slow] [Serial] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:PerformanceDNS] [Serial] Should answer DNS query for maximum number of services per cluster [Slow] [Suite:k8s]", + "labels": { + "Feature:PerformanceDNS": {}, + "Serial": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] should have ipv4 and ipv6 internal node ip [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] should create pod, add ipv6 and ipv4 ip to pod ips [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] should create pod, add ipv6 and ipv4 ip to host ips [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] should be able to reach pod on ipv4 and ipv6 ip [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] should create a single stack service with cluster ip from primary service range [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] should create service with ipv4 cluster ip [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] should create service with ipv6 cluster ip [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] should create service with ipv4,v6 cluster ip [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] should create service with ipv6,v4 cluster ip [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for pod-Service: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for pod-Service: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for pod-Service: sctp [Feature:SCTPConnectivity] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "Feature:SCTPConnectivity": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for node-Service: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for node-Service: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for endpoint-Service: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for endpoint-Service: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should update endpoints: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should update endpoints: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for client IP based session affinity: http [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for client IP based session affinity: udp [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should be able to handle large requests: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should be able to handle large requests: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for service endpoints using hostNetwork [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:IPv6DualStack": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] EndpointSlice should have Endpoints and EndpointSlices pointing to API Server [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] EndpointSlice should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] EndpointSlice should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] EndpointSlice should support creating EndpointSlice API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] EndpointSlice should support a Service with multiple ports specified in multiple EndpointSlices [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] EndpointSlice should support a Service with multiple endpoint IPs specified in multiple EndpointSlices [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] EndpointSliceMirroring should mirror a custom Endpoints resource through create update and delete [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] EndpointSliceMirroring should mirror a custom Endpoint with multiple subsets and same IP address [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] ClusterDns [Feature:Example] should create pod that uses dns [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Example": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] CVE-2021-29923 IPv4 Service Type ClusterIP with leading zeros should work interpreted as decimal [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] HostPort validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Ingress API should support creating Ingress API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] IngressClass [Feature:Ingress] should set default value on new IngressClass [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Ingress": {}, + "Serial": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] IngressClass [Feature:Ingress] should not set default value if no default IngressClass [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Ingress": {}, + "Serial": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] IngressClass [Feature:Ingress] should choose the one with the later CreationTimestamp, if equal the one with the lower name when two ingressClasses are marked as default [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Ingress": {}, + "Serial": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] IngressClass [Feature:Ingress] should allow IngressClass to have Namespace-scoped parameters [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Ingress": {}, + "Serial": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] IngressClass API should support creating IngressClass API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] KubeProxy should set TCP CLOSE_WAIT timeout [Privileged] [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] KubeProxy should update metric for tracking accepted packets destined for localhost nodeports [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers ExternalTrafficPolicy: Local [Feature:LoadBalancer] [Slow] should work for type=LoadBalancer [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers ExternalTrafficPolicy: Local [Feature:LoadBalancer] [Slow] should only target nodes with endpoints [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers ExternalTrafficPolicy: Local [Feature:LoadBalancer] [Slow] should work from pods [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers ExternalTrafficPolicy: Local [Feature:LoadBalancer] [Slow] should handle updates to ExternalTrafficPolicy field [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking IPerf2 [Feature:Networking-Performance] should run iperf2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Networking-Performance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] NoSNAT Should be able to send traffic between Pods without SNAT [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Connectivity Pod Lifecycle should be able to connect from a Pod to a terminating Pod [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Connectivity Pod Lifecycle should be able to connect to other Pod from a terminating Pod [Disabled:RebaseInProgress] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Proxy version v1 should proxy logs on node using proxy subresource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Proxy version v1 should proxy through a service and a pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Proxy version v1 A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Proxy version v1 A set of valid responses are returned for both pod and service Proxy [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should be able to change the type and ports of a TCP service [Slow] [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should be able to change the type and ports of a UDP service [Slow] [Skipped:alibabacloud] [Skipped:aws] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should only allow access from service loadbalancer source ranges [Slow] [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should have session affinity work for LoadBalancer service with Local traffic policy [Slow] [LinuxOnly] [Skipped:alibabacloud] [Skipped:aws] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should be able to switch session affinity for LoadBalancer service with Local traffic policy [Slow] [LinuxOnly] [Skipped:alibabacloud] [Skipped:aws] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should have session affinity work for LoadBalancer service with Cluster traffic policy [Slow] [LinuxOnly] [Skipped:alibabacloud] [Skipped:aws] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should be able to switch session affinity for LoadBalancer service with Cluster traffic policy [Slow] [LinuxOnly] [Skipped:alibabacloud] [Skipped:aws] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should handle load balancer cleanup finalizer for service [Slow] [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should be able to create LoadBalancer Service without NodePort and change it [Slow] [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should be able to preserve UDP traffic when server pod cycles for a LoadBalancer service on different nodes [Skipped:alibabacloud] [Skipped:aws] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should be able to preserve UDP traffic when server pod cycles for a LoadBalancer service on the same nodes [Skipped:alibabacloud] [Skipped:aws] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should not have connectivity disruption during rolling update with externalTrafficPolicy=Cluster [Slow] [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] LoadBalancers [Feature:LoadBalancer] should not have connectivity disruption during rolling update with externalTrafficPolicy=Local [Slow] [Skipped:alibabacloud] [Skipped:baremetal] [Skipped:ibmcloud] [Skipped:kubevirt] [Skipped:nutanix] [Skipped:openstack] [Skipped:ovirt] [Skipped:vsphere] [Suite:k8s]", + "labels": { + "Feature:LoadBalancer": {}, + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking should provide Internet connection for containers [Feature:Networking-IPv4] [Skipped:Disconnected] [Skipped:Proxy] [Skipped:azure] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Networking-IPv4": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking should provide Internet connection for containers [Feature:Networking-IPv6] [Experimental][LinuxOnly] [Disabled:Broken] [Skipped:Disconnected] [Skipped:Proxy] [Skipped:azure] [Suite:k8s]", + "labels": { + "Feature:Networking-IPv6": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking should provider Internet connection for containers using DNS [Feature:Networking-DNS] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Networking-DNS": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking should provide unchanging, static URL paths for kubernetes api services [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking should check kube-proxy urls [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for pod-Service: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for pod-Service: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for pod-Service: sctp [Feature:SCTPConnectivity] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SCTPConnectivity": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for node-Service: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for node-Service: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for node-Service: sctp [Feature:SCTPConnectivity] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SCTPConnectivity": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for endpoint-Service: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for endpoint-Service: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for endpoint-Service: sctp [Feature:SCTPConnectivity] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SCTPConnectivity": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for multiple endpoint-Services with same selector [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should update endpoints: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should update endpoints: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should update nodePort: http [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should support basic nodePort: udp functionality [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should update nodePort: udp [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for client IP based session affinity: http [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for client IP based session affinity: udp [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should be able to handle large requests: http [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should be able to handle large requests: udp [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking Granular Checks: Services should function for service endpoints using hostNetwork [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking should recreate its iptables rules if they are deleted [Disruptive] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Networking should allow creating a Pod with an SCTP HostPort [LinuxOnly] [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should provide secure master service [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should serve a basic endpoint from pods [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should serve multiport endpoints from pods [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be updated after adding or deleting ports [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should preserve source pod IP for traffic thru service cluster IP [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should allow pods to hairpin back to themselves through services [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to up and down services [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should work after the service has been recreated [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should work after restarting kube-proxy [Disruptive] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should work after restarting apiserver [Disruptive] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to create a functioning NodePort service [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be possible to connect to a service via ExternalIP when the external IP is not assigned to a node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to update service type to NodePort listening on same port number but different protocols [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should prevent NodePort collisions [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should check NodePort out-of-range [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should release NodePorts on delete [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should create endpoints for unready pods [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to connect to terminating and unready endpoints if PublishNotReadyAddresses is true [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should not be able to connect to terminating and unready endpoints if PublishNotReadyAddresses is false [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should have session affinity work for NodePort service [LinuxOnly] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should have session affinity timeout work for NodePort service [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should implement service.kubernetes.io/service-proxy-name [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should implement service.kubernetes.io/headless [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be rejected when no endpoints exist [Skipped:ibmroks] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should be rejected for evicted pods (no endpoints exist) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should respect internalTrafficPolicy=Local Pod to Pod [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should respect internalTrafficPolicy=Local Pod (hostNetwork: true) to Pod [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should respect internalTrafficPolicy=Local Pod and Node, to Pod (hostNetwork: true) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should support externalTrafficPolicy=Local for type=NodePort [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should fail health check node port if there are only terminating endpoints [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should fallback to terminating endpoints when there are no ready endpoints with internalTrafficPolicy=Cluster [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should fallback to local terminating endpoints when there are no ready endpoints with internalTrafficPolicy=Local [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should fallback to terminating endpoints when there are no ready endpoints with externallTrafficPolicy=Cluster [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should fallback to local terminating endpoints when there are no ready endpoints with externalTrafficPolicy=Local [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should find a service from listing all namespaces [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should test the lifecycle of an Endpoint [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should complete a service status lifecycle [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should delete a collection of services [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should serve endpoints on same port and different protocols [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should serve endpoints on same port and different protocol for internal traffic on Type LoadBalancer [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Services should allow creating a basic SCTP service with pod and endpoints [LinuxOnly] [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:ServiceCIDRs] should create Services and servce on different Service CIDRs [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:ServiceCIDRs": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] Service endpoints latency should not be very high [Conformance] [Serial] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:Topology Hints] should distribute endpoints evenly [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:Topology Hints": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] [Feature:Traffic Distribution] when Service has trafficDistribution=PreferClose should route traffic to an endpoint that is close to the client [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Traffic Distribution": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] AppArmor load AppArmor profiles should enforce an AppArmor profile specified on the pod [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] AppArmor load AppArmor profiles should enforce an AppArmor profile specified on the container [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] AppArmor load AppArmor profiles should enforce an AppArmor profile specified in annotations [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] AppArmor load AppArmor profiles can disable an AppArmor profile, using unconfined [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] crictl should be able to run crictl on the node [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [Feature:Example] Liveness liveness pods should be automatically restarted [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Example": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [Feature:Example] Secret should create a pod that reads a secret [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Example": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [Feature:Example] Downward API should create a pod that prints his name and namespace [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Example": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [Feature:GPUDevicePlugin] Sanity test for Nvidia Device should run nvidia-smi cli [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:GPUDevicePlugin": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [Feature:GPUDevicePlugin] Sanity test for Nvidia Device should run gpu based matrix multiplication [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:GPUDevicePlugin": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet Clean up pods on node kubelet should be able to delete 10 pods per node in 1m0s. [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet host cleanup with volume mounts [HostCleanup] [Flaky] Host cleanup after disrupting NFS volume [NFS] after stopping the nfs-server and deleting the (sleeping) client pod, the NFS mount and the pod's UID directory should be removed. [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet host cleanup with volume mounts [HostCleanup] [Flaky] Host cleanup after disrupting NFS volume [NFS] after stopping the nfs-server and deleting the (active) client pod, the NFS mount and the pod's UID directory should be removed. [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet kubectl get --raw \"/api/v1/nodes/\u003cinsert-node-name-here\u003e/proxy/logs/?query=/\u003cinsert-log-file-name-here\u003e [Feature:NodeLogQuery] should return the error with an empty --query option [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:NodeLogQuery": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet kubectl get --raw \"/api/v1/nodes/\u003cinsert-node-name-here\u003e/proxy/logs/?query=/\u003cinsert-log-file-name-here\u003e [Feature:NodeLogQuery] should return the kubelet logs [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:NodeLogQuery": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet kubectl get --raw \"/api/v1/nodes/\u003cinsert-node-name-here\u003e/proxy/logs/?query=/\u003cinsert-log-file-name-here\u003e [Feature:NodeLogQuery] should return the kubelet logs for the current boot [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:NodeLogQuery": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet kubectl get --raw \"/api/v1/nodes/\u003cinsert-node-name-here\u003e/proxy/logs/?query=/\u003cinsert-log-file-name-here\u003e [Feature:NodeLogQuery] should return the last three lines of the kubelet logs [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:NodeLogQuery": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet kubectl get --raw \"/api/v1/nodes/\u003cinsert-node-name-here\u003e/proxy/logs/?query=/\u003cinsert-log-file-name-here\u003e [Feature:NodeLogQuery] should return the kubelet logs for the current boot with the pattern container [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:NodeLogQuery": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet kubectl get --raw \"/api/v1/nodes/\u003cinsert-node-name-here\u003e/proxy/logs/?query=/\u003cinsert-log-file-name-here\u003e [Feature:NodeLogQuery] should return the kubelet logs since the current date and time [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:NodeLogQuery": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet kubectl get --raw \"/api/v1/nodes/\u003cinsert-node-name-here\u003e/proxy/logs/?query=/\u003cinsert-log-file-name-here\u003e [Feature:NodeLogQuery] should return the Microsoft-Windows-Security-SPP logs [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:NodeLogQuery": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet kubectl get --raw \"/api/v1/nodes/\u003cinsert-node-name-here\u003e/proxy/logs/?query=/\u003cinsert-log-file-name-here\u003e [Feature:NodeLogQuery] should return the last three lines of the Microsoft-Windows-Security-SPP logs [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:NodeLogQuery": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] kubelet kubectl get --raw \"/api/v1/nodes/\u003cinsert-node-name-here\u003e/proxy/logs/?query=/\u003cinsert-log-file-name-here\u003e [Feature:NodeLogQuery] should return the Microsoft-Windows-Security-SPP logs with the pattern Health [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:NodeLogQuery": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Kubelet [Serial] [Slow] regular resource usage tracking [Feature:RegularResourceUsageTracking] resource tracking for 0 pods per node [Suite:k8s]", + "labels": { + "Feature:RegularResourceUsageTracking": {}, + "Serial": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Kubelet [Serial] [Slow] regular resource usage tracking [Feature:RegularResourceUsageTracking] resource tracking for 100 pods per node [Suite:k8s]", + "labels": { + "Feature:RegularResourceUsageTracking": {}, + "Serial": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Kubelet [Serial] [Slow] experimental resource usage tracking [Feature:ExperimentalResourceUsageTracking] resource tracking for 100 pods per node [Suite:k8s]", + "labels": { + "Feature:ExperimentalResourceUsageTracking": {}, + "Serial": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Mount propagation should propagate mounts within defined scopes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NodeProblemDetector [NodeFeature:NodeProblemDetector] should run without error [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "NodeFeature:NodeProblemDetector": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod garbage collector [Feature:PodGarbageCollector] [Slow] should handle the creation of 1000 pods [Suite:k8s]", + "labels": { + "Feature:PodGarbageCollector": {}, + "Slow": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] [Serial] Pod InPlace Resize Container (scheduler-focused) [Feature:InPlacePodVerticalScaling] pod-resize-scheduler-tests [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Guaranteed QoS pod, one container - increase CPU \u0026 memory [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Guaranteed QoS pod, one container - decrease CPU \u0026 memory [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Guaranteed QoS pod, one container - increase CPU \u0026 decrease memory [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Guaranteed QoS pod, one container - decrease CPU \u0026 increase memory [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Guaranteed QoS pod, three containers (c1, c2, c3) - increase: CPU (c1,c3), memory (c2) ; decrease: CPU (c2), memory (c1,c3) [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease memory requests only [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease memory limits only [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase memory requests only [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase memory limits only [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU requests only [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU limits only [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase CPU requests only [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase CPU limits only [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU requests and limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase CPU requests and limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU requests and increase CPU limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase CPU requests and decrease CPU limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease memory requests and limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase memory requests and limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease memory requests and increase memory limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase memory requests and decrease memory limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU requests and increase memory limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase CPU requests and decrease memory limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease memory requests and increase CPU limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase memory requests and decrease CPU limits [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container with cpu \u0026 memory requests - decrease memory request [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Guaranteed QoS pod, one container - increase CPU (NotRequired) \u0026 memory (RestartContainer) [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, one container - decrease CPU (RestartContainer) \u0026 memory (NotRequired) [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, three containers - increase c1 resources, no change for c2, decrease c3 resources (no net change for pod) [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, three containers - decrease c1 resources, increase c2 resources, no change for c3 (net increase for pod) [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] Burstable QoS pod, three containers - no change for c1, increase c2 resources, decrease c3 (net decrease for pod) [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] pod-resize-resource-quota-test [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pod InPlace Resize Container [Feature:InPlacePodVerticalScaling] BestEffort pod - try requesting memory, expect error [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:InPlacePodVerticalScaling": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods Extended Delete Grace Period should be submitted and removed [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods Extended Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods Extended Pod Container Status should never report success for a pending container [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods Extended Pod Container Status should never report container start when an init container fails [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods Extended Pod Container lifecycle should not create extra sandbox if all containers are done [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods Extended Pod Container lifecycle evicted pods should be terminal [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Pods Extended Pod TerminationGracePeriodSeconds is negative pod with negative grace period [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] PreStop should call prestop when killing a pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] PreStop graceful pod terminated should wait until preStop hook completes the process [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should reject a Pod requesting a RuntimeClass with conflicting node selector [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should run a Pod requesting a RuntimeClass with scheduling with taints [Serial] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] RuntimeClass should run a Pod requesting a RuntimeClass with scheduling without taints [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support pod.Spec.SecurityContext.SupplementalGroups [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context when if the container's primary UID belongs to some groups in the image [LinuxOnly] should add pod.Spec.SecurityContext.SupplementalGroups to them [LinuxOnly] in resultant supplementary groups for the container processes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context [sig-node] SupplementalGroupsPolicy [Feature:SupplementalGroupsPolicy] when SupplementalGroupsPolicy was not set if the container's primary UID belongs to some groups in the image, it should add SupplementalGroups to them [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SupplementalGroupsPolicy": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context [sig-node] SupplementalGroupsPolicy [Feature:SupplementalGroupsPolicy] when SupplementalGroupsPolicy was set to Merge if the container's primary UID belongs to some groups in the image, it should add SupplementalGroups to them [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SupplementalGroupsPolicy": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context [sig-node] SupplementalGroupsPolicy [Feature:SupplementalGroupsPolicy] when SupplementalGroupsPolicy was set to Strict even if the container's primary UID belongs to some groups in the image, it should not add SupplementalGroups to them [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:SupplementalGroupsPolicy": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support pod.Spec.SecurityContext.RunAsUser [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support container.SecurityContext.RunAsUser [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support volume SELinux relabeling [Flaky] [LinuxOnly] [Suite:k8s]", + "labels": { + "Flaky": {}, + "LinuxOnly": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support volume SELinux relabeling when using hostIPC [Flaky] [LinuxOnly] [Suite:k8s]", + "labels": { + "Flaky": {}, + "LinuxOnly": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support volume SELinux relabeling when using hostPID [Flaky] [LinuxOnly] [Suite:k8s]", + "labels": { + "Flaky": {}, + "LinuxOnly": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support seccomp unconfined on the container [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support seccomp unconfined on the pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support seccomp runtime/default [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] Security Context should support seccomp default which is unconfined [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] SSH should SSH to all nodes and run commands [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NoExecuteTaintManager Single Pod [Serial] evicts pods from tainted nodes [Skipped:SingleReplicaTopology] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NoExecuteTaintManager Single Pod [Serial] doesn't evict pod with tolerations from tainted nodes [Skipped:SingleReplicaTopology] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NoExecuteTaintManager Single Pod [Serial] eventually evict pod with finite tolerations from tainted nodes [Skipped:SingleReplicaTopology] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NoExecuteTaintManager Single Pod [Serial] removing taint cancels eviction [Disruptive] [Conformance] [Skipped:SingleReplicaTopology] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Disruptive": {}, + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NoExecuteTaintManager Single Pod [Serial] pods evicted from tainted nodes have pod disruption condition [Skipped:SingleReplicaTopology] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NoExecuteTaintManager Multiple Pods [Serial] only evicts pods without tolerations from tainted nodes [Skipped:SingleReplicaTopology] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-node] NoExecuteTaintManager Multiple Pods [Serial] evicts pods with minTolerationSeconds [Disruptive] [Conformance] [Skipped:SingleReplicaTopology] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Disruptive": {}, + "Serial": {}, + "sig-node": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSIInlineVolumes should support CSIVolumeSource in Pod API [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSIInlineVolumes should run through the lifecycle of a CSIDriver [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable-stress [Feature:VolumeSnapshotDataSource] should support snapshotting of many volumes repeatedly [Slow] [Serial] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable-stress [Feature:VolumeSnapshotDataSource] should support snapshotting of many volumes repeatedly [Slow] [Serial] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volume-lifecycle-performance should provision volumes at scale within performance constraints [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should preempt lower priority pods using ReadWriteOncePod volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "MinimumKubeletVersion:1.27": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should block a second pod from using an in-use ReadWriteOncePod volume on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "MinimumKubeletVersion:1.27": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should create a volume with VAC [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume with no VAC [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume that already has a VAC [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should create a volume with VAC [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume with no VAC [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume that already has a VAC [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should create a volume with VAC [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "Feature:Windows": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume with no VAC [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "Feature:Windows": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume that already has a VAC [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "Feature:Windows": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ext3)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ext4)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read-only inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read/write inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support two pods which have the same volume definition [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support multiple inline ephemeral volumes [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent) [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable-stress [Feature:VolumeSnapshotDataSource] should support snapshotting of many volumes repeatedly [Slow] [Serial] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable-stress [Feature:VolumeSnapshotDataSource] should support snapshotting of many volumes repeatedly [Slow] [Serial] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volume-lifecycle-performance should provision volumes at scale within performance constraints [Slow] [Serial] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should preempt lower priority pods using ReadWriteOncePod volumes [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "MinimumKubeletVersion:1.27": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should block a second pod from using an in-use ReadWriteOncePod volume on the same node [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "MinimumKubeletVersion:1.27": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should create a volume with VAC [Disabled:Alpha] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume with no VAC [Disabled:Alpha] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume that already has a VAC [Disabled:Alpha] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should create a volume with VAC [Disabled:Alpha] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume with no VAC [Disabled:Alpha] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume that already has a VAC [Disabled:Alpha] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should create a volume with VAC [Disabled:Alpha] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "Feature:Windows": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume with no VAC [Disabled:Alpha] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "Feature:Windows": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should modify volume that already has a VAC [Disabled:Alpha] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "Feature:Windows": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSIStorageCapacity should support CSIStorageCapacities API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Feature:Flexvolumes] Detaching volumes should not work when mount is in progress [Slow] [Suite:k8s]", + "labels": { + "Feature:Flexvolumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir wrapper volumes should not conflict [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] EmptyDir wrapper volumes should not cause race condition when used for git_repo [Serial] [Slow] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : secret [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : configmap [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : projected [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Flexvolumes should be mountable when non-attachable [Skipped:gce] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Flexvolumes should be mountable when attachable [Feature:Flexvolumes] [Skipped:gce] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Flexvolumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Feature:Flexvolumes] Mounted flexvolume expand [Slow] Should verify mounted flex volumes can be resized [Suite:k8s]", + "labels": { + "Feature:Flexvolumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Feature:Flexvolumes] Mounted flexvolume volume expand [Slow] should be resizable when mounted [Suite:k8s]", + "labels": { + "Feature:Flexvolumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] GenericPersistentVolume [Disruptive] When kubelet restarts Should test that a file written to the mount before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] GenericPersistentVolume [Disruptive] When kubelet restarts Should test that a volume mounted to a pod that is deleted while the kubelet is down unmounts when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] GenericPersistentVolume [Disruptive] When kubelet restarts Should test that a volume mounted to a pod that is force deleted while the kubelet is down unmounts when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Directory [Slow] Should fail on mounting non-existent directory 'does-not-exist-dir' when HostPathType is HostPathDirectory [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Directory [Slow] Should be able to mount directory 'adir' successfully when HostPathType is HostPathDirectory [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Directory [Slow] Should be able to mount directory 'adir' successfully when HostPathType is HostPathUnset [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Directory [Slow] Should fail on mounting directory 'adir' when HostPathType is HostPathFile [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Directory [Slow] Should fail on mounting directory 'adir' when HostPathType is HostPathSocket [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Directory [Slow] Should fail on mounting directory 'adir' when HostPathType is HostPathCharDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Directory [Slow] Should fail on mounting directory 'adir' when HostPathType is HostPathBlockDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType File [Slow] Should fail on mounting non-existent file 'does-not-exist-file' when HostPathType is HostPathFile [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType File [Slow] Should be able to mount file 'afile' successfully when HostPathType is HostPathFile [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType File [Slow] Should be able to mount file 'afile' successfully when HostPathType is HostPathUnset [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType File [Slow] Should fail on mounting file 'afile' when HostPathType is HostPathDirectory [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType File [Slow] Should fail on mounting file 'afile' when HostPathType is HostPathSocket [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType File [Slow] Should fail on mounting file 'afile' when HostPathType is HostPathCharDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType File [Slow] Should fail on mounting file 'afile' when HostPathType is HostPathBlockDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Socket [Slow] Should fail on mounting non-existent socket 'does-not-exist-socket' when HostPathType is HostPathSocket [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Socket [Slow] Should be able to mount socket 'asocket' successfully when HostPathType is HostPathSocket [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Socket [Slow] Should be able to mount socket 'asocket' successfully when HostPathType is HostPathUnset [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Socket [Slow] Should fail on mounting socket 'asocket' when HostPathType is HostPathDirectory [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Socket [Slow] Should fail on mounting socket 'asocket' when HostPathType is HostPathFile [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Socket [Slow] Should fail on mounting socket 'asocket' when HostPathType is HostPathCharDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Socket [Slow] Should fail on mounting socket 'asocket' when HostPathType is HostPathBlockDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Character Device [Slow] Should fail on mounting non-existent character device 'does-not-exist-char-dev' when HostPathType is HostPathCharDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Character Device [Slow] Should be able to mount character device 'achardev' successfully when HostPathType is HostPathCharDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Character Device [Slow] Should be able to mount character device 'achardev' successfully when HostPathType is HostPathUnset [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Character Device [Slow] Should fail on mounting character device 'achardev' when HostPathType is HostPathDirectory [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Character Device [Slow] Should fail on mounting character device 'achardev' when HostPathType is HostPathFile [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Character Device [Slow] Should fail on mounting character device 'achardev' when HostPathType is HostPathSocket [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Character Device [Slow] Should fail on mounting character device 'achardev' when HostPathType is HostPathBlockDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Block Device [Slow] Should fail on mounting non-existent block device 'does-not-exist-blk-dev' when HostPathType is HostPathBlockDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Block Device [Slow] Should be able to mount block device 'ablkdev' successfully when HostPathType is HostPathBlockDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Block Device [Slow] Should be able to mount block device 'ablkdev' successfully when HostPathType is HostPathUnset [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Block Device [Slow] Should fail on mounting block device 'ablkdev' when HostPathType is HostPathDirectory [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Block Device [Slow] Should fail on mounting block device 'ablkdev' when HostPathType is HostPathFile [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Block Device [Slow] Should fail on mounting block device 'ablkdev' when HostPathType is HostPathSocket [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] HostPathType Block Device [Slow] Should fail on mounting block device 'ablkdev' when HostPathType is HostPathCharDev [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (ext3)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext3)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (ext4)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Disabled:Broken] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Disabled:Broken] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Volumes": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: iscsi] [Feature:Volumes] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Disabled:Broken] [Suite:k8s]", + "labels": { + "Feature:Volumes": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Disabled:Unsupported] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Disabled:Unsupported] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Disabled:Unsupported] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Disabled:Unsupported] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Skipped:azure] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Skipped:azure] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext3)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext3)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext4)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (xfs)] [Slow] volumes should store data [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (xfs)] [Slow] volumes should allow exec of files on the volume [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should store data [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (xfs)] [Slow] volumes should allow exec of files on the volume [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should store data [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] volumes should allow exec of files on the volume [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volumes should store data [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volumeIO should write files of various sizes, verify size, validate content [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumeIO should write files of various sizes, verify size, validate content [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to create pod by failing to mount volume [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail in binding dynamic provisioned PV to PVC [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should fail to use a volume in a pod with mismatched mode [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directory [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using directory as subpath [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should verify container cannot write to subpath readonly volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath directory is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath file is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if non-existent subpath is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should fail if subpath with backstepping is outside the volume [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support creating multiple subpath from same volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using directory as subpath [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support restarting containers using file as subpath [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should unmount if pod is force deleted while kubelet is down [Disruptive] [Slow] [LinuxOnly] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should verify container cannot write to subpath readonly volumes [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode) [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source in parallel [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode) [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source in parallel [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode) [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source in parallel [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (xfs)] [Slow] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with the same volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should access to two volumes with different volume mode and retain data across pod recreation on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and restored snapshot from pods on the same node [LinuxOnly] [Feature:VolumeSnapshotDataSource] [Feature:VolumeSourceXFS] [Skipped:NoOptionalCapabilities] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the volume and its clone from pods on the same node [LinuxOnly] [Feature:VolumeSourceXFS] [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:VolumeSourceXFS": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single read-only volume from pods on the same node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] multiVolume [Slow] should concurrently access the single volume from pods on different node [Skipped:gce] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down is usable by a new pod with a different SELinux context when kubelet returns [Feature:SELinux] [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:SELinux": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv written before kubelet restart is readable after restart. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] disruptive [Disruptive] [LinuxOnly] Should test that pv used in a pod that is force deleted while the kubelet is down cleans up when the kubelet returns. [Skipped:gce] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "LinuxOnly": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial] [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-stress multiple pods should access different volumes repeatedly [Slow] [Serial] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Mounted volume expand [Feature:StorageProvider] Should verify mounted devices can be resized [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:StorageProvider": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] NFSPersistentVolumes [Disruptive] [Flaky] when kube-controller-manager restarts should delete a bound PVC from a clientPod, restart the kube-control-manager, and ensure the kube-controller-manager does not crash [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] NFSPersistentVolumes [Disruptive] [Flaky] when kubelet restarts Should test that a file written to the mount before kubelet restart is readable after restart. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] NFSPersistentVolumes [Disruptive] [Flaky] when kubelet restarts Should test that a volume mounted to a pod that is deleted while the kubelet is down unmounts when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] NFSPersistentVolumes [Disruptive] [Flaky] when kubelet restarts Should test that a volume mounted to a pod that is force deleted while the kubelet is down unmounts when the kubelet returns. [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Feature:NodeOutOfServiceVolumeDetach] [Disruptive] [LinuxOnly] NonGracefulNodeShutdown [NonGracefulNodeShutdown] pod that uses a persistent volume via gce pd driver should get immediately rescheduled to a different node after non graceful node shutdown [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:NodeOutOfServiceVolumeDetach": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-expansion loopback local block volume should support online expansion on node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] One pod requesting one prebound PVC should be able to mount volume and read from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] One pod requesting one prebound PVC should be able to mount volume and write from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] Set fsGroup for local volume should set fsGroup for one pod [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted [Flaky] [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] One pod requesting one prebound PVC should be able to mount volume and read from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] One pod requesting one prebound PVC should be able to mount volume and write from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] Set fsGroup for local volume should set fsGroup for one pod [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted [Flaky] [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] One pod requesting one prebound PVC should be able to mount volume and read from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] One pod requesting one prebound PVC should be able to mount volume and write from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] Set fsGroup for local volume should set fsGroup for one pod [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted [Flaky] [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] One pod requesting one prebound PVC should be able to mount volume and read from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] One pod requesting one prebound PVC should be able to mount volume and write from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] Set fsGroup for local volume should set fsGroup for one pod [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted [Flaky] [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] One pod requesting one prebound PVC should be able to mount volume and read from pod1 [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] One pod requesting one prebound PVC should be able to mount volume and write from pod1 [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2 [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2 [Skipped:gce] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] Set fsGroup for local volume should set fsGroup for one pod [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow] [Skipped:gce] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted [Flaky] [Skipped:gce] [Suite:k8s]", + "labels": { + "Flaky": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir] One pod requesting one prebound PVC should be able to mount volume and read from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir] One pod requesting one prebound PVC should be able to mount volume and write from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir] Set fsGroup for local volume should set fsGroup for one pod [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted [Flaky] [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] One pod requesting one prebound PVC should be able to mount volume and read from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] One pod requesting one prebound PVC should be able to mount volume and write from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] Set fsGroup for local volume should set fsGroup for one pod [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted [Flaky] [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] One pod requesting one prebound PVC should be able to mount volume and read from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] One pod requesting one prebound PVC should be able to mount volume and write from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] Set fsGroup for local volume should set fsGroup for one pod [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted [Flaky] [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: block] One pod requesting one prebound PVC should be able to mount volume and read from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: block] One pod requesting one prebound PVC should be able to mount volume and write from pod1 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: block] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: block] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2 [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: block] Set fsGroup for local volume should set fsGroup for one pod [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: block] Set fsGroup for local volume should set same fsGroup for two pods simultaneously [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local [Volume type: block] Set fsGroup for local volume should set different fsGroup for second pod if first pod is deleted [Flaky] [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local Local volume that cannot be mounted [Slow] should fail due to non-existent path [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local Local volume that cannot be mounted [Slow] should fail due to wrong node [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local Pod with node different from PV's NodeAffinity should fail scheduling due to different NodeAffinity [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local Pod with node different from PV's NodeAffinity should fail scheduling due to different NodeSelector [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local StatefulSet with pod affinity [Slow] should use volumes spread across nodes when pod has anti-affinity [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local StatefulSet with pod affinity [Slow] should use volumes on one node when pod has affinity [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local StatefulSet with pod affinity [Slow] should use volumes spread across nodes when pod management is parallel and pod has anti-affinity [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local StatefulSet with pod affinity [Slow] should use volumes on one node when pod management is parallel and pod has affinity [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes-local Stress with local volumes [Serial] should be able to process many pods and reuse local volumes [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs should create a non-pre-bound PV and PVC: test write access [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PVC and non-pre-bound PV: test write access [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PVC and a pre-bound PV: test write access [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PV and a pre-bound PVC: test write access [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PV: test phase transition timestamp is set and phase is Available [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PV and a pre-bound PVC: test phase transition timestamp is set [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PV and a pre-bound PVC: test phase transition timestamp multiple updates [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with multiple PVs and PVCs all in same ns should create 2 PVs and 4 PVCs: test write access [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with multiple PVs and PVCs all in same ns should create 3 PVs and 3 PVCs: test write access [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS with multiple PVs and PVCs all in same ns should create 4 PVs and 2 PVCs: test write access [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes NFS when invoking the Recycle reclaim policy should test that a PV becomes Available and is clean after the PVC is deleted. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes CSI Conformance should run through the lifecycle of a PV and a PVC [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes CSI Conformance should apply changes to a pv/pvc status [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PersistentVolumes Default StorageClass [LinuxOnly] pods that use multiple volumes should be reschedulable [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PV Protection Verify \"immediate\" deletion of a PV that is not bound to a PVC [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PV Protection Verify that PV bound to a PVC is not removed immediately [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PVC Protection Verify \"immediate\" deletion of a PVC that is not in active use by a pod [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PVC Protection Verify that PVC in active use by a pod is not removed immediately [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] PVC Protection Verify that scheduling of a pod that uses PVC that is being deleted fails and the pod becomes Unschedulable [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Persistent Volume Claim and StorageClass Retroactive StorageClass assignment [Serial] [Disruptive] should assign default SC to PVCs that have no SC set [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] StaticPods [Feature:Kind] should run after kubelet stopped with CSI volume mounted [Disruptive] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Kind": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] StorageClasses CSI Conformance should run through the lifecycle of a StorageClass [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Subpath Container restart should verify that container can restart successfully after configmaps modified [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] VolumeAttachment Conformance should run through the lifecycle of a VolumeAttachment [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVC should create prometheus metrics for volume provisioning and attach/detach [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVC should create prometheus metrics for volume provisioning errors [Slow] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVC should create volume metrics with the correct FilesystemMode PVC ref [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVC should create volume metrics with the correct BlockMode PVC ref [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVC should create metrics for total time taken in volume operations in P/V Controller [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVC should create volume metrics in Volume Manager [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVC should create metrics for total number of volumes in A/D Controller [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics Ephemeral should create prometheus metrics for volume provisioning and attach/detach [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics Ephemeral should create prometheus metrics for volume provisioning errors [Slow] [Suite:k8s]", + "labels": { + "Serial": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics Ephemeral should create volume metrics with the correct FilesystemMode PVC ref [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics Ephemeral should create volume metrics with the correct BlockMode PVC ref [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics Ephemeral should create metrics for total time taken in volume operations in P/V Controller [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics Ephemeral should create volume metrics in Volume Manager [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics Ephemeral should create metrics for total number of volumes in A/D Controller [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVController should create none metrics for pvc controller before creating any PV or PVC [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVController should create unbound pv count metrics for pvc controller after creating pv only [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVController should create unbound pvc count metrics for pvc controller after creating pvc only [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVController should create bound pv/pvc count metrics for pvc controller after creating both pv and pvc [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVController should create unbound pvc count metrics for pvc controller with volume attributes class dimension after creating pvc only [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVController should create bound pv/pvc count metrics for pvc controller with volume attributes class dimension after creating both pv and pvc [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] [Serial] Volume metrics PVController should create total pv count metrics for with plugin and volume mode labels after creating pv [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Dynamic Provisioning DynamicProvisioner [Slow] [Feature:StorageProvider] should provision storage with different parameters [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:StorageProvider": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Dynamic Provisioning DynamicProvisioner [Slow] [Feature:StorageProvider] should provision storage with non-default reclaim policy Retain [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:StorageProvider": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Dynamic Provisioning DynamicProvisioner [Slow] [Feature:StorageProvider] should test that deleting a claim before the volume is provisioned deletes the volume. [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:StorageProvider": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Dynamic Provisioning DynamicProvisioner [Slow] [Feature:StorageProvider] deletion should be idempotent [Skipped:NoOptionalCapabilities] [Suite:k8s]", + "labels": { + "Feature:StorageProvider": {}, + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Dynamic Provisioning DynamicProvisioner External should let an external dynamic provisioner create and delete persistent volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Dynamic Provisioning DynamicProvisioner Default should create and delete default persistent volumes [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Dynamic Provisioning DynamicProvisioner Default should be disabled by changing the default annotation [Serial] [Disruptive] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Dynamic Provisioning DynamicProvisioner Default should be disabled by removing the default annotation [Serial] [Disruptive] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Dynamic Provisioning Invalid AWS KMS key should report an error and create no PV [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] VolumeAttributesClass [Feature:VolumeAttributesClass] [FeatureGate:VolumeAttributesClass] [Beta] should run through the lifecycle of a VolumeAttributesClass [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:VolumeAttributesClass": {}, + "FeatureGate:VolumeAttributesClass": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] Volumes ConfigMap should be mountable [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume attach CSI attach test using mock driver should not require VolumeAttach for drivers without attachment [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume attach CSI attach test using mock driver should require VolumeAttach for drivers with attachment [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume attach CSI attach test using mock driver should require VolumeAttach for ephemermal volume and drivers with attachment [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume attach CSI attach test using mock driver should preserve attachment policy when no CSIDriver present [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume attach CSI CSIDriver deployment after pod creation using non-attachable mock driver should bringup pod after deploying CSIDriver attach=false [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock fsgroup as mount option Delegate FSGroup to CSI driver [LinuxOnly] should pass FSGroup to CSI driver if it is set in pod and driver supports VOLUME_MOUNT_GROUP [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock fsgroup as mount option Delegate FSGroup to CSI driver [LinuxOnly] should not pass FSGroup to CSI driver if it is set in pod and driver supports VOLUME_MOUNT_GROUP [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy [LinuxOnly] should modify fsGroup if fsGroupPolicy=default [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy [LinuxOnly] should modify fsGroup if fsGroupPolicy=File [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy [LinuxOnly] should not modify fsGroup if fsGroupPolicy=None [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should update fsGroup if update from None to File [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should update fsGroup if update from None to default [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should not update fsGroup if update from File to None [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should update fsGroup if update from File to default [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should not update fsGroup if update from detault to None [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should update fsGroup if update from detault to File [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock honor pv reclaim policy [Feature:HonorPVReclaimPolicy] [FeatureGate:HonorPVReclaimPolicy] [Beta] CSI honor pv reclaim policy using mock driver Dynamic provisioning should honor pv delete reclaim policy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:HonorPVReclaimPolicy": {}, + "FeatureGate:HonorPVReclaimPolicy": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock honor pv reclaim policy [Feature:HonorPVReclaimPolicy] [FeatureGate:HonorPVReclaimPolicy] [Beta] CSI honor pv reclaim policy using mock driver Dynamic provisioning should honor pv retain reclaim policy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:HonorPVReclaimPolicy": {}, + "FeatureGate:HonorPVReclaimPolicy": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock honor pv reclaim policy [Feature:HonorPVReclaimPolicy] [FeatureGate:HonorPVReclaimPolicy] [Beta] CSI honor pv reclaim policy using mock driver Static provisioning should honor pv delete reclaim policy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:HonorPVReclaimPolicy": {}, + "FeatureGate:HonorPVReclaimPolicy": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock honor pv reclaim policy [Feature:HonorPVReclaimPolicy] [FeatureGate:HonorPVReclaimPolicy] [Beta] CSI honor pv reclaim policy using mock driver Static provisioning should honor pv retain reclaim policy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:HonorPVReclaimPolicy": {}, + "FeatureGate:HonorPVReclaimPolicy": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume node stage CSI NodeStage error cases [Slow] should call NodeUnstage after NodeStage success [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume node stage CSI NodeStage error cases [Slow] should retry NodeStage after NodeStage final error [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume node stage CSI NodeStage error cases [Slow] should retry NodeStage after NodeStage ephemeral error [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume node stage CSI NodeStage error cases [Slow] should call NodeUnstage after NodeStage ephemeral error [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume node stage CSI NodeStage error cases [Slow] should not call NodeUnstage after NodeStage final error [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume node stage CSI NodeUnstage error cases [Slow] should call NodeStage after NodeUnstage success [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume node stage CSI NodeUnstage error cases [Slow] two pods: should call NodeStage after previous NodeUnstage final error [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume node stage CSI NodeUnstage error cases [Slow] two pods: should call NodeStage after previous NodeUnstage transient error [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should pass SELinux mount option for RWOP volume and Pod with SELinux context set [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should add SELinux mount option to existing mount options [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not pass SELinux mount option for RWO volume with SELinuxMount disabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Feature:SELinuxMountReadWriteOncePodOnly] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "Feature:SELinuxMountReadWriteOncePodOnly": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should pass SELinux mount option for RWO volume with SELinuxMount enabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Alpha] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Alpha": {}, + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMount": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not pass SELinux mount option for Pod without SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not pass SELinux mount option for CSI driver that does not support SELinux mount [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not unstage RWOP volume when starting a second pod with the same SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should unstage RWOP volume when starting a second pod with different SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not unstage RWO volume when starting a second pod with the same SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Alpha] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Alpha": {}, + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMount": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should unstage RWO volume when starting a second pod with different SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Alpha] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Alpha": {}, + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMount": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount metrics SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] warning is not bumped on two Pods with the same context on RWO volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Feature:SELinuxMountReadWriteOncePodOnly] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "Feature:SELinuxMountReadWriteOncePodOnly": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount metrics SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] warning is bumped on two Pods with a different context on RWO volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Feature:SELinuxMountReadWriteOncePodOnly] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "Feature:SELinuxMountReadWriteOncePodOnly": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount metrics SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is not bumped on two Pods with the same context on RWO volume and SELinuxMount enabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Alpha] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Alpha": {}, + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMount": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount metrics SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with a different context on RWO volume and SELinuxMount enabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Alpha] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Alpha": {}, + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMount": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount metrics SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with a different context on RWX volume and SELinuxMount enabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Alpha] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:Alpha": {}, + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMount": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock selinux on mount metrics SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with a different context on RWOP volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "Feature:SELinux": {}, + "FeatureGate:SELinuxMountReadWriteOncePod": {}, + "Serial": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume service account token CSIServiceAccountToken token should not be plumbed down when csiServiceAccountTokenEnabled=false [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume service account token CSIServiceAccountToken token should not be plumbed down when CSIDriver is not deployed [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume service account token CSIServiceAccountToken token should be plumbed down when csiServiceAccountTokenEnabled=true [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume snapshot CSI Volume Snapshots [Feature:VolumeSnapshotDataSource] volumesnapshotcontent and pvc in Bound state with deletion timestamp set should not get deleted while snapshot finalizer exists [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume snapshot CSI Volume Snapshots secrets [Feature:VolumeSnapshotDataSource] volume snapshot create/delete with secrets [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume snapshot CSI Snapshot Controller metrics [Feature:VolumeSnapshotDataSource] snapshot controller should emit dynamic CreateSnapshot, CreateSnapshotAndReady, and DeleteSnapshot metrics [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume snapshot CSI Snapshot Controller metrics [Feature:VolumeSnapshotDataSource] snapshot controller should emit pre-provisioned CreateSnapshot, CreateSnapshotAndReady, and DeleteSnapshot metrics [Skipped:NoOptionalCapabilities] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:VolumeSnapshotDataSource": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume storage capacity storage capacity unlimited [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume storage capacity storage capacity exhausted, immediate binding [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume storage capacity storage capacity exhausted, late binding, no topology [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume storage capacity storage capacity exhausted, late binding, with topology [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity unused [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity disabled [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity used, no capacity [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity used, insufficient capacity [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity used, have capacity [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should expand volume without restarting pod if nodeExpansion=off [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should expand volume by restarting pod if attach=on, nodeExpansion=on [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should not have staging_path missing in node expand volume pod if attach=on, nodeExpansion=on [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should expand volume by restarting pod if attach=off, nodeExpansion=on [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should not expand volume if resizingOnDriver=off, resizingOnSC=on [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion CSI online volume expansion with secret should expand volume without restarting pod if attach=on, nodeExpansion=on, csiNodeExpandSecret=on [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion CSI online volume expansion should expand volume without restarting pod if attach=on, nodeExpansion=on [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion CSI online volume expansion should expand volume without restarting pod if attach=off, nodeExpansion=on [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion Expansion with recovery [Feature:RecoverVolumeExpansionFailure] should record target size in allocated resources [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:RecoverVolumeExpansionFailure": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion Expansion with recovery [Feature:RecoverVolumeExpansionFailure] should allow recovery if controller expansion fails with final error [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:RecoverVolumeExpansionFailure": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume expansion Expansion with recovery [Feature:RecoverVolumeExpansionFailure] recovery should not be possible in partially expanded volumes [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:RecoverVolumeExpansionFailure": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume limit CSI volume limit information using mock driver should report attach limit when limit is bigger than 0 [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume limit CSI volume limit information using mock driver should report attach limit for generic ephemeral volume when persistent volume is attached [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock volume limit CSI volume limit information using mock driver should report attach limit for persistent volume when generic ephemeral volume is attached [Slow] [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock workload info CSI workload information using mock driver should be passed when podInfoOnMount=true [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock workload info CSI workload information using mock driver contain ephemeral=true when using inline volume [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock workload info CSI workload information using mock driver should not be passed when podInfoOnMount=false [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock workload info CSI workload information using mock driver should not be passed when CSIDriver does not exist [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock workload info CSI PodInfoOnMount Update should not be passed when update from true to false [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-storage] CSI Mock workload info CSI PodInfoOnMount Update should be passed when update from false to true [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-storage": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-apps] stateful Upgrade [Feature:StatefulUpgrade] stateful upgrade should maintain a functioning cluster [Disabled:Unimplemented] [Suite:k8s]", + "labels": { + "Feature:StatefulUpgrade": {}, + "sig-apps": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-auth] ServiceAccount admission controller migration [Feature:BoundServiceAccountTokenVolume] master upgrade should maintain a functioning cluster [Disabled:Unimplemented] [Suite:k8s]", + "labels": { + "Feature:BoundServiceAccountTokenVolume": {}, + "sig-auth": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] kube-proxy migration [Feature:KubeProxyDaemonSetMigration] Upgrade kube-proxy from static pods to a DaemonSet should maintain a functioning cluster [Feature:KubeProxyDaemonSetUpgrade] [Disabled:Unimplemented] [Suite:k8s]", + "labels": { + "Feature:KubeProxyDaemonSetMigration": {}, + "Feature:KubeProxyDaemonSetUpgrade": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-network] kube-proxy migration [Feature:KubeProxyDaemonSetMigration] Downgrade kube-proxy from a DaemonSet to static pods should maintain a functioning cluster [Feature:KubeProxyDaemonSetDowngrade] [Disabled:Unimplemented] [Suite:k8s]", + "labels": { + "Feature:KubeProxyDaemonSetDowngrade": {}, + "Feature:KubeProxyDaemonSetMigration": {}, + "sig-network": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Upgrade [Feature:Upgrade] master upgrade should maintain a functioning cluster [Feature:MasterUpgrade] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:MasterUpgrade": {}, + "Feature:Upgrade": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Upgrade [Feature:Upgrade] cluster upgrade should maintain a functioning cluster [Feature:ClusterUpgrade] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:ClusterUpgrade": {}, + "Feature:Upgrade": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Downgrade [Feature:Downgrade] cluster downgrade should maintain a functioning cluster [Feature:ClusterDowngrade] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:ClusterDowngrade": {}, + "Feature:Downgrade": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] GKE node pools [Feature:GKENodePool] should create a cluster with multiple node pools [Feature:GKENodePool] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:GKENodePool": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] HA-master [Feature:HAMaster] survive addition/removal replicas same zone [Serial] [Disruptive] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:HAMaster": {}, + "Serial": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] HA-master [Feature:HAMaster] survive addition/removal replicas different zones [Serial] [Disruptive] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:HAMaster": {}, + "Serial": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] HA-master [Feature:HAMaster] survive addition/removal replicas multizone workers [Serial] [Disruptive] [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:HAMaster": {}, + "Serial": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Ports Security Check [Feature:KubeletSecurity] should not be able to proxy to the readonly kubelet port 10255 using proxy subresource [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:KubeletSecurity": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Ports Security Check [Feature:KubeletSecurity] should not be able to proxy to cadvisor port 4194 using proxy subresource [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:KubeletSecurity": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Ports Security Check [Feature:KubeletSecurity] should not have port 10255 open on its all public IP addresses [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:KubeletSecurity": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Ports Security Check [Feature:KubeletSecurity] should not have port 4194 open on its all public IP addresses [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:KubeletSecurity": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] [Disruptive] NodeLease NodeLease deletion node lease should be deleted when corresponding node is deleted [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Reboot [Disruptive] [Feature:Reboot] each node by ordering clean reboot and ensure they function upon restart [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Reboot": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Reboot [Disruptive] [Feature:Reboot] each node by ordering unclean reboot and ensure they function upon restart [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Reboot": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Reboot [Disruptive] [Feature:Reboot] each node by triggering kernel panic and ensure they function upon restart [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Reboot": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Reboot [Disruptive] [Feature:Reboot] each node by switching off the network interface and ensure they function upon switch on [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Reboot": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Reboot [Disruptive] [Feature:Reboot] each node by dropping all inbound packets for a while and ensure they function afterwards [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Reboot": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Reboot [Disruptive] [Feature:Reboot] each node by dropping all outbound packets for a while and ensure they function afterwards [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Reboot": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Nodes [Disruptive] Resize [Slow] should be able to delete nodes [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Nodes [Disruptive] Resize [Slow] should be able to add nodes [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Slow": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider-gcp] Restart [Disruptive] [KubeUp] should restart all nodes and ensure all nodes and pods recover [Disabled:SpecialConfig] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "KubeUp": {}, + "sig-cloud-provider-gcp": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cloud-provider] [Feature:CloudProvider] [Disruptive] Nodes should be deleted on API server if it doesn't exist in the cloud provider [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:CloudProvider": {}, + "sig-cloud-provider": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Cpu Resources [Serial] Container limits should not be exceeded after waiting 2 minutes [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Density [Serial] [Slow] create a batch of pods latency/resource should be within limit when create 10 pods with 0s interval [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:GPUDevicePlugin] Device Plugin should be able to create a functioning device plugin for Windows [Disabled:SpecialConfig] [Suite:k8s]", + "labels": { + "Feature:GPUDevicePlugin": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] DNS should support configurable pod DNS servers [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Eviction [Serial] [Slow] [Disruptive] should evict a pod when a node experiences memory pressure [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] GMSA Full [Serial] [Slow] GMSA support works end to end [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] GMSA Full [Serial] [Slow] GMSA support can read and write file to remote SMB folder [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] GMSA Kubelet [Slow] kubelet GMSA support when creating a pod with correct GMSA credential specs passes the credential specs down to the Pod's containers [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Slow": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should run as a process on the host/node [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:WindowsHostProcessContainers": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should support init containers [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:WindowsHostProcessContainers": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers container command path validation [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:WindowsHostProcessContainers": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should support various volume mount types [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:WindowsHostProcessContainers": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers metrics should report count of started and failed to start HostProcess containers [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:WindowsHostProcessContainers": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers container stats validation [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:WindowsHostProcessContainers": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should support querying api-server using in-cluster config [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:WindowsHostProcessContainers": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should run as localgroup accounts [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:WindowsHostProcessContainers": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] Hybrid cluster network for all supported CNIs should have stable networking for Linux and Windows pods [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] Hybrid cluster network for all supported CNIs should provide Internet connection for Linux containers [Feature:Networking-IPv4] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Networking-IPv4": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] Hybrid cluster network for all supported CNIs should provide Internet connection and DNS for Windows containers [Feature:Networking-IPv4] [Feature:Networking-DNS] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Networking-DNS": {}, + "Feature:Networking-IPv4": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:WindowsHyperVContainers] HyperV containers should start a hyperv isolated container [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:WindowsHyperVContainers": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Kubelet-Stats [Serial] Kubelet stats collection for Windows nodes when running 10 pods should return within 10 seconds [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Kubelet-Stats Kubelet stats collection for Windows nodes when windows is booted should return bootid within 10 seconds [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Kubelet-Stats Kubelet stats collection for Windows nodes when running 3 pods should return within 10 seconds [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Memory Limits [Serial] [Slow] Allocatable node memory should be equal to a calculated allocatable memory value [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Memory Limits [Serial] [Slow] attempt to deploy past allocatable memory limits should fail deployments of pods once there isn't enough memory [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] [Excluded:WindowsDocker] [MinimumKubeletVersion:1.22] RebootHost containers [Serial] [Disruptive] [Slow] should run as a reboot process on the host/node [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:Windows": {}, + "Serial": {}, + "Slow": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] SecurityContext should be able create pods and run containers with a given username [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] SecurityContext should not be able to create pods with unknown usernames at Pod level [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] SecurityContext should not be able to create pods with unknown usernames at Container level [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] SecurityContext should override SecurityContext username if set [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] SecurityContext should ignore Linux Specific SecurityContext if set [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] SecurityContext should not be able to create pods with containers running as ContainerAdministrator when runAsNonRoot is true [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] SecurityContext should not be able to create pods with containers running as CONTAINERADMINISTRATOR when runAsNonRoot is true [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] SecurityContext should be able to create pod and run containers [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] Services should be able to create a functioning NodePort service for Windows [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Windows volume mounts check volume mount permissions container should have readOnly permissions on emptyDir [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-windows] [Feature:Windows] Windows volume mounts check volume mount permissions container should have readOnly permissions on hostMapPath [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Windows": {}, + "sig-windows": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AggregatedDiscovery should support raw aggregated discovery endpoint Accept headers [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AggregatedDiscovery should support raw aggregated discovery request for CRDs [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AggregatedDiscovery should support aggregated discovery interface [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AggregatedDiscovery should support aggregated discovery interface for CRDs [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] kube-apiserver identity [Feature:APIServerIdentity] kube-apiserver identity should persist after restart [Disruptive] [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "Feature:APIServerIdentity": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ServerSideApply should create an applied object if it does not already exist [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ServerSideApply should work for subresources [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ServerSideApply should remove a field if it is owned but removed in the apply request [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ServerSideApply should not remove a field if an owner unsets the field but other managers still have ownership of the field [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ServerSideApply should ignore conflict errors if force apply is used [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ServerSideApply should work for CRDs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ServerSideApply should give up ownership of a field if forced applied by a controller [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Servers with support for API chunking should return chunks of results for list calls [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Servers with support for API chunking should support continue listing from the last key if the original version has been compacted away, though the list is inconsistent [Slow] [Conformance] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Slow": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] [Flaky] kubectl explain works for CR with the same resource name as built-in object. [Suite:k8s]", + "labels": { + "Flaky": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceFieldSelectors [Privileged:ClusterAdmin] [FeatureGate:CustomResourceFieldSelectors] [Beta] CustomResourceFieldSelectors MUST list and watch custom resources matching the field selector [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "FeatureGate:CustomResourceFieldSelectors": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] [Beta] MUST NOT fail to update a resource due to JSONSchema errors on unchanged correlatable fields [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "FeatureGate:CRDValidationRatcheting": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] [Beta] MUST fail to update a resource due to JSONSchema errors on unchanged uncorrelatable fields [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "FeatureGate:CRDValidationRatcheting": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] [Beta] MUST fail to update a resource due to JSONSchema errors on changed fields [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "FeatureGate:CRDValidationRatcheting": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] [Beta] MUST NOT fail to update a resource due to CRD Validation Rule errors on unchanged correlatable fields [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "FeatureGate:CRDValidationRatcheting": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] [Beta] MUST fail to update a resource due to CRD Validation Rule errors on unchanged uncorrelatable fields [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "FeatureGate:CRDValidationRatcheting": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] [Beta] MUST fail to update a resource due to CRD Validation Rule errors on changed fields [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "FeatureGate:CRDValidationRatcheting": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] [Beta] MUST NOT ratchet errors raised by transition rules [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "FeatureGate:CRDValidationRatcheting": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] [Beta] MUST evaluate a CRD Validation Rule with oldSelf = nil for new values when optionalOldSelf is true [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:Beta": {}, + "FeatureGate:CRDValidationRatcheting": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST NOT fail validation for create of a custom resource that satisfies the x-kubernetes-validations rules [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail validation for create of a custom resource that does not satisfy the x-kubernetes-validations rules [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail create of a custom resource definition that contains a x-kubernetes-validations rule that refers to a property that do not exist [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail create of a custom resource definition that contains an x-kubernetes-validations rule that contains a syntax error [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail create of a custom resource definition that contains an x-kubernetes-validations rule that exceeds the estimated cost limit [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail create of a custom resource that exceeds the runtime cost limit for x-kubernetes-validations rule execution [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail update of a custom resource that does not satisfy a x-kubernetes-validations transition rule [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Discovery should accurately determine present and missing resources [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Discovery Custom resource should have storage version hash [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Discovery should validate PreferredVersion for each APIGroup [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Discovery should locate the groupVersion and a resource within each APIGroup [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Etcd failure [Disruptive] should recover from network partition with master [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Etcd failure [Disruptive] should recover from SIGKILL [Serial] [Suite:k8s]", + "labels": { + "Disruptive": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] FieldValidation should detect unknown and duplicate fields of a typed object [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] FieldValidation should detect unknown metadata fields of a typed object [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] FieldValidation should create/apply a valid CR for CRD with validation schema [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] FieldValidation should create/apply a CR with unknown fields for CRD with no validation schema [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] FieldValidation should create/apply an invalid CR with extra properties for CRD with validation schema [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] FieldValidation should detect unknown metadata fields in both the root and embedded object of a CR [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] FieldValidation should detect duplicates in a CR when preserving unknown fields [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] API priority and fairness should ensure that requests can be classified by adding FlowSchema and PriorityLevelConfiguration [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] API priority and fairness should ensure that requests can't be drowned out (priority) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] API priority and fairness should ensure that requests can't be drowned out (fairness) [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] API priority and fairness should support FlowSchema API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] API priority and fairness should support PriorityLevelConfiguration API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should orphan pods created by rc if deleteOptions.OrphanDependents is nil [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should support cascading deletion of custom resources [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should support orphan deletion of custom resources [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Garbage collector should delete jobs and pods created by cronjob [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Generated clientset should create pods, set the deletionTimestamp and deletionGracePeriodSeconds of the pod [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Generated clientset should create v1 cronJobs, delete cronJobs, watch cronJobs [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] health handlers should contain necessary checks [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Namespaces [Serial] should delete fast enough (90 percent of 100 namespaces in 150 seconds) [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Namespaces [Serial] should always delete fast (ALL of 100 namespaces in 150 seconds) [Feature:ComprehensiveNamespaceDraining] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Feature:ComprehensiveNamespaceDraining": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Namespaces [Serial] should apply changes to a namespace status [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Namespaces [Serial] should apply an update to a Namespace [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Namespaces [Serial] should apply a finalizer to a Namespace [Conformance] [Suite:openshift/conformance/serial/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] OpenAPIV3 should round trip OpenAPI V3 for all built-in group versions [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] OpenAPIV3 should publish OpenAPI V3 for CustomResourceDefinition [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] OpenAPIV3 should contain OpenAPI V3 for Aggregated APIServer [Serial] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] client-go should negotiate watch and report errors with accept \"application/json\" [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] client-go should negotiate watch and report errors with accept \"application/vnd.kubernetes.protobuf\" [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] client-go should negotiate watch and report errors with accept \"application/vnd.kubernetes.protobuf,application/json\" [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] client-go should negotiate watch and report errors with accept \"application/json,application/vnd.kubernetes.protobuf\" [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Server request timeout should return HTTP status code 400 if the user specifies an invalid timeout in the request URL [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Server request timeout the request should be served with a default timeout if the specified timeout in the request URL exceeds maximum allowed [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Server request timeout default timeout should be used if the specified timeout in the request URL is 0s [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a ResourceClaim [Feature:DynamicResourceAllocation] [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:DynamicResourceAllocation": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a persistent volume claim [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a persistent volume claim with a storage class [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a custom resource. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should manage the lifecycle of a ResourceQuota [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should apply changes to a resourcequota status [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope using scope-selectors. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes through scope selectors. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against a pod with same priority class. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:PodPriority": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against 2 pods with same priority class. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:PodPriority": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against 2 pods with different priority class. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:PodPriority": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's multiple priority class scope (quota set to pod count: 2) against 2 pods with same priority classes. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:PodPriority": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against a pod with different priority class (ScopeSelectorOpNotIn). [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:PodPriority": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against a pod with different priority class (ScopeSelectorOpExists). [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:PodPriority": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (cpu, memory quota set) against a pod with same priority class. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "Feature:PodPriority": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with cross namespace pod affinity scope using scope-selectors. [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] server version should find the server version [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] StorageVersion resources [Feature:StorageVersionAPI] storage version with non-existing id should be GC'ed [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:StorageVersionAPI": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Servers with support for Table transformation should return pod details [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Servers with support for Table transformation should return chunks of table results for list calls [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Servers with support for Table transformation should return generic metadata details across all namespaces for nodes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should validate against a Deployment [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should type check validation expressions [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should allow expressions to refer variables. [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should type check a CRD [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should support ValidatingAdmissionPolicy API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should support ValidatingAdmissionPolicyBinding API operations [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] API Streaming (aka. WatchList) [Serial] [Feature:WatchList] should be requested by informers when WatchListClient is enabled [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:WatchList": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] API Streaming (aka. WatchList) [Serial] [Feature:WatchList] should be requested by client-go's List method when WatchListClient is enabled [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:WatchList": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] API Streaming (aka. WatchList) [Serial] [Feature:WatchList] should be requested by dynamic client's List method when WatchListClient is enabled [Disabled:Alpha] [Suite:k8s]", + "labels": { + "Feature:WatchList": {}, + "Serial": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to create and update validating webhook configurations with match conditions [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to create and update mutating webhook configurations with match conditions [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should reject validating webhook configurations with invalid match conditions [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should reject mutating webhook configurations with invalid match conditions [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate everything except 'skip-me' configmaps [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-api-machinery": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl delete interactive based on user confirmation input [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl exec should be able to execute 1000 times in a container [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl logs logs should be able to retrieve and filter logs [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl logs default container logs the second container is the default-container by annotation should log default container if not specified [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl logs all pod logs the Deployment has 2 replicas and each pod has 2 containers should get logs from all pods based on default container [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl logs all pod logs the Deployment has 2 replicas and each pod has 2 containers should get logs from each pod and each container in Deployment [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance] [Slow] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should support exec [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should support exec using resource/name [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should support exec through an HTTP proxy [Skipped:Proxy] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should support exec through kubectl proxy [Skipped:Proxy] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should return command exit codes execing into a container with a successful command [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should return command exit codes execing into a container with a failing command [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should return command exit codes should support port-forward [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should return command exit codes should handle in-cluster config [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod Kubectl run running a successful command [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod Kubectl run running a failing command [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod Kubectl run [Slow] running a failing command without --restart=Never [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod Kubectl run [Slow] running a failing command without --restart=Never, but with --rm [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod Kubectl run [Slow] running a failing command with --leave-stdin-open [Suite:k8s]", + "labels": { + "Slow": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should support inline execution and attach [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should support inline execution and attach with websockets or fallback to spdy [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Simple pod should contain last line of the log [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl get componentstatuses should get componentstatuses [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl prune with applyset should apply and prune objects [Disabled:RebaseInProgress] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl apply should apply a new configuration to an existing RC [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl apply should reuse port when apply to an existing SVC [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl apply apply set/view last-applied [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl diff should check if kubectl diff finds a difference for Deployments [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl server-side dry-run should check if kubectl can dry-run update Pods [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl validation should create/apply a CR with unknown fields for CRD with no validation schema [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl validation should create/apply a valid CR for CRD with validation schema [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl validation should create/apply an invalid/valid CR with arbitrary-extra properties for CRD with partially-specified validation schema [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl validation should detect unknown metadata fields in both the root and embedded object of a CR [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl validation should detect unknown metadata fields of a typed object [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes control plane services is included in cluster-info [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl cluster-info dump should check if cluster-info dump succeeds [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for cronjob [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl expose should create services for rc [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl copy should copy a file from a running Pod [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl replace should update a single-container pod's image [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Proxy server should support --unix-socket=/path [Conformance] [Suite:openshift/conformance/parallel/minimal] [Suite:k8s]", + "labels": { + "Conformance": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl taint [Serial] should update the taint on a node [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl taint [Serial] should remove all the taints with the same key off a node [Skipped:SingleReplicaTopology] [Suite:openshift/conformance/serial] [Suite:k8s]", + "labels": { + "Serial": {}, + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl events should show event when pod is created [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl create quota should create a quota without scopes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl create quota should create a quota with scopes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client Kubectl create quota should reject quota with invalid scopes [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client kubectl wait should ignore not found error with --for=delete [Disabled:Broken] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client kubectl subresource flag should not be used in a bulk GET [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl client kubectl subresource flag GET on status subresource of built-in type (node) returns identical info as GET on the built-in type [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 that expects a client request should support a client that connects, sends NO DATA, and disconnects [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 that expects a client request should support a client that connects, sends DATA, and disconnects [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 that expects NO client request should support a client that connects, sends DATA, and disconnects [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 should support forwarding over websockets [Skipped:Proxy] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl Port forwarding With a server listening on localhost that expects a client request should support a client that connects, sends NO DATA, and disconnects [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl Port forwarding With a server listening on localhost that expects a client request should support a client that connects, sends DATA, and disconnects [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl Port forwarding With a server listening on localhost that expects NO client request should support a client that connects, sends DATA, and disconnects [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl Port forwarding With a server listening on localhost should support forwarding over websockets [Skipped:Proxy] [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + }, + { + "name": "[sig-cli] Kubectl rollout undo undo should rollback and update deployment env [Suite:openshift/conformance/parallel] [Suite:k8s]", + "labels": { + "sig-cli": {} + }, + "resources": { + "isolation": {} + }, + "source": "openshift:payload:hyperkube", + "lifecycle": "blocking" + } +] \ No newline at end of file diff --git a/LICENSES/vendor/github.com/openshift-eng/openshift-tests-extension/LICENSE b/LICENSES/vendor/github.com/openshift-eng/openshift-tests-extension/LICENSE new file mode 100644 index 0000000000000..f67fc367d04a2 --- /dev/null +++ b/LICENSES/vendor/github.com/openshift-eng/openshift-tests-extension/LICENSE @@ -0,0 +1,205 @@ += vendor/github.com/openshift-eng/openshift-tests-extension licensed under: = + + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "[]" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright [yyyy] [name of copyright owner] + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. + += vendor/github.com/openshift-eng/openshift-tests-extension/LICENSE 86d3f3a95c324c9479bd8986968f4327 diff --git a/LICENSES/vendor/golang.org/x/mod/LICENSE b/LICENSES/vendor/golang.org/x/mod/LICENSE index 2bef705f0d991..3e7f5d4a79cc7 100644 --- a/LICENSES/vendor/golang.org/x/mod/LICENSE +++ b/LICENSES/vendor/golang.org/x/mod/LICENSE @@ -1,6 +1,6 @@ = vendor/golang.org/x/mod licensed under: = -Copyright (c) 2009 The Go Authors. All rights reserved. +Copyright 2009 The Go Authors. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are @@ -12,7 +12,7 @@ notice, this list of conditions and the following disclaimer. copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. - * Neither the name of Google Inc. nor the names of its + * Neither the name of Google LLC nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. @@ -28,4 +28,4 @@ THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -= vendor/golang.org/x/mod/LICENSE 5d4950ecb7b26d2c5e4e7b4e0dd74707 += vendor/golang.org/x/mod/LICENSE 7998cb338f82d15c0eff93b7004d272a diff --git a/LICENSES/vendor/golang.org/x/oauth2/LICENSE b/LICENSES/vendor/golang.org/x/oauth2/LICENSE index f3045780ff052..36377941e14ad 100644 --- a/LICENSES/vendor/golang.org/x/oauth2/LICENSE +++ b/LICENSES/vendor/golang.org/x/oauth2/LICENSE @@ -1,6 +1,6 @@ = vendor/golang.org/x/oauth2 licensed under: = -Copyright (c) 2009 The Go Authors. All rights reserved. +Copyright 2009 The Go Authors. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are @@ -12,7 +12,7 @@ notice, this list of conditions and the following disclaimer. copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. - * Neither the name of Google Inc. nor the names of its + * Neither the name of Google LLC nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. @@ -28,4 +28,4 @@ THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -= vendor/golang.org/x/oauth2/LICENSE 5d4950ecb7b26d2c5e4e7b4e0dd74707 += vendor/golang.org/x/oauth2/LICENSE 7998cb338f82d15c0eff93b7004d272a diff --git a/LICENSES/vendor/golang.org/x/tools/LICENSE b/LICENSES/vendor/golang.org/x/tools/LICENSE index 6532fef7a76f8..7f9f3b253f77b 100644 --- a/LICENSES/vendor/golang.org/x/tools/LICENSE +++ b/LICENSES/vendor/golang.org/x/tools/LICENSE @@ -1,6 +1,6 @@ = vendor/golang.org/x/tools licensed under: = -Copyright (c) 2009 The Go Authors. All rights reserved. +Copyright 2009 The Go Authors. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are @@ -12,7 +12,7 @@ notice, this list of conditions and the following disclaimer. copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. - * Neither the name of Google Inc. nor the names of its + * Neither the name of Google LLC nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. @@ -28,4 +28,4 @@ THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -= vendor/golang.org/x/tools/LICENSE 5d4950ecb7b26d2c5e4e7b4e0dd74707 += vendor/golang.org/x/tools/LICENSE 7998cb338f82d15c0eff93b7004d272a diff --git a/go.mod b/go.mod index d1726268c0a77..6223ba07fbbde 100644 --- a/go.mod +++ b/go.mod @@ -44,10 +44,11 @@ require ( github.com/moby/ipvs v1.1.0 github.com/mrunalp/fileutils v0.5.1 github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822 - github.com/onsi/ginkgo/v2 v2.17.2 + github.com/onsi/ginkgo/v2 v2.20.2 github.com/onsi/gomega v1.33.1 github.com/opencontainers/runc v1.1.13 github.com/opencontainers/selinux v1.11.0 + github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f github.com/openshift/apiserver-library-go v0.0.0-20241001175710-6064b62894a6 github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f @@ -77,14 +78,14 @@ require ( go.uber.org/zap v1.26.0 golang.org/x/crypto v0.27.0 golang.org/x/net v0.29.0 - golang.org/x/oauth2 v0.21.0 + golang.org/x/oauth2 v0.22.0 golang.org/x/sync v0.8.0 golang.org/x/sys v0.25.0 golang.org/x/term v0.24.0 golang.org/x/time v0.3.0 - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 - google.golang.org/grpc v1.65.0 + golang.org/x/tools v0.24.0 + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 + google.golang.org/grpc v1.67.0 google.golang.org/protobuf v1.34.2 gopkg.in/evanphx/json-patch.v4 v4.12.0 gopkg.in/natefinch/lumberjack.v2 v2.2.1 @@ -164,7 +165,7 @@ require ( github.com/go-task/slim-sprig/v3 v3.0.0 // indirect github.com/golang-jwt/jwt/v4 v4.5.0 // indirect github.com/google/btree v1.0.1 // indirect - github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 // indirect + github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 // indirect github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510 // indirect github.com/gorilla/websocket v1.5.0 // indirect github.com/gregjones/httpcache v0.0.0-20180305231024-9cad4c3443a7 // indirect @@ -219,10 +220,10 @@ require ( go.uber.org/atomic v1.7.0 // indirect go.uber.org/multierr v1.11.0 // indirect golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc // indirect - golang.org/x/mod v0.17.0 // indirect + golang.org/x/mod v0.20.0 // indirect golang.org/x/text v0.18.0 // indirect google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect gopkg.in/inf.v0 v0.9.1 // indirect k8s.io/gengo/v2 v2.0.0-20240228010128-51d4e06bde70 // indirect sigs.k8s.io/apiserver-network-proxy/konnectivity-client v0.30.3 // indirect @@ -234,7 +235,7 @@ require ( replace ( github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ./staging/src/k8s.io/api k8s.io/apiextensions-apiserver => ./staging/src/k8s.io/apiextensions-apiserver k8s.io/apimachinery => ./staging/src/k8s.io/apimachinery diff --git a/go.sum b/go.sum index b3bb01a5978e4..092451a292b76 100644 --- a/go.sum +++ b/go.sum @@ -1,7 +1,7 @@ bazil.org/fuse v0.0.0-20160811212531-371fbbdaa898/go.mod h1:Xbm+BRKSBEpa4q4hTSxohYNQpsxXPbPry4JJWOB3LB8= bitbucket.org/bertimus9/systemstat v0.5.0 h1:n0aLnh2Jo4nBUBym9cE5PJDG8GT6g+4VuS2Ya2jYYpA= bitbucket.org/bertimus9/systemstat v0.5.0/go.mod h1:EkUWPp8lKFPMXP8vnbpT5JDI0W/sTiLZAvN8ONWErHY= -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= cloud.google.com/go v0.26.0/go.mod h1:aQUYkXzVsufM+DwF1aE+0xfcU+56JwCaLick0ClmMTw= cloud.google.com/go v0.34.0/go.mod h1:aQUYkXzVsufM+DwF1aE+0xfcU+56JwCaLick0ClmMTw= cloud.google.com/go v0.110.7/go.mod h1:+EYjdK8e5RME/VY/qLCAtuyALQ9q67dvuum8i+H5xsI= @@ -30,7 +30,7 @@ cloud.google.com/go/cloudbuild v1.13.0/go.mod h1:lyJg7v97SUIPq4RC2sGsz/9tNczhyv2 cloud.google.com/go/clouddms v1.6.1/go.mod h1:Ygo1vL52Ov4TBZQquhz5fiw2CQ58gvu+PlS6PVXCpZI= cloud.google.com/go/cloudtasks v1.12.1/go.mod h1:a9udmnou9KO2iulGscKR0qBYjreuX8oHwpmFsKspEvM= cloud.google.com/go/compute v1.23.0/go.mod h1:4tCnrn48xsqlwSAiLf1HXMQk8CONslYbdiEZc9FEIbM= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= cloud.google.com/go/contactcenterinsights v1.10.0/go.mod h1:bsg/R7zGLYMVxFFzfh9ooLTruLRCG9fnzhH9KznHhbM= cloud.google.com/go/container v1.24.0/go.mod h1:lTNExE2R7f+DLbAN+rJiKTisauFCaoDq6NURZ83eVH4= cloud.google.com/go/containeranalysis v0.10.1/go.mod h1:Ya2jiILITMY68ZLPaogjmOMNkwsDrWBSTyBubGXO7j0= @@ -144,6 +144,7 @@ github.com/alecthomas/template v0.0.0-20160405071501-a0175ee3bccc/go.mod h1:LOuy github.com/alecthomas/units v0.0.0-20151022065526-2efee857e7cf/go.mod h1:ybxpYRFXyAe+OPACYpWeL0wqObRcbAqCMya13uyzqw0= github.com/alecthomas/units v0.0.0-20211218093645-b94a6e3cc137/go.mod h1:OMCwj8VM1Kc9e19TLln2VL61YJF0x1XFtfdL4JdbSyE= github.com/antihax/optional v1.0.0/go.mod h1:uupD/76wgC+ih3iEmQUL+0Ugr19nfwCT1kdvxnR2qWY= +github.com/antlr/antlr4/runtime/Go/antlr/v4 v4.0.0-20230305170008-8188dc5388df/go.mod h1:pSwJ0fSY5KhvocuWSx4fz3BA8OrA1bQn+K1Eli3BRwM= github.com/antlr4-go/antlr/v4 v4.13.0 h1:lxCg3LAv+EUK6t1i0y1V6/SLeUi0eKEKdhQAlS8TVTI= github.com/antlr4-go/antlr/v4 v4.13.0/go.mod h1:pfChB/xh/Unjila75QW7+VU4TSnWnnk9UTnmpPaOR2g= github.com/armon/circbuf v0.0.0-20150827004946-bbbad097214e h1:QEF07wC0T1rKkctt1RINW/+RMTVmiwxETico2l3gxJA= @@ -181,7 +182,7 @@ github.com/cilium/ebpf v0.9.1 h1:64sn2K3UKw8NbP/blsixRpF3nXuyhz/VjRlRzvlBRu4= github.com/cilium/ebpf v0.9.1/go.mod h1:+OhNOIXx/Fnu1IE8bJz2dzOA+VSfyTfdNUVdlQnxUFY= github.com/client9/misspell v0.3.4/go.mod h1:qj6jICC3Q7zFZvVWo7KLAzC3yx5G7kyvSDkc90ppPyw= github.com/cncf/udpa/go v0.0.0-20191209042840-269d4d468f6f/go.mod h1:M8M6+tZqaGXZJjfX53e64911xZQV5JYwmTeXPW+k8Sc= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/cockroachdb/datadriven v1.0.2 h1:H9MtNqVoVhvd9nCBwOyDjUEdZCREqbIdCJD93PBm/jA= github.com/cockroachdb/datadriven v1.0.2/go.mod h1:a9RdTaap04u637JoCzcUoIcDmvwSUtcUFtT/C3kJlTU= github.com/container-storage-interface/spec v1.9.0 h1:zKtX4STsq31Knz3gciCYCi1SXtO2HJDecIjDVboYavY= @@ -258,9 +259,9 @@ github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRr github.com/envoyproxy/go-control-plane v0.9.0/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymFceY/DCBVvsKhRF0iEA4= github.com/envoyproxy/go-control-plane v0.9.1-0.20191026205805-5f8ba28d4473/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymFceY/DCBVvsKhRF0iEA4= github.com/envoyproxy/go-control-plane v0.9.4/go.mod h1:6rpuAdCZL397s3pYoYcLgu1mIlRU8Am5FuJP05cCM98= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= github.com/envoyproxy/protoc-gen-validate v0.1.0/go.mod h1:iSmxcyjqTsJpI2R4NaDN7+kN2VEUnK/pcBlmesArF7c= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/euank/go-kmsg-parser v2.0.0+incompatible h1:cHD53+PLQuuQyLZeriD1V/esuG4MuU0Pjs5y6iknohY= github.com/euank/go-kmsg-parser v2.0.0+incompatible/go.mod h1:MhmAMZ8V4CYH4ybgdRwPr2TU5ThnS43puaKEMpja1uw= github.com/evanphx/json-patch v4.12.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= @@ -323,7 +324,7 @@ github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69 github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= github.com/golang/glog v0.0.0-20160126235308-23def4e6c14b/go.mod h1:SBH7ygxi8pfUlaOkMMuAQtPIUF8ecWP5IEl/CR7VP2Q= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20190129154638-5b532d6fd5ef/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/groupcache v0.0.0-20190702054246-869f871628b6/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/groupcache v0.0.0-20200121045136-8c9f03a8e57e/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= @@ -378,8 +379,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510 h1:El6M4kTTCOh6aBiKaUGG7oYTSPP8MxqL4YI3kZKwcP4= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.1.1/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= @@ -497,6 +498,8 @@ github.com/opencontainers/runtime-spec v1.0.3-0.20220909204839-494a5a6aca78 h1:R github.com/opencontainers/runtime-spec v1.0.3-0.20220909204839-494a5a6aca78/go.mod h1:jwyrGlmzljRJv/Fgzds9SsS/C5hL+LL3ko9hs6T5lQ0= github.com/opencontainers/selinux v1.11.0 h1:+5Zbo97w3Lbmb3PeqQtpmTkMwsW5nRI3YaLpt7tQ7oU= github.com/opencontainers/selinux v1.11.0/go.mod h1:E5dMC3VPuVvVHDYmi78qvhJp8+M586T4DlDRYpFkyec= +github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf h1:WY0VzsbhiQQNcS8Y9AXuiqqY8YHljePp1QhrIrue4+c= +github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf/go.mod h1:1OhaNsaU9vuy/dlYZLEve7bgE2Ed+yTV5VSbYvGXt4s= github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f h1:ya1OmyZm3LIIxI3U9VE9Nyx3ehCHgBwxyFUPflYPWls= github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f/go.mod h1:Shkl4HanLwDiiBzakv+con/aMGnVE2MAGvoKp5oyYUo= github.com/openshift/apiserver-library-go v0.0.0-20241001175710-6064b62894a6 h1:Wban+ggY6sbg611SQSOeavUeug2cRJGz0rEeXxTxIH0= @@ -508,8 +511,8 @@ github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 h1:ls1C5cvJbA5CbOw github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2/go.mod h1:s6Fqwb2KiWG6leCegVhw4KW40tf9f7m+SF1aXiE8Wsk= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060 h1:jiDC7d8d+jmjv2WfiMY0+Uf55q11MGyYkGGqXnfqWTU= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/opentracing/opentracing-go v1.1.0/go.mod h1:UkNAQd3GIcIGf0SeVgPpRdFStlNbqXla1AfSYxPUl2o= github.com/pelletier/go-toml v1.2.0/go.mod h1:5z9KED0ma1S8pY6P1sdut58dfprrGBbd/94hg7ilaic= github.com/peterbourgon/diskv v2.0.1+incompatible h1:UBdAOUP5p4RWqPBg048CAvpKN+vxiaj6gdUUzhl4XmI= @@ -519,6 +522,7 @@ github.com/pkg/errors v0.8.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINE github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/profile v1.7.0/go.mod h1:8Uer0jas47ZQMJ7VD+OHknK4YDY07LPUC6dEvqDjvNo= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -693,8 +697,8 @@ golang.org/x/lint v0.0.0-20190227174305-5b3e6a55c961/go.mod h1:wehouNa3lNwaWXcvx golang.org/x/lint v0.0.0-20190313153728-d0100b6bd8b3/go.mod h1:6SW0HCj/g11FgYtHlgUYUwCkIfeOF89ocIRzGO/8vkc= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0 h1:zY54UmvipHiNd+pm+m0x9KhZ9hl1/7QNMyxXbc6ICqA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0 h1:utOm6MM3R3dnawAiJgn0y+xvuYRsm1RKM/4giyfDgV0= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20181114220301-adae6a3d119a/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= @@ -716,8 +720,8 @@ golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U= golang.org/x/oauth2 v0.0.0-20200107190931-bf48bf16ab8d/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181108010431-42b317875d0f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181221193216-37e7f081c4d4/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -754,7 +758,7 @@ golang.org/x/sys v0.0.0-20220715151400-c0bba94af5f8/go.mod h1:oPkhp1MJrh7nUepCBc golang.org/x/sys v0.1.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.0.0-20201126162022-7de9c90e9dd1/go.mod h1:bj7SfCRtBDWHUb9snDiAeCFNEtKQo2Wmx5Cou7ajbmo= golang.org/x/term v0.0.0-20220526004731-065cf7ba2467/go.mod h1:jbD1KX2456YbFQfuXm/mYQcufACuNUgVhRMnK/tPxf8= golang.org/x/term v0.24.0 h1:Mh5cbb+Zk2hqqXNO7S1iTjEphVL+jb8ZWaqh/g+JWkM= @@ -778,8 +782,8 @@ golang.org/x/tools v0.0.0-20190624222133-a101b041ded4/go.mod h1:/rFqwRUd4F7ZHNgw golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -797,10 +801,10 @@ google.golang.org/genproto v0.0.0-20200513103714-09dca8ec2884/go.mod h1:55QSHmfG google.golang.org/genproto v0.0.0-20200526211855-cb27e3aa2013/go.mod h1:NbSheEEYHJ7i3ixzK3sjbqSGDJWnxyFXZblF3eUsNvo= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d h1:VBu5YqKPv6XiJ199exd8Br+Aetz+o08F+PLMnwJQHAY= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= google.golang.org/grpc v1.19.0/go.mod h1:mqu4LbDTu4XGKhr4mRzUsmM4RtVoemTSY81AxZiDr8c= google.golang.org/grpc v1.20.1/go.mod h1:10oTOabMzJvdu6/UiuZezV6QK5dSlG84ov/aaiqXj38= google.golang.org/grpc v1.21.0/go.mod h1:oYelfM1adQP15Ek0mdvEgi9Df8B9CZIaU1084ijfRaM= @@ -811,8 +815,8 @@ google.golang.org/grpc v1.27.1/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8 google.golang.org/grpc v1.29.1/go.mod h1:itym6AZVZYACWQqET3MqgPpjcuV5QH3BxFS3IjizoKk= google.golang.org/grpc v1.33.1/go.mod h1:fr5YgcSWrqhRRxogOsw7RzIpsmvOZ6IcH4kBYTpR3n0= google.golang.org/grpc v1.33.2/go.mod h1:JMHMWHQWaTccqQQlmk3MJZS+GWXOdAesneDmEnv2fbc= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v0.0.0-20200109180630-ec00e32a8dfd/go.mod h1:DFci5gLYBciE7Vtevhsrf46CRTquxDuWsQurQQe4oz8= google.golang.org/protobuf v0.0.0-20200221191635-4d8936d0db64/go.mod h1:kwYJMbMJ01Woi6D6+Kah6886xMZcty6N08ah7+eCXa0= google.golang.org/protobuf v0.0.0-20200228230310-ab0ca4ff8a60/go.mod h1:cfTl7dwQJ+fmap5saPgwCLgHXTUD7jkjRqWcaiX5VyM= diff --git a/go.work.sum b/go.work.sum index 05b8d8d19defd..e085e445af670 100644 --- a/go.work.sum +++ b/go.work.sum @@ -1,5 +1,6 @@ bazil.org/fuse v0.0.0-20160811212531-371fbbdaa898 h1:SC+c6A1qTFstO9qmB86mPV2IpYme/2ZoEQ0hrP+wo+Q= cel.dev/expr v0.15.0 h1:O1jzfJCQBfL5BFoYktaxwIhuttaQPsVWerH9/EEKx0w= +cel.dev/expr v0.16.0 h1:yloc84fytn4zmJX2GU3TkXGsaieaV7dQ057Qs4sIG2Y= cloud.google.com/go v0.110.7 h1:rJyC7nWRg2jWGZ4wSJ5nY65GTdYJkg0cd/uXb+ACI6o= cloud.google.com/go v0.112.0/go.mod h1:3jEEVwZ/MHU4djK5t5RHuKOA/GbLddgTdVubX1qnPD4= cloud.google.com/go v0.112.1 h1:uJSeirPke5UNZHIb4SxfZklVSiWWVqW4oXlETwZziwM= @@ -81,6 +82,7 @@ cloud.google.com/go/compute v1.23.0 h1:tP41Zoavr8ptEqaW6j+LQOnyBBhO7OkOMAGrgLopT cloud.google.com/go/compute v1.23.4/go.mod h1:/EJMj55asU6kAFnuZET8zqgwgJ9FvXWXOkkfQZa4ioI= cloud.google.com/go/compute v1.24.0/go.mod h1:kw1/T+h/+tK2LJK0wiPPx1intgdAM3j/g3hFDlscY40= cloud.google.com/go/compute/metadata v0.3.0 h1:Tz+eQXMEqDIKRsmY3cHTL6FVaynIjX2QxYC4trgAKZc= +cloud.google.com/go/compute/metadata v0.5.0 h1:Zr0eK8JbFv6+Wi4ilXAR8FJ3wyNdpxHKJNPos6LTZOY= cloud.google.com/go/contactcenterinsights v1.10.0 h1:YR2aPedGVQPpFBZXJnPkqRj8M//8veIZZH5ZvICoXnI= cloud.google.com/go/contactcenterinsights v1.13.0 h1:6Vs/YnDG5STGjlWMEjN/xtmft7MrOTOnOZYUZtGTx0w= cloud.google.com/go/contactcenterinsights v1.13.0/go.mod h1:ieq5d5EtHsu8vhe2y3amtZ+BE+AQwX5qAy7cpo0POsI= @@ -372,6 +374,7 @@ github.com/alecthomas/template v0.0.0-20160405071501-a0175ee3bccc h1:cAKDfWh5Vpd github.com/alecthomas/units v0.0.0-20211218093645-b94a6e3cc137 h1:s6gZFSlWYmbqAuRjVTiNNhvNRfY2Wxp9nhfyel4rklc= github.com/andybalholm/brotli v1.0.4/go.mod h1:fO7iG3H7G2nSZ7m0zPUDn85XEX2GTukHGRSepvi9Eig= github.com/antihax/optional v1.0.0 h1:xK2lYat7ZLaVVcIuj82J8kIro4V6kDe0AUDFboUCwcg= +github.com/antlr/antlr4/runtime/Go/antlr/v4 v4.0.0-20230305170008-8188dc5388df h1:7RFfzj4SSt6nnvCPbCqijJi1nWCd+TqAT3bYCStRC18= github.com/apache/arrow/go/v12 v12.0.0/go.mod h1:d+tV/eHZZ7Dz7RPrFKtPK02tpr+c9/PEd/zm8mDS9Vg= github.com/apache/arrow/go/v14 v14.0.2 h1:N8OkaJEOfI3mEZt07BIkvo4sC6XDbL+48MBPWO5IONw= github.com/apache/arrow/go/v14 v14.0.2/go.mod h1:u3fgh3EdgN/YQ8cVQRguVW3R+seMybFg8QBQ5LU+eBY= @@ -380,7 +383,6 @@ github.com/armon/consul-api v0.0.0-20180202201655-eb2c6b5be1b6 h1:G1bPvciwNyF7IU github.com/armon/go-metrics v0.0.0-20180917152333-f0300d1749da h1:8GUt8eRujhVEGZFFEjBj46YV4rDjvGrNxb0KMWYkL2I= github.com/armon/go-radix v0.0.0-20180808171621-7fddfc383310 h1:BUAU3CGlLvorLI26FmByPp2eC2qla6E1Tw+scpcg/to= github.com/aws/aws-sdk-go v1.35.24 h1:U3GNTg8+7xSM6OAJ8zksiSM4bRqxBWmVwwehvOSNG3A= -github.com/aws/aws-sdk-go v1.35.24/go.mod h1:tlPOdRjfxPBpNIwqDj61rmsnA85v9jc0Ps9+muhnW+k= github.com/bgentry/speakeasy v0.1.0 h1:ByYyxL9InA1OWqxJqqp2A5pYHUrCiAL6K3J+LKSsQkY= github.com/bketelsen/crypt v0.0.3-0.20200106085610-5cbc8cc4026c h1:+0HFd5KSZ/mm3JmhmrDukiId5iR6w4+BdFtfSy4yWIc= github.com/census-instrumentation/opencensus-proto v0.4.1 h1:iKLQ0xPNFxR/2hzXZMrBo8f1j86j5WHzznCCQxV/b8g= @@ -402,6 +404,7 @@ github.com/cncf/xds/go v0.0.0-20210805033703-aa0b78936158/go.mod h1:eXthEFrGJvWH github.com/cncf/xds/go v0.0.0-20210922020428-25de7278fc84/go.mod h1:eXthEFrGJvWHgFFCl3hGmgk+/aYT6PnTQLykKQRLhEs= github.com/cncf/xds/go v0.0.0-20211011173535-cb28da3451f1/go.mod h1:eXthEFrGJvWHgFFCl3hGmgk+/aYT6PnTQLykKQRLhEs= github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b h1:ga8SEFjZ60pxLcmhnThWgvH2wg8376yUJmPhEH4H3kw= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20 h1:N+3sFI5GUjRKBi+i0TxYVST9h4Ie192jJWpHvthBBgg= github.com/containerd/containerd v1.4.9 h1:JIw9mjVw4LsGmnA/Bqg9j9e+XB7soOJufrKUpA6n2Ns= github.com/containerd/continuity v0.1.0 h1:UFRRY5JemiAhPZrr/uE0n8fMTLcZsUvySPr1+D7pgr8= github.com/containerd/fifo v1.0.0 h1:6PirWBr9/L7GDamKr+XM0IeUFXu5mf3M/BPpH9gaLBU= @@ -420,7 +423,9 @@ github.com/envoyproxy/go-control-plane v0.9.9-0.20210217033140-668b12f5399d/go.m github.com/envoyproxy/go-control-plane v0.9.10-0.20210907150352-cf90f659a021/go.mod h1:AFq3mo9L8Lqqiid3OhADV3RfLJnjiw63cSpi+fDTRC0= github.com/envoyproxy/go-control-plane v0.11.1 h1:wSUXTlLfiAQRWs2F+p+EKOY9rUyis1MyGqJ2DIk5HpM= github.com/envoyproxy/go-control-plane v0.12.0 h1:4X+VP1GHd1Mhj6IB5mMeGbLCleqxjletLK6K0rbxyZI= +github.com/envoyproxy/go-control-plane v0.13.0 h1:HzkeUz1Knt+3bK+8LG1bxOO/jzWZmdxpwC51i202les= github.com/envoyproxy/protoc-gen-validate v1.0.4 h1:gVPz/FMfvh57HdSJQyvBtF00j8JU4zdyUgIUNhlgg0A= +github.com/envoyproxy/protoc-gen-validate v1.1.0 h1:tntQDh69XqOCOZsDz0lVJQez/2L6Uu2PdjCQwWCJ3bM= github.com/evanphx/json-patch v4.12.0+incompatible h1:4onqiflcdA9EOZ4RxV643DvftH5pOlLGNtQ5lPWQu84= github.com/fatih/color v1.7.0 h1:DkWD4oS2D8LGGgTQ6IvwJJXSL5Vp2ffcQg58nFV38Ys= github.com/felixge/fgprof v0.9.4 h1:ocDNwMFlnA0NU0zSB3I52xkO4sFXk80VK9lXjLClu88= @@ -440,9 +445,9 @@ github.com/goccy/go-json v0.9.11/go.mod h1:6MelG93GURQebXPDq3khkgXZkazVtN9CRI+MG github.com/goccy/go-json v0.10.2 h1:CrxCmQqYDkv1z7lO7Wbh2HN93uovUHgrECaO5ZrCXAU= github.com/goccy/go-json v0.10.2/go.mod h1:6MelG93GURQebXPDq3khkgXZkazVtN9CRI+MGFi0w8I= github.com/gogo/googleapis v1.4.1 h1:1Yx4Myt7BxzvUr5ldGSbwYiZG6t9wGBZ+8/fX3Wvtq0= -github.com/gogo/googleapis v1.4.1/go.mod h1:2lpHqI5OcWCtVElxXnPt+s8oJvMpySlOyM6xDCrzib4= github.com/golang/glog v1.2.0 h1:uCdmnmatrKCgMBlM4rMuJZWOkPDqdbZPnrMXDY4gI68= github.com/golang/glog v1.2.1 h1:OptwRhECazUx5ix5TTWC3EZhsZEHWcYWY4FQHTIubm4= +github.com/golang/glog v1.2.2 h1:1+mZ9upx1Dh6FmUTFR1naJ77miKiXgALjWOZ3NVFPmY= github.com/golang/mock v1.1.1 h1:G5FRp8JnTd7RQH5kemVNlMeyXQAztQ3mOWV95KxsXH8= github.com/golang/mock v1.3.1 h1:qGJ6qTW+x6xX/my+8YUVl4WNpX9B7+/l2tRsHGZ7f2s= github.com/golang/mock v1.6.0 h1:ErTB+efbowRARo13NNdxyJji2egdxLGQhRaY+DUumQc= @@ -508,7 +513,6 @@ github.com/iancoleman/strcase v0.3.0/go.mod h1:iwCmte+B7n89clKwxIoIXy/HfoL7AsD47 github.com/ianlancetaylor/demangle v0.0.0-20200824232613-28f6c0f3b639 h1:mV02weKRL81bEnm8A0HT1/CAelMQDBuQIfLw8n+d6xI= github.com/ianlancetaylor/demangle v0.0.0-20240312041847-bd984b5ce465 h1:KwWnWVWCNtNq/ewIX7HIKnELmEx2nDP42yskD/pi7QE= github.com/jmespath/go-jmespath v0.4.0 h1:BEgLn5cpjn8UN1mAw4NjwDrS35OdebyEtFe+9YPoQUg= -github.com/jmespath/go-jmespath v0.4.0/go.mod h1:T8mJZnbsbmF+m6zOOFylbeCJqk5+pHWvzYPziyZiYoo= github.com/jpillora/backoff v1.0.0 h1:uvFg412JmmHBHw7iwprIxkPMI+sGQ4kzOWsMeHnm2EA= github.com/jstemmer/go-junit-report v0.0.0-20190106144839-af01ea7f8024 h1:rBMNdlhTLzJjJSDIjNEXX1Pz3Hmwmz91v+zycvx9PJc= github.com/jstemmer/go-junit-report v0.9.1 h1:6QPYqodiu3GuPL+7mfx+NwDdp2eTkp9IfEUpgAwUN0o= @@ -533,7 +537,6 @@ github.com/magiconair/properties v1.8.1 h1:ZC2Vc7/ZFkGmsVC9KvOjumD+G5lXy2RtTKyzR github.com/mattn/go-colorable v0.0.9 h1:UVL0vNpWh04HeJXV0KLcaT7r06gOH2l4OW6ddYRUIY4= github.com/mattn/go-isatty v0.0.3 h1:ns/ykhmWi7G9O+8a448SecJU3nSMBXJfqQkl0upE1jI= github.com/matttproud/golang_protobuf_extensions v1.0.2 h1:hAHbPm5IJGijwng3PWk09JkG9WeqChjprR5s9bBZ+OM= -github.com/matttproud/golang_protobuf_extensions v1.0.2/go.mod h1:BSXmuO+STAnVfrANrmjBb36TMTDstsz7MSK+HVaYKv4= github.com/miekg/dns v1.0.14 h1:9jZdLNd/P4+SfEJ0TNyxYpsK8N4GtfylBLqtbYN1sbA= github.com/minio/asm2plan9s v0.0.0-20200509001527-cdd76441f9d8/go.mod h1:mC1jAcsrzbxHt8iiaC+zU4b1ylILSosueou12R++wfY= github.com/minio/c2goasm v0.0.0-20190812172519-36a3d3bbc4f3/go.mod h1:RagcQ7I8IeTMnF8JTXieKnO4Z6JCsikNEzj0DwauVzE= @@ -544,7 +547,6 @@ github.com/mitchellh/gox v0.4.0 h1:lfGJxY7ToLJQjHHwi0EX6uYBdK78egf954SQl13PQJc= github.com/mitchellh/iochan v1.0.0 h1:C+X3KsSTLFVBr/tK1eYN/vs4rJcvsiLU338UhYPJWeY= github.com/mitchellh/mapstructure v1.1.2 h1:fmNYVwqnSfB9mZU6OS2O6GsXM+wcskZDuKQzvN1EDeE= github.com/morikuni/aec v1.0.0 h1:nP9CBfwrvYnBRgY6qfDQkygYDmYwOilePFkwzv4dU8A= -github.com/morikuni/aec v1.0.0/go.mod h1:BbKIizmSmc5MMPqRYbxO4ZU0S0+P200+tUnFx7PXmsc= github.com/mwitkow/go-conntrack v0.0.0-20190716064945-2f068394615f h1:KUppIJq7/+SVif2QVs3tOP0zanoHgBEVAwHxUSIzRqU= github.com/oklog/ulid v1.3.1 h1:EGfNDEx6MqHz8B3uNV6QAib1UR2Lm97sHi3ocA6ESJ4= github.com/openshift/build-machinery-go v0.0.0-20240613134303-8359781da660 h1:F0zE2bmdVvaEd18VXuGYQdJJ1FYJu4MIDW9PYZWc9No= @@ -556,6 +558,7 @@ github.com/pierrec/lz4/v4 v4.1.18 h1:xaKrnTkyoqfh1YItXl56+6KJNVYWlEEPuAQW9xsplYQ github.com/pierrec/lz4/v4 v4.1.18/go.mod h1:gZWDp/Ze/IJXGXf23ltt2EXimqmTUXEy0GFuRQyBid4= github.com/pkg/diff v0.0.0-20210226163009-20ebb0f2a09e h1:aoZm08cpOy4WuID//EZDgcC4zIxODThtZNPirFr42+A= github.com/pkg/profile v1.7.0 h1:hnbDkaNWPCLMO9wGLdBFTIZvzDrDfBM2072E1S9gJkA= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10 h1:GFCKgmp0tecUJ0sJuv4pzYCqS9+RGSn52M3FUwPs+uo= github.com/posener/complete v1.1.1 h1:ccV59UEOTzVDnDUEFdT95ZzHVZ+5+158q8+SJb2QV5w= github.com/prometheus-operator/prometheus-operator/pkg/apis/monitoring v0.74.0 h1:AHzMWDxNiAVscJL6+4wkvFRTpMnJqiaZFEKA/osaBXE= github.com/prometheus/client_model v0.5.0/go.mod h1:dTiFglRmd66nLR9Pv9f0mZi7B7fk5Pm3gvsjB5tr+kI= @@ -648,6 +651,7 @@ golang.org/x/sys v0.18.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= golang.org/x/sys v0.20.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= golang.org/x/telemetry v0.0.0-20240208230135-b75ee8823808 h1:+Kc94D8UVEVxJnLXp/+FMfqQARZtWHfVrcRtcG8aT3g= golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2 h1:IRJeR9r1pYWsHKTRe/IInb7lYvbBVIqOgsX/u0mbOWY= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457 h1:zf5N6UOrA487eEFacMePxjXAJctxKmyjKUsjA11Uzuk= golang.org/x/term v0.0.0-20210927222741-03fcf44c2211/go.mod h1:jbD1KX2456YbFQfuXm/mYQcufACuNUgVhRMnK/tPxf8= golang.org/x/text v0.3.7/go.mod h1:u+2+/6zg+i71rQMx5EYifcz6MCKuco9NR6JIITiCfzQ= golang.org/x/text v0.3.8/go.mod h1:E6s5w1FMmriuDzIBO73fBruAKo1PCIq6d2Q6DHfQ8WQ= diff --git a/hack/lib/golang.sh b/hack/lib/golang.sh index fd111fea3cdae..356642b3ed923 100755 --- a/hack/lib/golang.sh +++ b/hack/lib/golang.sh @@ -80,6 +80,7 @@ kube::golang::server_targets() { cluster/gce/gci/mounter cmd/watch-termination openshift-hack/cmd/k8s-tests + openshift-hack/cmd/k8s-tests-ext ) echo "${targets[@]}" } diff --git a/openshift-hack/cmd/k8s-tests-ext/k8s-tests.go b/openshift-hack/cmd/k8s-tests-ext/k8s-tests.go new file mode 100644 index 0000000000000..5314527b92f09 --- /dev/null +++ b/openshift-hack/cmd/k8s-tests-ext/k8s-tests.go @@ -0,0 +1,105 @@ +package main + +import ( + "os" + + "github.com/spf13/cobra" + "github.com/spf13/pflag" + + "github.com/openshift-eng/openshift-tests-extension/pkg/cmd" + e "github.com/openshift-eng/openshift-tests-extension/pkg/extension" + "github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests" + g "github.com/openshift-eng/openshift-tests-extension/pkg/ginkgo" + v "github.com/openshift-eng/openshift-tests-extension/pkg/version" + + "k8s.io/client-go/pkg/version" + utilflag "k8s.io/component-base/cli/flag" + "k8s.io/component-base/logs" + "k8s.io/kubernetes/openshift-hack/e2e/annotate/generated" + + // initialize framework extensions + _ "k8s.io/kubernetes/test/e2e/framework/debug/init" + _ "k8s.io/kubernetes/test/e2e/framework/metrics/init" +) + +func main() { + logs.InitLogs() + defer logs.FlushLogs() + pflag.CommandLine.SetNormalizeFunc(utilflag.WordSepNormalizeFunc) + + // Get version info from kube + kubeVersion := version.Get() + v.GitTreeState = kubeVersion.GitTreeState + v.BuildDate = kubeVersion.BuildDate + v.CommitFromGit = kubeVersion.GitCommit + + // Create our registry of openshift-tests extensions + extensionRegistry := e.NewRegistry() + kubeTestsExtension := e.NewExtension("openshift", "payload", "hyperkube") + extensionRegistry.Register(kubeTestsExtension) + + // Carve up the kube tests into our openshift suites... + kubeTestsExtension.AddSuite(e.Suite{ + Name: "kubernetes/conformance/parallel", + Parents: []string{ + "openshift/conformance/parallel", + "openshift/conformance/parallel/minimal", + }, + Qualifiers: []string{`!labels.exists(l, l == "Serial") && labels.exists(l, l == "Conformance")`}, + }) + + kubeTestsExtension.AddSuite(e.Suite{ + Name: "kubernetes/conformance/serial", + Parents: []string{ + "openshift/conformance/serial", + "openshift/conformance/serial/minimal", + }, + Qualifiers: []string{`labels.exists(l, l == "Serial") && labels.exists(l, l == "Conformance")`}, + }) + + //FIXME(stbenjam): what other suites does k8s-test contribute to? + + // Build our specs from ginkgo + specs, err := g.BuildExtensionTestSpecsFromOpenShiftGinkgoSuite() + if err != nil { + panic(err) + } + + // Initialization for kube ginkgo test framework needs to run before all tests execute + specs.AddBeforeAll(func() { + if err := initializeTestFramework(os.Getenv("TEST_PROVIDER")); err != nil { + panic(err) + } + }) + + // Annotations get appended to test names, these are additions to upstream + // tests for controlling skips, suite membership, etc. + // + // TODO: + // - Remove this annotation code, and migrate to Labels/Tags and + // the environmental skip code from the enhancement once its implemented. + // - Make sure to account for test renames that occur because of removal of these + // annotations + specs.Walk(func(spec *extensiontests.ExtensionTestSpec) { + if annotations, ok := generated.Annotations[spec.Name]; ok { + spec.Name += annotations + } + }) + + kubeTestsExtension.AddSpecs(specs) + + // Cobra stuff + root := &cobra.Command{ + Long: "Kubernetes tests extension for OpenShift", + } + + root.AddCommand( + cmd.DefaultExtensionCommands(extensionRegistry)..., + ) + + if err := func() error { + return root.Execute() + }(); err != nil { + os.Exit(1) + } +} diff --git a/openshift-hack/cmd/k8s-tests-ext/provider.go b/openshift-hack/cmd/k8s-tests-ext/provider.go new file mode 100644 index 0000000000000..cdc948a45c652 --- /dev/null +++ b/openshift-hack/cmd/k8s-tests-ext/provider.go @@ -0,0 +1,147 @@ +package main + +import ( + "context" + "encoding/json" + "fmt" + "os" + "path/filepath" + "strings" + + "github.com/onsi/ginkgo/v2" + "github.com/onsi/gomega" + + corev1 "k8s.io/api/core/v1" + kclientset "k8s.io/client-go/kubernetes" + "k8s.io/client-go/tools/clientcmd" + "k8s.io/kubernetes/openshift-hack/e2e" + conformancetestdata "k8s.io/kubernetes/test/conformance/testdata" + "k8s.io/kubernetes/test/e2e/framework" + "k8s.io/kubernetes/test/e2e/framework/testfiles" + "k8s.io/kubernetes/test/e2e/storage/external" + e2etestingmanifests "k8s.io/kubernetes/test/e2e/testing-manifests" + testfixtures "k8s.io/kubernetes/test/fixtures" + + // this appears to inexplicably auto-register global flags. + _ "k8s.io/kubernetes/test/e2e/storage/drivers" + + // these are loading important global flags that we need to get and set + _ "k8s.io/kubernetes/test/e2e" + _ "k8s.io/kubernetes/test/e2e/lifecycle" +) + +// copied directly from github.com/openshift/origin/cmd/openshift-tests/provider.go +// and github.com/openshift/origin/test/extended/util/test.go +func initializeTestFramework(provider string) error { + providerInfo := &ClusterConfiguration{} + if err := json.Unmarshal([]byte(provider), &providerInfo); err != nil { + return fmt.Errorf("provider must be a JSON object with the 'type' key at a minimum: %v", err) + } + if len(providerInfo.ProviderName) == 0 { + return fmt.Errorf("provider must be a JSON object with the 'type' key") + } + config := &ClusterConfiguration{} + if err := json.Unmarshal([]byte(provider), config); err != nil { + return fmt.Errorf("provider must decode into the ClusterConfig object: %v", err) + } + + // update testContext with loaded config + testContext := &framework.TestContext + testContext.Provider = config.ProviderName + testContext.CloudConfig = framework.CloudConfig{ + ProjectID: config.ProjectID, + Region: config.Region, + Zone: config.Zone, + Zones: config.Zones, + NumNodes: config.NumNodes, + MultiMaster: config.MultiMaster, + MultiZone: config.MultiZone, + ConfigFile: config.ConfigFile, + } + testContext.AllowedNotReadyNodes = -1 + testContext.MinStartupPods = -1 + testContext.MaxNodesToGather = 0 + testContext.KubeConfig = os.Getenv("KUBECONFIG") + + // allow the CSI tests to access test data, but only briefly + // TODO: ideally CSI would not use any of these test methods + // var err error + // exutil.WithCleanup(func() { err = initCSITests(dryRun) }) + // TODO: for now I'm only initializing CSI directly, but we probably need that + // WithCleanup here as well + if err := initCSITests(); err != nil { + return err + } + + if ad := os.Getenv("ARTIFACT_DIR"); len(strings.TrimSpace(ad)) == 0 { + os.Setenv("ARTIFACT_DIR", filepath.Join(os.TempDir(), "artifacts")) + } + + testContext.DeleteNamespace = os.Getenv("DELETE_NAMESPACE") != "false" + testContext.VerifyServiceAccount = true + testfiles.AddFileSource(e2etestingmanifests.GetE2ETestingManifestsFS()) + testfiles.AddFileSource(testfixtures.GetTestFixturesFS()) + testfiles.AddFileSource(conformancetestdata.GetConformanceTestdataFS()) + testContext.KubectlPath = "kubectl" + // context.KubeConfig = KubeConfigPath() + testContext.KubeConfig = os.Getenv("KUBECONFIG") + + // "debian" is used when not set. At least GlusterFS tests need "custom". + // (There is no option for "rhel" or "centos".) + testContext.NodeOSDistro = "custom" + testContext.MasterOSDistro = "custom" + + // load and set the host variable for kubectl + clientConfig := clientcmd.NewNonInteractiveDeferredLoadingClientConfig(&clientcmd.ClientConfigLoadingRules{ExplicitPath: testContext.KubeConfig}, &clientcmd.ConfigOverrides{}) + cfg, err := clientConfig.ClientConfig() + if err != nil { + return err + } + testContext.Host = cfg.Host + + // Ensure that Kube tests run privileged (like they do upstream) + testContext.CreateTestingNS = func(ctx context.Context, baseName string, c kclientset.Interface, labels map[string]string) (*corev1.Namespace, error) { + return e2e.CreateTestingNS(ctx, baseName, c, labels, true) + } + + gomega.RegisterFailHandler(ginkgo.Fail) + + framework.AfterReadingAllFlags(testContext) + testContext.DumpLogsOnFailure = true + + // these constants are taken from kube e2e and used by tests + testContext.IPFamily = "ipv4" + if config.HasIPv6 && !config.HasIPv4 { + testContext.IPFamily = "ipv6" + } + + testContext.ReportDir = os.Getenv("TEST_JUNIT_DIR") + + return nil +} + +const ( + manifestEnvVar = "TEST_CSI_DRIVER_FILES" +) + +// copied directly from github.com/openshift/origin/cmd/openshift-tests/csi.go +// Initialize openshift/csi suite, i.e. define CSI tests from TEST_CSI_DRIVER_FILES. +func initCSITests() error { + manifestList := os.Getenv(manifestEnvVar) + if manifestList != "" { + manifests := strings.Split(manifestList, ",") + for _, manifest := range manifests { + if err := external.AddDriverDefinition(manifest); err != nil { + return fmt.Errorf("failed to load manifest from %q: %s", manifest, err) + } + // Register the base dir of the manifest file as a file source. + // With this we can reference the CSI driver's storageClass + // in the manifest file (FromFile field). + testfiles.AddFileSource(testfiles.RootFileSource{ + Root: filepath.Dir(manifest), + }) + } + } + + return nil +} diff --git a/openshift-hack/cmd/k8s-tests-ext/types.go b/openshift-hack/cmd/k8s-tests-ext/types.go new file mode 100644 index 0000000000000..b43652499537d --- /dev/null +++ b/openshift-hack/cmd/k8s-tests-ext/types.go @@ -0,0 +1,47 @@ +package main + +// copied directly from github.com/openshift/origin/test/extended/util/cluster/cluster.go +type ClusterConfiguration struct { + ProviderName string `json:"type"` + + // These fields (and the "type" tag for ProviderName) chosen to match + // upstream's e2e.CloudConfig. + ProjectID string + Region string + Zone string + NumNodes int + MultiMaster bool + MultiZone bool + Zones []string + ConfigFile string + + // Disconnected is set for test jobs without external internet connectivity + Disconnected bool + + // SingleReplicaTopology is set for disabling disruptive tests or tests + // that require high availability + SingleReplicaTopology bool + + // NetworkPlugin is the "official" plugin name + NetworkPlugin string + // NetworkPluginMode is an optional sub-identifier for the NetworkPlugin. + // (Currently it is only used for OpenShiftSDN.) + NetworkPluginMode string `json:",omitempty"` + + // HasIPv4 and HasIPv6 determine whether IPv4-specific, IPv6-specific, + // and dual-stack-specific tests are run + HasIPv4 bool + HasIPv6 bool + + // HasSCTP determines whether SCTP connectivity tests can be run in the cluster + HasSCTP bool + + // IsProxied determines whether we are accessing the cluster through an HTTP proxy + IsProxied bool + + // IsIBMROKS determines whether the cluster is Managed IBM Cloud (ROKS) + IsIBMROKS bool + + // IsNoOptionalCapabilities indicates the cluster has no optional capabilities enabled + HasNoOptionalCapabilities bool +} diff --git a/openshift-hack/e2e/annotate/rules_test.go b/openshift-hack/e2e/annotate/rules_test.go index b929a573d9020..6a413a2ef92ac 100644 --- a/openshift-hack/e2e/annotate/rules_test.go +++ b/openshift-hack/e2e/annotate/rules_test.go @@ -22,6 +22,10 @@ func (n *testNode) AppendText(text string) { n.text += text } +func (n *testNode) Labels() []string { + return nil +} + func TestStockRules(t *testing.T) { tests := []struct { name string diff --git a/openshift-hack/images/hyperkube/Dockerfile.rhel b/openshift-hack/images/hyperkube/Dockerfile.rhel index e133800481f39..74ef5d6e7b2e4 100644 --- a/openshift-hack/images/hyperkube/Dockerfile.rhel +++ b/openshift-hack/images/hyperkube/Dockerfile.rhel @@ -1,11 +1,12 @@ FROM registry.ci.openshift.org/ocp/builder:rhel-9-golang-1.22-openshift-4.18 AS builder WORKDIR /go/src/k8s.io/kubernetes COPY . . -RUN make WHAT='cmd/kube-apiserver cmd/kube-controller-manager cmd/kube-scheduler cmd/kubelet cmd/watch-termination openshift-hack/cmd/k8s-tests' && \ +RUN make WHAT='cmd/kube-apiserver cmd/kube-controller-manager cmd/kube-scheduler cmd/kubelet cmd/watch-termination openshift-hack/cmd/k8s-tests openshift-hack/cmd/k8s-tests-ext' && \ mkdir -p /tmp/build && \ cp openshift-hack/images/hyperkube/hyperkube openshift-hack/images/hyperkube/kubensenter /tmp/build && \ - cp /go/src/k8s.io/kubernetes/_output/local/bin/linux/$(go env GOARCH)/{kube-apiserver,kube-controller-manager,kube-scheduler,kubelet,watch-termination,k8s-tests} \ - /tmp/build + cp /go/src/k8s.io/kubernetes/_output/local/bin/linux/$(go env GOARCH)/{kube-apiserver,kube-controller-manager,kube-scheduler,kubelet,watch-termination,k8s-tests,k8s-tests-ext} \ + /tmp/build && \ + gzip /tmp/build/k8s-tests-ext FROM registry.ci.openshift.org/ocp/4.18:base-rhel9 RUN yum install -y --setopt=tsflags=nodocs --setopt=skip_missing_names_on_install=False iproute && yum clean all diff --git a/staging/src/k8s.io/api/go.mod b/staging/src/k8s.io/api/go.mod index deb5688a2e380..c52509170b553 100644 --- a/staging/src/k8s.io/api/go.mod +++ b/staging/src/k8s.io/api/go.mod @@ -41,7 +41,7 @@ require ( replace ( github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apiextensions-apiserver => ../apiextensions-apiserver k8s.io/apimachinery => ../apimachinery diff --git a/staging/src/k8s.io/api/go.sum b/staging/src/k8s.io/api/go.sum index 0e294918c486f..f1b49733ce3b4 100644 --- a/staging/src/k8s.io/api/go.sum +++ b/staging/src/k8s.io/api/go.sum @@ -72,7 +72,7 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/websocket v1.5.0/go.mod h1:YR8l580nyteQvAITg2hZ9XVh4b55+EU/adAjf1fMHhE= @@ -123,12 +123,13 @@ github.com/opencontainers/go-digest v1.0.0/go.mod h1:0JzlMkj0TRzQZfJkVvzbP0HBR3I github.com/opencontainers/runc v1.1.13/go.mod h1:R016aXacfp/gwQBYw2FDGa9m+n6atbLWrYY8hNMT/sA= github.com/opencontainers/runtime-spec v1.0.3-0.20220909204839-494a5a6aca78/go.mod h1:jwyrGlmzljRJv/Fgzds9SsS/C5hL+LL3ko9hs6T5lQ0= github.com/opencontainers/selinux v1.11.0/go.mod h1:E5dMC3VPuVvVHDYmi78qvhJp8+M586T4DlDRYpFkyec= +github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf/go.mod h1:1OhaNsaU9vuy/dlYZLEve7bgE2Ed+yTV5VSbYvGXt4s= github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f/go.mod h1:Shkl4HanLwDiiBzakv+con/aMGnVE2MAGvoKp5oyYUo= github.com/openshift/apiserver-library-go v0.0.0-20241001175710-6064b62894a6/go.mod h1:9Anrq7+DZmmw1Brchx4zmh26hAZbe6Dv7bGXRclnhYI= github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f/go.mod h1:KiZi2mJRH1TOJ3FtBDYS6YvUL30s/iIXaGSUrSa36mo= github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2/go.mod h1:s6Fqwb2KiWG6leCegVhw4KW40tf9f7m+SF1aXiE8Wsk= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/diff v0.0.0-20210226163009-20ebb0f2a09e/go.mod h1:pJLUxLENpZxwdsKMEsNbx1VGcRFpLqf3715MtcvvzbA= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= @@ -198,14 +199,14 @@ golang.org/x/crypto v0.27.0/go.mod h1:1Xngt8kV6Dvbssa53Ziq6Eqn0HqbZi5Z6R0ZpwQzt7 golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc/go.mod h1:V1LtkGg67GoY2N1AnLN78QLrzxkLyJw7RJb1gzOOz9w= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -224,15 +225,15 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/apiextensions-apiserver/go.mod b/staging/src/k8s.io/apiextensions-apiserver/go.mod index 90d28872a4604..442595cbfb6ae 100644 --- a/staging/src/k8s.io/apiextensions-apiserver/go.mod +++ b/staging/src/k8s.io/apiextensions-apiserver/go.mod @@ -20,7 +20,7 @@ require ( go.etcd.io/etcd/client/v3 v3.5.14 go.opentelemetry.io/otel v1.28.0 go.opentelemetry.io/otel/trace v1.28.0 - google.golang.org/grpc v1.65.0 + google.golang.org/grpc v1.67.0 google.golang.org/protobuf v1.34.2 gopkg.in/evanphx/json-patch.v4 v4.12.0 gopkg.in/yaml.v2 v2.4.0 @@ -109,18 +109,18 @@ require ( go.uber.org/zap v1.26.0 // indirect golang.org/x/crypto v0.27.0 // indirect golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc // indirect - golang.org/x/mod v0.17.0 // indirect + golang.org/x/mod v0.20.0 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect + golang.org/x/tools v0.24.0 // indirect google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect gopkg.in/inf.v0 v0.9.1 // indirect gopkg.in/natefinch/lumberjack.v2 v2.2.1 // indirect gopkg.in/yaml.v3 v3.0.1 // indirect @@ -131,7 +131,7 @@ require ( replace ( github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apiextensions-apiserver => ../apiextensions-apiserver k8s.io/apimachinery => ../apimachinery diff --git a/staging/src/k8s.io/apiextensions-apiserver/go.sum b/staging/src/k8s.io/apiextensions-apiserver/go.sum index 85ba0d41d7f57..9a8cf8943eba1 100644 --- a/staging/src/k8s.io/apiextensions-apiserver/go.sum +++ b/staging/src/k8s.io/apiextensions-apiserver/go.sum @@ -1,5 +1,5 @@ bitbucket.org/bertimus9/systemstat v0.5.0/go.mod h1:EkUWPp8lKFPMXP8vnbpT5JDI0W/sTiLZAvN8ONWErHY= -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= cloud.google.com/go v0.26.0/go.mod h1:aQUYkXzVsufM+DwF1aE+0xfcU+56JwCaLick0ClmMTw= cloud.google.com/go v0.34.0/go.mod h1:aQUYkXzVsufM+DwF1aE+0xfcU+56JwCaLick0ClmMTw= cloud.google.com/go v0.110.7/go.mod h1:+EYjdK8e5RME/VY/qLCAtuyALQ9q67dvuum8i+H5xsI= @@ -28,7 +28,7 @@ cloud.google.com/go/cloudbuild v1.13.0/go.mod h1:lyJg7v97SUIPq4RC2sGsz/9tNczhyv2 cloud.google.com/go/clouddms v1.6.1/go.mod h1:Ygo1vL52Ov4TBZQquhz5fiw2CQ58gvu+PlS6PVXCpZI= cloud.google.com/go/cloudtasks v1.12.1/go.mod h1:a9udmnou9KO2iulGscKR0qBYjreuX8oHwpmFsKspEvM= cloud.google.com/go/compute v1.23.0/go.mod h1:4tCnrn48xsqlwSAiLf1HXMQk8CONslYbdiEZc9FEIbM= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= cloud.google.com/go/contactcenterinsights v1.10.0/go.mod h1:bsg/R7zGLYMVxFFzfh9ooLTruLRCG9fnzhH9KznHhbM= cloud.google.com/go/container v1.24.0/go.mod h1:lTNExE2R7f+DLbAN+rJiKTisauFCaoDq6NURZ83eVH4= cloud.google.com/go/containeranalysis v0.10.1/go.mod h1:Ya2jiILITMY68ZLPaogjmOMNkwsDrWBSTyBubGXO7j0= @@ -154,7 +154,7 @@ github.com/checkpoint-restore/go-criu/v5 v5.3.0/go.mod h1:E/eQpaFtUKGOOSEBZgmKAc github.com/cilium/ebpf v0.9.1/go.mod h1:+OhNOIXx/Fnu1IE8bJz2dzOA+VSfyTfdNUVdlQnxUFY= github.com/client9/misspell v0.3.4/go.mod h1:qj6jICC3Q7zFZvVWo7KLAzC3yx5G7kyvSDkc90ppPyw= github.com/cncf/udpa/go v0.0.0-20191209042840-269d4d468f6f/go.mod h1:M8M6+tZqaGXZJjfX53e64911xZQV5JYwmTeXPW+k8Sc= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/cockroachdb/datadriven v1.0.2 h1:H9MtNqVoVhvd9nCBwOyDjUEdZCREqbIdCJD93PBm/jA= github.com/cockroachdb/datadriven v1.0.2/go.mod h1:a9RdTaap04u637JoCzcUoIcDmvwSUtcUFtT/C3kJlTU= github.com/container-storage-interface/spec v1.9.0/go.mod h1:ZfDu+3ZRyeVqxZM0Ds19MVLkN2d1XJ5MAfi1L3VjlT0= @@ -189,9 +189,9 @@ github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRr github.com/envoyproxy/go-control-plane v0.9.0/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymFceY/DCBVvsKhRF0iEA4= github.com/envoyproxy/go-control-plane v0.9.1-0.20191026205805-5f8ba28d4473/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymFceY/DCBVvsKhRF0iEA4= github.com/envoyproxy/go-control-plane v0.9.4/go.mod h1:6rpuAdCZL397s3pYoYcLgu1mIlRU8Am5FuJP05cCM98= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= github.com/envoyproxy/protoc-gen-validate v0.1.0/go.mod h1:iSmxcyjqTsJpI2R4NaDN7+kN2VEUnK/pcBlmesArF7c= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/euank/go-kmsg-parser v2.0.0+incompatible/go.mod h1:MhmAMZ8V4CYH4ybgdRwPr2TU5ThnS43puaKEMpja1uw= github.com/evanphx/json-patch v4.12.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= github.com/exponent-io/jsonpath v0.0.0-20151013193312-d6023ce2651d/go.mod h1:ZZMPRZwes7CROmyNKgQzC3XPs6L/G2EJLHddWejkmf4= @@ -236,7 +236,7 @@ github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69 github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= github.com/golang/glog v0.0.0-20160126235308-23def4e6c14b/go.mod h1:SBH7ygxi8pfUlaOkMMuAQtPIUF8ecWP5IEl/CR7VP2Q= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/mock v1.1.1/go.mod h1:oTYuIxOrZwtPieC+H1uAHpcLFnEyAGVDL/k47Jfbm0A= @@ -265,8 +265,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.1.2/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= @@ -343,6 +343,7 @@ github.com/opencontainers/image-spec v1.0.2/go.mod h1:BtxoFyWECRxE4U/7sNtV5W15zM github.com/opencontainers/runc v1.1.13/go.mod h1:R016aXacfp/gwQBYw2FDGa9m+n6atbLWrYY8hNMT/sA= github.com/opencontainers/runtime-spec v1.0.3-0.20220909204839-494a5a6aca78/go.mod h1:jwyrGlmzljRJv/Fgzds9SsS/C5hL+LL3ko9hs6T5lQ0= github.com/opencontainers/selinux v1.11.0/go.mod h1:E5dMC3VPuVvVHDYmi78qvhJp8+M586T4DlDRYpFkyec= +github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf/go.mod h1:1OhaNsaU9vuy/dlYZLEve7bgE2Ed+yTV5VSbYvGXt4s= github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f h1:ya1OmyZm3LIIxI3U9VE9Nyx3ehCHgBwxyFUPflYPWls= github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f/go.mod h1:Shkl4HanLwDiiBzakv+con/aMGnVE2MAGvoKp5oyYUo= github.com/openshift/apiserver-library-go v0.0.0-20241001175710-6064b62894a6/go.mod h1:9Anrq7+DZmmw1Brchx4zmh26hAZbe6Dv7bGXRclnhYI= @@ -351,14 +352,15 @@ github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f/go.mod h1:KiZi github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2/go.mod h1:s6Fqwb2KiWG6leCegVhw4KW40tf9f7m+SF1aXiE8Wsk= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060 h1:jiDC7d8d+jmjv2WfiMY0+Uf55q11MGyYkGGqXnfqWTU= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/opentracing/opentracing-go v1.1.0/go.mod h1:UkNAQd3GIcIGf0SeVgPpRdFStlNbqXla1AfSYxPUl2o= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.8.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/profile v1.7.0/go.mod h1:8Uer0jas47ZQMJ7VD+OHknK4YDY07LPUC6dEvqDjvNo= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -483,8 +485,8 @@ golang.org/x/lint v0.0.0-20190227174305-5b3e6a55c961/go.mod h1:wehouNa3lNwaWXcvx golang.org/x/lint v0.0.0-20190313153728-d0100b6bd8b3/go.mod h1:6SW0HCj/g11FgYtHlgUYUwCkIfeOF89ocIRzGO/8vkc= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0 h1:zY54UmvipHiNd+pm+m0x9KhZ9hl1/7QNMyxXbc6ICqA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0 h1:utOm6MM3R3dnawAiJgn0y+xvuYRsm1RKM/4giyfDgV0= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= @@ -501,8 +503,8 @@ golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U= golang.org/x/oauth2 v0.0.0-20200107190931-bf48bf16ab8d/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181108010431-42b317875d0f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181221193216-37e7f081c4d4/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -524,7 +526,7 @@ golang.org/x/sys v0.0.0-20210510120138-977fb7262007/go.mod h1:oPkhp1MJrh7nUepCBc golang.org/x/sys v0.0.0-20220715151400-c0bba94af5f8/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.0.0-20201126162022-7de9c90e9dd1/go.mod h1:bj7SfCRtBDWHUb9snDiAeCFNEtKQo2Wmx5Cou7ajbmo= golang.org/x/term v0.24.0 h1:Mh5cbb+Zk2hqqXNO7S1iTjEphVL+jb8ZWaqh/g+JWkM= golang.org/x/term v0.24.0/go.mod h1:lOBK/LVxemqiMij05LGJ0tzNr8xlmwBRJ81PX6wVLH8= @@ -543,8 +545,8 @@ golang.org/x/tools v0.0.0-20190524140312-2c0ae7006135/go.mod h1:RgjU9mgBXZiqYHBn golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -559,18 +561,18 @@ google.golang.org/genproto v0.0.0-20200423170343-7949de9c1215/go.mod h1:55QSHmfG google.golang.org/genproto v0.0.0-20200513103714-09dca8ec2884/go.mod h1:55QSHmfGQM9UVYDPBsyGGes0y52j32PQ3BqQfXhyH3c= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d h1:VBu5YqKPv6XiJ199exd8Br+Aetz+o08F+PLMnwJQHAY= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= google.golang.org/grpc v1.19.0/go.mod h1:mqu4LbDTu4XGKhr4mRzUsmM4RtVoemTSY81AxZiDr8c= google.golang.org/grpc v1.23.0/go.mod h1:Y5yQAOtifL1yxbo5wqy6BxZv8vAUGQwXBOALyacEbxg= google.golang.org/grpc v1.25.1/go.mod h1:c3i+UQWmh7LiEpx4sFZnkU36qjEYZ0imhYfXVyQciAY= google.golang.org/grpc v1.27.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk= google.golang.org/grpc v1.29.1/go.mod h1:itym6AZVZYACWQqET3MqgPpjcuV5QH3BxFS3IjizoKk= google.golang.org/grpc v1.33.1/go.mod h1:fr5YgcSWrqhRRxogOsw7RzIpsmvOZ6IcH4kBYTpR3n0= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/apimachinery/go.mod b/staging/src/k8s.io/apimachinery/go.mod index 804d103e3cc42..c693bf256b730 100644 --- a/staging/src/k8s.io/apimachinery/go.mod +++ b/staging/src/k8s.io/apimachinery/go.mod @@ -16,7 +16,7 @@ require ( github.com/google/uuid v1.6.0 github.com/moby/spdystream v0.4.0 github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f - github.com/onsi/ginkgo/v2 v2.17.2 + github.com/onsi/ginkgo/v2 v2.20.2 github.com/spf13/pflag v1.0.5 github.com/stretchr/testify v1.9.0 golang.org/x/net v0.29.0 @@ -37,7 +37,7 @@ require ( github.com/go-openapi/jsonreference v0.20.2 // indirect github.com/go-openapi/swag v0.22.4 // indirect github.com/go-task/slim-sprig/v3 v3.0.0 // indirect - github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 // indirect + github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 // indirect github.com/josharian/intern v1.0.0 // indirect github.com/json-iterator/go v1.1.12 // indirect github.com/kr/pretty v0.3.1 // indirect @@ -50,13 +50,13 @@ require ( github.com/x448/float16 v0.8.4 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/text v0.18.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect + golang.org/x/tools v0.24.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/yaml.v2 v2.4.0 // indirect gopkg.in/yaml.v3 v3.0.1 // indirect ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/apimachinery => ../apimachinery ) diff --git a/staging/src/k8s.io/apimachinery/go.sum b/staging/src/k8s.io/apimachinery/go.sum index 6011bee606a61..b4168cbe4bff1 100644 --- a/staging/src/k8s.io/apimachinery/go.sum +++ b/staging/src/k8s.io/apimachinery/go.sum @@ -34,8 +34,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/ianlancetaylor/demangle v0.0.0-20240312041847-bd984b5ce465/go.mod h1:gx7rwoVhcfuVKG5uya9Hs3Sxj7EIvldVofAWIUtGouw= @@ -66,8 +66,8 @@ github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f h1:y5//uYreIhSUg3J github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/pkg/diff v0.0.0-20210226163009-20ebb0f2a09e/go.mod h1:pJLUxLENpZxwdsKMEsNbx1VGcRFpLqf3715MtcvvzbA= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= @@ -100,7 +100,7 @@ golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9/go.mod h1:LzIPMQfyMNhhGPh golang.org/x/crypto v0.27.0/go.mod h1:1Xngt8kV6Dvbssa53Ziq6Eqn0HqbZi5Z6R0ZpwQzt70= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= @@ -116,7 +116,7 @@ golang.org/x/sys v0.0.0-20190412213103-97732733099d/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20200930185726-fdedc70b468f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.24.0/go.mod h1:lOBK/LVxemqiMij05LGJ0tzNr8xlmwBRJ81PX6wVLH8= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= golang.org/x/text v0.3.3/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ= @@ -128,8 +128,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/staging/src/k8s.io/apiserver/go.mod b/staging/src/k8s.io/apiserver/go.mod index dd9cea42b2acd..d2b07c87355d9 100644 --- a/staging/src/k8s.io/apiserver/go.mod +++ b/staging/src/k8s.io/apiserver/go.mod @@ -41,8 +41,8 @@ require ( golang.org/x/sync v0.8.0 golang.org/x/sys v0.25.0 golang.org/x/time v0.3.0 - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 - google.golang.org/grpc v1.65.0 + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f + google.golang.org/grpc v1.67.0 google.golang.org/protobuf v1.34.2 gopkg.in/evanphx/json-patch.v4 v4.12.0 gopkg.in/natefinch/lumberjack.v2 v2.2.1 @@ -116,11 +116,11 @@ require ( go.opentelemetry.io/proto/otlp v1.3.1 // indirect go.uber.org/multierr v1.11.0 // indirect golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect gopkg.in/inf.v0 v0.9.1 // indirect gopkg.in/yaml.v2 v2.4.0 // indirect gopkg.in/yaml.v3 v3.0.1 // indirect @@ -128,7 +128,7 @@ require ( replace ( github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apiextensions-apiserver => ../apiextensions-apiserver k8s.io/apimachinery => ../apimachinery diff --git a/staging/src/k8s.io/apiserver/go.sum b/staging/src/k8s.io/apiserver/go.sum index 0caec638978bf..08f8911275ab0 100644 --- a/staging/src/k8s.io/apiserver/go.sum +++ b/staging/src/k8s.io/apiserver/go.sum @@ -1,5 +1,5 @@ bitbucket.org/bertimus9/systemstat v0.5.0/go.mod h1:EkUWPp8lKFPMXP8vnbpT5JDI0W/sTiLZAvN8ONWErHY= -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= cloud.google.com/go v0.26.0/go.mod h1:aQUYkXzVsufM+DwF1aE+0xfcU+56JwCaLick0ClmMTw= cloud.google.com/go v0.34.0/go.mod h1:aQUYkXzVsufM+DwF1aE+0xfcU+56JwCaLick0ClmMTw= cloud.google.com/go v0.110.7/go.mod h1:+EYjdK8e5RME/VY/qLCAtuyALQ9q67dvuum8i+H5xsI= @@ -28,7 +28,7 @@ cloud.google.com/go/cloudbuild v1.13.0/go.mod h1:lyJg7v97SUIPq4RC2sGsz/9tNczhyv2 cloud.google.com/go/clouddms v1.6.1/go.mod h1:Ygo1vL52Ov4TBZQquhz5fiw2CQ58gvu+PlS6PVXCpZI= cloud.google.com/go/cloudtasks v1.12.1/go.mod h1:a9udmnou9KO2iulGscKR0qBYjreuX8oHwpmFsKspEvM= cloud.google.com/go/compute v1.23.0/go.mod h1:4tCnrn48xsqlwSAiLf1HXMQk8CONslYbdiEZc9FEIbM= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= cloud.google.com/go/contactcenterinsights v1.10.0/go.mod h1:bsg/R7zGLYMVxFFzfh9ooLTruLRCG9fnzhH9KznHhbM= cloud.google.com/go/container v1.24.0/go.mod h1:lTNExE2R7f+DLbAN+rJiKTisauFCaoDq6NURZ83eVH4= cloud.google.com/go/containeranalysis v0.10.1/go.mod h1:Ya2jiILITMY68ZLPaogjmOMNkwsDrWBSTyBubGXO7j0= @@ -154,7 +154,7 @@ github.com/checkpoint-restore/go-criu/v5 v5.3.0/go.mod h1:E/eQpaFtUKGOOSEBZgmKAc github.com/cilium/ebpf v0.9.1/go.mod h1:+OhNOIXx/Fnu1IE8bJz2dzOA+VSfyTfdNUVdlQnxUFY= github.com/client9/misspell v0.3.4/go.mod h1:qj6jICC3Q7zFZvVWo7KLAzC3yx5G7kyvSDkc90ppPyw= github.com/cncf/udpa/go v0.0.0-20191209042840-269d4d468f6f/go.mod h1:M8M6+tZqaGXZJjfX53e64911xZQV5JYwmTeXPW+k8Sc= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/cockroachdb/datadriven v1.0.2 h1:H9MtNqVoVhvd9nCBwOyDjUEdZCREqbIdCJD93PBm/jA= github.com/cockroachdb/datadriven v1.0.2/go.mod h1:a9RdTaap04u637JoCzcUoIcDmvwSUtcUFtT/C3kJlTU= github.com/container-storage-interface/spec v1.9.0/go.mod h1:ZfDu+3ZRyeVqxZM0Ds19MVLkN2d1XJ5MAfi1L3VjlT0= @@ -190,9 +190,9 @@ github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRr github.com/envoyproxy/go-control-plane v0.9.0/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymFceY/DCBVvsKhRF0iEA4= github.com/envoyproxy/go-control-plane v0.9.1-0.20191026205805-5f8ba28d4473/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymFceY/DCBVvsKhRF0iEA4= github.com/envoyproxy/go-control-plane v0.9.4/go.mod h1:6rpuAdCZL397s3pYoYcLgu1mIlRU8Am5FuJP05cCM98= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= github.com/envoyproxy/protoc-gen-validate v0.1.0/go.mod h1:iSmxcyjqTsJpI2R4NaDN7+kN2VEUnK/pcBlmesArF7c= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/euank/go-kmsg-parser v2.0.0+incompatible/go.mod h1:MhmAMZ8V4CYH4ybgdRwPr2TU5ThnS43puaKEMpja1uw= github.com/evanphx/json-patch v4.12.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= github.com/exponent-io/jsonpath v0.0.0-20151013193312-d6023ce2651d/go.mod h1:ZZMPRZwes7CROmyNKgQzC3XPs6L/G2EJLHddWejkmf4= @@ -237,7 +237,7 @@ github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69 github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= github.com/golang/glog v0.0.0-20160126235308-23def4e6c14b/go.mod h1:SBH7ygxi8pfUlaOkMMuAQtPIUF8ecWP5IEl/CR7VP2Q= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/mock v1.1.1/go.mod h1:oTYuIxOrZwtPieC+H1uAHpcLFnEyAGVDL/k47Jfbm0A= @@ -266,8 +266,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.1.2/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= @@ -344,6 +344,7 @@ github.com/opencontainers/image-spec v1.0.2/go.mod h1:BtxoFyWECRxE4U/7sNtV5W15zM github.com/opencontainers/runc v1.1.13/go.mod h1:R016aXacfp/gwQBYw2FDGa9m+n6atbLWrYY8hNMT/sA= github.com/opencontainers/runtime-spec v1.0.3-0.20220909204839-494a5a6aca78/go.mod h1:jwyrGlmzljRJv/Fgzds9SsS/C5hL+LL3ko9hs6T5lQ0= github.com/opencontainers/selinux v1.11.0/go.mod h1:E5dMC3VPuVvVHDYmi78qvhJp8+M586T4DlDRYpFkyec= +github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf/go.mod h1:1OhaNsaU9vuy/dlYZLEve7bgE2Ed+yTV5VSbYvGXt4s= github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f/go.mod h1:Shkl4HanLwDiiBzakv+con/aMGnVE2MAGvoKp5oyYUo= github.com/openshift/apiserver-library-go v0.0.0-20241001175710-6064b62894a6/go.mod h1:9Anrq7+DZmmw1Brchx4zmh26hAZbe6Dv7bGXRclnhYI= github.com/openshift/build-machinery-go v0.0.0-20240613134303-8359781da660/go.mod h1:8jcm8UPtg2mCAsxfqKil1xrmRMI3a+XU2TZ9fF8A7TE= @@ -351,14 +352,15 @@ github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f/go.mod h1:KiZi github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2/go.mod h1:s6Fqwb2KiWG6leCegVhw4KW40tf9f7m+SF1aXiE8Wsk= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060 h1:jiDC7d8d+jmjv2WfiMY0+Uf55q11MGyYkGGqXnfqWTU= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/opentracing/opentracing-go v1.1.0/go.mod h1:UkNAQd3GIcIGf0SeVgPpRdFStlNbqXla1AfSYxPUl2o= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.8.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/profile v1.7.0/go.mod h1:8Uer0jas47ZQMJ7VD+OHknK4YDY07LPUC6dEvqDjvNo= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -484,7 +486,7 @@ golang.org/x/lint v0.0.0-20190227174305-5b3e6a55c961/go.mod h1:wehouNa3lNwaWXcvx golang.org/x/lint v0.0.0-20190313153728-d0100b6bd8b3/go.mod h1:6SW0HCj/g11FgYtHlgUYUwCkIfeOF89ocIRzGO/8vkc= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= @@ -501,8 +503,8 @@ golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U= golang.org/x/oauth2 v0.0.0-20200107190931-bf48bf16ab8d/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181108010431-42b317875d0f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181221193216-37e7f081c4d4/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -542,8 +544,8 @@ golang.org/x/tools v0.0.0-20190524140312-2c0ae7006135/go.mod h1:RgjU9mgBXZiqYHBn golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -558,18 +560,18 @@ google.golang.org/genproto v0.0.0-20200423170343-7949de9c1215/go.mod h1:55QSHmfG google.golang.org/genproto v0.0.0-20200513103714-09dca8ec2884/go.mod h1:55QSHmfGQM9UVYDPBsyGGes0y52j32PQ3BqQfXhyH3c= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d h1:VBu5YqKPv6XiJ199exd8Br+Aetz+o08F+PLMnwJQHAY= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= google.golang.org/grpc v1.19.0/go.mod h1:mqu4LbDTu4XGKhr4mRzUsmM4RtVoemTSY81AxZiDr8c= google.golang.org/grpc v1.23.0/go.mod h1:Y5yQAOtifL1yxbo5wqy6BxZv8vAUGQwXBOALyacEbxg= google.golang.org/grpc v1.25.1/go.mod h1:c3i+UQWmh7LiEpx4sFZnkU36qjEYZ0imhYfXVyQciAY= google.golang.org/grpc v1.27.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk= google.golang.org/grpc v1.29.1/go.mod h1:itym6AZVZYACWQqET3MqgPpjcuV5QH3BxFS3IjizoKk= google.golang.org/grpc v1.33.1/go.mod h1:fr5YgcSWrqhRRxogOsw7RzIpsmvOZ6IcH4kBYTpR3n0= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/cli-runtime/go.mod b/staging/src/k8s.io/cli-runtime/go.mod index 92c0eabd3ea67..c2fb5d8fd61bb 100644 --- a/staging/src/k8s.io/cli-runtime/go.mod +++ b/staging/src/k8s.io/cli-runtime/go.mod @@ -62,7 +62,7 @@ require ( github.com/xlab/treeprint v1.2.0 // indirect go.starlark.net v0.0.0-20230525235612-a134d8f9ddca // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/time v0.3.0 // indirect @@ -74,7 +74,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/cli-runtime => ../cli-runtime diff --git a/staging/src/k8s.io/cli-runtime/go.sum b/staging/src/k8s.io/cli-runtime/go.sum index a7002f7088d0b..fadf48a0b0f1a 100644 --- a/staging/src/k8s.io/cli-runtime/go.sum +++ b/staging/src/k8s.io/cli-runtime/go.sum @@ -71,8 +71,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510 h1:El6M4kTTCOh6aBiKaUGG7oYTSPP8MxqL4YI3kZKwcP4= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= @@ -116,8 +116,8 @@ github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8m github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible h1:UBdAOUP5p4RWqPBg048CAvpKN+vxiaj6gdUUzhl4XmI= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= @@ -179,8 +179,8 @@ golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwY golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181108010431-42b317875d0f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -214,8 +214,8 @@ golang.org/x/tools v0.0.0-20190524140312-2c0ae7006135/go.mod h1:RgjU9mgBXZiqYHBn golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/staging/src/k8s.io/client-go/go.mod b/staging/src/k8s.io/client-go/go.mod index 6a5f0f322729d..2d59cfe679593 100644 --- a/staging/src/k8s.io/client-go/go.mod +++ b/staging/src/k8s.io/client-go/go.mod @@ -20,7 +20,7 @@ require ( github.com/stretchr/testify v1.9.0 go.uber.org/goleak v1.3.0 golang.org/x/net v0.29.0 - golang.org/x/oauth2 v0.21.0 + golang.org/x/oauth2 v0.22.0 golang.org/x/term v0.24.0 golang.org/x/time v0.3.0 google.golang.org/protobuf v1.34.2 @@ -45,7 +45,7 @@ require ( github.com/go-openapi/swag v0.22.4 // indirect github.com/go-task/slim-sprig/v3 v3.0.0 // indirect github.com/google/btree v1.0.1 // indirect - github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 // indirect + github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 // indirect github.com/josharian/intern v1.0.0 // indirect github.com/json-iterator/go v1.1.12 // indirect github.com/mailru/easyjson v0.7.7 // indirect @@ -54,20 +54,20 @@ require ( github.com/modern-go/reflect2 v1.0.2 // indirect github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822 // indirect github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f // indirect - github.com/onsi/ginkgo/v2 v2.17.2 // indirect + github.com/onsi/ginkgo/v2 v2.20.2 // indirect github.com/pkg/errors v0.9.1 // indirect github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 // indirect github.com/x448/float16 v0.8.4 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/text v0.18.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect + golang.org/x/tools v0.24.0 // indirect gopkg.in/inf.v0 v0.9.1 // indirect gopkg.in/yaml.v2 v2.4.0 // indirect gopkg.in/yaml.v3 v3.0.1 // indirect ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/client-go => ../client-go diff --git a/staging/src/k8s.io/client-go/go.sum b/staging/src/k8s.io/client-go/go.sum index b204a3b501777..f9d5eac9591f0 100644 --- a/staging/src/k8s.io/client-go/go.sum +++ b/staging/src/k8s.io/client-go/go.sum @@ -40,8 +40,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/websocket v1.5.0 h1:PPwGk2jz7EePpoHN/+ClbZu8SPxiqlu12wZP/3sWmnc= @@ -79,8 +79,8 @@ github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f h1:y5//uYreIhSUg3J github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible h1:UBdAOUP5p4RWqPBg048CAvpKN+vxiaj6gdUUzhl4XmI= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= @@ -115,15 +115,15 @@ golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9/go.mod h1:LzIPMQfyMNhhGPh golang.org/x/crypto v0.27.0/go.mod h1:1Xngt8kV6Dvbssa53Ziq6Eqn0HqbZi5Z6R0ZpwQzt70= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -133,7 +133,7 @@ golang.org/x/sys v0.0.0-20190412213103-97732733099d/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20200930185726-fdedc70b468f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.24.0 h1:Mh5cbb+Zk2hqqXNO7S1iTjEphVL+jb8ZWaqh/g+JWkM= golang.org/x/term v0.24.0/go.mod h1:lOBK/LVxemqiMij05LGJ0tzNr8xlmwBRJ81PX6wVLH8= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= @@ -146,8 +146,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/staging/src/k8s.io/cloud-provider/go.mod b/staging/src/k8s.io/cloud-provider/go.mod index 56b92c8492ef3..dbf0d9456a8cc 100644 --- a/staging/src/k8s.io/cloud-provider/go.mod +++ b/staging/src/k8s.io/cloud-provider/go.mod @@ -87,15 +87,15 @@ require ( golang.org/x/crypto v0.27.0 // indirect golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect - google.golang.org/grpc v1.65.0 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect + google.golang.org/grpc v1.67.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/evanphx/json-patch.v4 v4.12.0 // indirect gopkg.in/inf.v0 v0.9.1 // indirect @@ -111,7 +111,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apiextensions-apiserver => ../apiextensions-apiserver k8s.io/apimachinery => ../apimachinery diff --git a/staging/src/k8s.io/cloud-provider/go.sum b/staging/src/k8s.io/cloud-provider/go.sum index 3757d080c5229..3d71c4a1795da 100644 --- a/staging/src/k8s.io/cloud-provider/go.sum +++ b/staging/src/k8s.io/cloud-provider/go.sum @@ -1,5 +1,5 @@ -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1 h1:UQHMgLO+TxOElx5B5HZ4hJQsoJ/PvUvKRhJHDQXO8P8= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1/go.mod h1:xomTg63KZ2rFqZQzSB4Vz2SUXa1BpHTVz9L5PTmPC4E= github.com/Azure/go-ntlmssp v0.0.0-20211209120228-48547f28849e/go.mod h1:chxPXzSsl7ZWRAuOIE23GDNzjWuZquvFlgA8xmpunjU= @@ -23,7 +23,7 @@ github.com/cenkalti/backoff/v4 v4.3.0/go.mod h1:Y3VNntkOUPxTVeUxJ/G5vcM//AlwfmyY github.com/census-instrumentation/opencensus-proto v0.4.1/go.mod h1:4T9NM4+4Vw91VeyqjLS6ao50K5bOcLKN6Q42XnYaRYw= github.com/cespare/xxhash/v2 v2.3.0 h1:UL815xU9SqsFlibzuggzjXhog7bL6oX9BbNZnL2UFvs= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/coreos/go-oidc v2.2.1+incompatible/go.mod h1:CgnwVTmzoESiwO9qyAFEMiHoZ1nMCKZlZ9V6mm3/LKc= github.com/coreos/go-semver v0.3.1 h1:yi21YpKnrx1gt5R+la8n5WgS0kCrsPp33dmEyHReZr4= github.com/coreos/go-semver v0.3.1/go.mod h1:irMmmIw/7yzSRPWryHsK7EYSg09caPQL03VsM8rvUec= @@ -44,8 +44,8 @@ github.com/dustin/go-humanize v1.0.1 h1:GzkhY7T5VNhEkwH0PVJgjz+fX1rhBrR7pRT3mDkp github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto= github.com/emicklei/go-restful/v3 v3.11.0 h1:rAQeMHw1c7zTmncogyy8VvRZwtkmkZ4FxERmMY4rD+g= github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRry+4bhvbpWn3mrbc= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/evanphx/json-patch v4.12.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= github.com/felixge/fgprof v0.9.4/go.mod h1:yKl+ERSa++RYOs32d8K6WEXCB4uXdLls4ZaZPpayhMM= github.com/felixge/httpsnoop v1.0.4 h1:NFTV2Zj1bL4mc9sqWACXbQFVBBg2W3GPvqp8/ESS2Wg= @@ -80,7 +80,7 @@ github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= @@ -103,8 +103,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/mux v1.8.1/go.mod h1:AKf9I4AEqPTmMytcMc0KkNouC66V3BtZ4qD5fmWSiMQ= @@ -167,12 +167,13 @@ github.com/openshift/build-machinery-go v0.0.0-20240613134303-8359781da660/go.mo github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f/go.mod h1:KiZi2mJRH1TOJ3FtBDYS6YvUL30s/iIXaGSUrSa36mo= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060 h1:jiDC7d8d+jmjv2WfiMY0+Uf55q11MGyYkGGqXnfqWTU= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/profile v1.7.0/go.mod h1:8Uer0jas47ZQMJ7VD+OHknK4YDY07LPUC6dEvqDjvNo= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -279,8 +280,8 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -304,8 +305,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -314,12 +315,12 @@ golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNq google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d h1:VBu5YqKPv6XiJ199exd8Br+Aetz+o08F+PLMnwJQHAY= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/cluster-bootstrap/go.mod b/staging/src/k8s.io/cluster-bootstrap/go.mod index 834cf93daa5b7..cbb44dc552bb2 100644 --- a/staging/src/k8s.io/cluster-bootstrap/go.mod +++ b/staging/src/k8s.io/cluster-bootstrap/go.mod @@ -36,7 +36,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/cluster-bootstrap => ../cluster-bootstrap diff --git a/staging/src/k8s.io/cluster-bootstrap/go.sum b/staging/src/k8s.io/cluster-bootstrap/go.sum index ed1ae8d212227..f68bd3a1f1111 100644 --- a/staging/src/k8s.io/cluster-bootstrap/go.sum +++ b/staging/src/k8s.io/cluster-bootstrap/go.sum @@ -22,7 +22,7 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/josharian/intern v1.0.0/go.mod h1:5DoeVV0s6jJacbCEi61lwdGj/aVlrQvzHFFd8Hwg//Y= github.com/json-iterator/go v1.1.12 h1:PV8peI4a0ysnczrg+LtxykD8LfKY9ML6u2jnxaEnrnM= @@ -41,7 +41,7 @@ github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd/go.mod h1:6dJ github.com/modern-go/reflect2 v1.0.2 h1:xBagoLtFs94CBntxluKeaWgTMpvLxC4ur3nMaC9Gz0M= github.com/modern-go/reflect2 v1.0.2/go.mod h1:yWuevngMOJpCy52FWWMvUC8ws7m/LJsjYzDa0/r8luk= github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= @@ -91,7 +91,7 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/staging/src/k8s.io/code-generator/examples/go.mod b/staging/src/k8s.io/code-generator/examples/go.mod index 59229b11bd31f..36cb7d0c8940a 100644 --- a/staging/src/k8s.io/code-generator/examples/go.mod +++ b/staging/src/k8s.io/code-generator/examples/go.mod @@ -35,7 +35,7 @@ require ( github.com/pkg/errors v0.9.1 // indirect github.com/x448/float16 v0.8.4 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect @@ -59,4 +59,4 @@ replace ( replace github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 -replace github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 +replace github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 diff --git a/staging/src/k8s.io/code-generator/examples/go.sum b/staging/src/k8s.io/code-generator/examples/go.sum index 17cca057c1c1b..3f99ec042808e 100644 --- a/staging/src/k8s.io/code-generator/examples/go.sum +++ b/staging/src/k8s.io/code-generator/examples/go.sum @@ -30,8 +30,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/josharian/intern v1.0.0 h1:vlS4z54oSdjm0bgjRigI+G1HpF+tI+9rE5LLzOg8HmY= @@ -58,8 +58,8 @@ github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822 h1:C3w9PqII01/Oq github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8mK8HuQTcFwEeznm/DIxMOiR9yIdICNftLE1DvQ= github.com/onsi/gomega v1.19.0 h1:4ieX6qQjPP/BfC3mpsAtIGGlxTWPeA3Inl/7DtXw1tw= github.com/onsi/gomega v1.19.0/go.mod h1:LY+I3pBVzYsTBU1AnDwOSxaYi9WoWiqgwooUqq9yPro= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -93,8 +93,8 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -115,8 +115,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/staging/src/k8s.io/code-generator/go.mod b/staging/src/k8s.io/code-generator/go.mod index 0f6e283d3ec92..a717d227b78a7 100644 --- a/staging/src/k8s.io/code-generator/go.mod +++ b/staging/src/k8s.io/code-generator/go.mod @@ -34,9 +34,9 @@ require ( github.com/modern-go/reflect2 v1.0.2 // indirect github.com/onsi/gomega v1.33.1 // indirect github.com/x448/float16 v0.8.4 // indirect - golang.org/x/mod v0.17.0 // indirect + golang.org/x/mod v0.20.0 // indirect golang.org/x/sync v0.8.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect + golang.org/x/tools v0.24.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/yaml.v3 v3.0.1 // indirect sigs.k8s.io/json v0.0.0-20221116044647-bc3834ca7abd // indirect @@ -44,7 +44,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/apimachinery => ../apimachinery k8s.io/code-generator => ../code-generator ) diff --git a/staging/src/k8s.io/code-generator/go.sum b/staging/src/k8s.io/code-generator/go.sum index 35193ae0423f1..73b52b7895019 100644 --- a/staging/src/k8s.io/code-generator/go.sum +++ b/staging/src/k8s.io/code-generator/go.sum @@ -33,8 +33,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/josharian/intern v1.0.0 h1:vlS4z54oSdjm0bgjRigI+G1HpF+tI+9rE5LLzOg8HmY= github.com/josharian/intern v1.0.0/go.mod h1:5DoeVV0s6jJacbCEi61lwdGj/aVlrQvzHFFd8Hwg//Y= @@ -61,8 +61,8 @@ github.com/munnerz/goautoneg v0.0.0-20120707110453-a547fc61f48d/go.mod h1:+n7T8m github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= @@ -90,8 +90,8 @@ golang.org/x/crypto v0.0.0-20191011191535-87dc89f01550/go.mod h1:yigFU9vqHzYiE8U golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9/go.mod h1:LzIPMQfyMNhhGPhUkYOs5KpL4U8rLKemX1yGLhDgUto= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0 h1:zY54UmvipHiNd+pm+m0x9KhZ9hl1/7QNMyxXbc6ICqA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0 h1:utOm6MM3R3dnawAiJgn0y+xvuYRsm1RKM/4giyfDgV0= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= @@ -108,7 +108,7 @@ golang.org/x/sys v0.0.0-20190412213103-97732733099d/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20200930185726-fdedc70b468f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= golang.org/x/text v0.3.3/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ= golang.org/x/text v0.18.0 h1:XvMDiNzPAl0jr17s6W9lcaIhGUfUORdGCNsuLmPG224= @@ -118,8 +118,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/staging/src/k8s.io/component-base/go.mod b/staging/src/k8s.io/component-base/go.mod index dfacc96f8c6bd..9b1a569557187 100644 --- a/staging/src/k8s.io/component-base/go.mod +++ b/staging/src/k8s.io/component-base/go.mod @@ -65,13 +65,13 @@ require ( go.opentelemetry.io/proto/otlp v1.3.1 // indirect go.uber.org/multierr v1.11.0 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect - google.golang.org/grpc v1.65.0 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect + google.golang.org/grpc v1.67.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/inf.v0 v0.9.1 // indirect gopkg.in/yaml.v3 v3.0.1 // indirect @@ -83,7 +83,7 @@ require ( replace ( github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apiextensions-apiserver => ../apiextensions-apiserver k8s.io/apimachinery => ../apimachinery diff --git a/staging/src/k8s.io/component-base/go.sum b/staging/src/k8s.io/component-base/go.sum index e0f88c13393da..7641b85fd88b3 100644 --- a/staging/src/k8s.io/component-base/go.sum +++ b/staging/src/k8s.io/component-base/go.sum @@ -1,6 +1,6 @@ bitbucket.org/bertimus9/systemstat v0.5.0/go.mod h1:EkUWPp8lKFPMXP8vnbpT5JDI0W/sTiLZAvN8ONWErHY= -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1 h1:UQHMgLO+TxOElx5B5HZ4hJQsoJ/PvUvKRhJHDQXO8P8= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1/go.mod h1:xomTg63KZ2rFqZQzSB4Vz2SUXa1BpHTVz9L5PTmPC4E= github.com/Azure/go-ntlmssp v0.0.0-20211209120228-48547f28849e/go.mod h1:chxPXzSsl7ZWRAuOIE23GDNzjWuZquvFlgA8xmpunjU= @@ -28,7 +28,7 @@ github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XL github.com/chai2010/gettext-go v1.0.2/go.mod h1:y+wnP2cHYaVj19NZhYKAwEMH2CI1gNHeQQ+5AjwawxA= github.com/checkpoint-restore/go-criu/v5 v5.3.0/go.mod h1:E/eQpaFtUKGOOSEBZgmKAcn+zUUwWxqcaKZlF54wK8E= github.com/cilium/ebpf v0.9.1/go.mod h1:+OhNOIXx/Fnu1IE8bJz2dzOA+VSfyTfdNUVdlQnxUFY= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/container-storage-interface/spec v1.9.0/go.mod h1:ZfDu+3ZRyeVqxZM0Ds19MVLkN2d1XJ5MAfi1L3VjlT0= github.com/containerd/cgroups v1.1.0/go.mod h1:6ppBcbh/NOOUU+dMKrykgaBnK9lCIBxHqJDGwsa1mIw= github.com/containerd/console v1.0.3/go.mod h1:7LqA/THxQ86k76b8c/EMSiaJ3h1eZkMkXar0TQ1gf3U= @@ -53,8 +53,8 @@ github.com/docker/go-units v0.5.0/go.mod h1:fgPhTUdO+D/Jk86RDLlptpiXQzgHJF7gydDD github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto= github.com/emicklei/go-restful/v3 v3.11.0 h1:rAQeMHw1c7zTmncogyy8VvRZwtkmkZ4FxERmMY4rD+g= github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRry+4bhvbpWn3mrbc= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/euank/go-kmsg-parser v2.0.0+incompatible/go.mod h1:MhmAMZ8V4CYH4ybgdRwPr2TU5ThnS43puaKEMpja1uw= github.com/exponent-io/jsonpath v0.0.0-20151013193312-d6023ce2651d/go.mod h1:ZZMPRZwes7CROmyNKgQzC3XPs6L/G2EJLHddWejkmf4= github.com/fatih/camelcase v1.0.0/go.mod h1:yN2Sb0lFhZJUdVvtELVWefmrXpuZESvPmqwoZc+/fpc= @@ -88,7 +88,7 @@ github.com/godbus/dbus/v5 v5.1.0/go.mod h1:xhWf0FNVPg57R7Z0UbKHbJfkEywrmjJnf7w5x github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= @@ -103,8 +103,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= @@ -166,16 +166,18 @@ github.com/opencontainers/go-digest v1.0.0/go.mod h1:0JzlMkj0TRzQZfJkVvzbP0HBR3I github.com/opencontainers/runc v1.1.13/go.mod h1:R016aXacfp/gwQBYw2FDGa9m+n6atbLWrYY8hNMT/sA= github.com/opencontainers/runtime-spec v1.0.3-0.20220909204839-494a5a6aca78/go.mod h1:jwyrGlmzljRJv/Fgzds9SsS/C5hL+LL3ko9hs6T5lQ0= github.com/opencontainers/selinux v1.11.0/go.mod h1:E5dMC3VPuVvVHDYmi78qvhJp8+M586T4DlDRYpFkyec= +github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf/go.mod h1:1OhaNsaU9vuy/dlYZLEve7bgE2Ed+yTV5VSbYvGXt4s= github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f/go.mod h1:Shkl4HanLwDiiBzakv+con/aMGnVE2MAGvoKp5oyYUo= github.com/openshift/apiserver-library-go v0.0.0-20241001175710-6064b62894a6/go.mod h1:9Anrq7+DZmmw1Brchx4zmh26hAZbe6Dv7bGXRclnhYI= github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f/go.mod h1:KiZi2mJRH1TOJ3FtBDYS6YvUL30s/iIXaGSUrSa36mo= github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2/go.mod h1:s6Fqwb2KiWG6leCegVhw4KW40tf9f7m+SF1aXiE8Wsk= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -264,15 +266,15 @@ golang.org/x/crypto v0.27.0/go.mod h1:1Xngt8kV6Dvbssa53Ziq6Eqn0HqbZi5Z6R0ZpwQzt7 golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc/go.mod h1:V1LtkGg67GoY2N1AnLN78QLrzxkLyJw7RJb1gzOOz9w= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -295,8 +297,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -304,12 +306,12 @@ golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8T golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNqU5K3HdILfMLy1f3aNYFI/wnl100a8= google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/component-helpers/go.mod b/staging/src/k8s.io/component-helpers/go.mod index 5f25205945c3a..3833ea775eb53 100644 --- a/staging/src/k8s.io/component-helpers/go.mod +++ b/staging/src/k8s.io/component-helpers/go.mod @@ -37,7 +37,7 @@ require ( github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 // indirect github.com/x448/float16 v0.8.4 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect @@ -55,7 +55,7 @@ require ( replace ( github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apiextensions-apiserver => ../apiextensions-apiserver k8s.io/apimachinery => ../apimachinery diff --git a/staging/src/k8s.io/component-helpers/go.sum b/staging/src/k8s.io/component-helpers/go.sum index ef5fbb4abf155..51a82b2d83716 100644 --- a/staging/src/k8s.io/component-helpers/go.sum +++ b/staging/src/k8s.io/component-helpers/go.sum @@ -80,8 +80,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= @@ -137,13 +137,14 @@ github.com/opencontainers/go-digest v1.0.0/go.mod h1:0JzlMkj0TRzQZfJkVvzbP0HBR3I github.com/opencontainers/runc v1.1.13/go.mod h1:R016aXacfp/gwQBYw2FDGa9m+n6atbLWrYY8hNMT/sA= github.com/opencontainers/runtime-spec v1.0.3-0.20220909204839-494a5a6aca78/go.mod h1:jwyrGlmzljRJv/Fgzds9SsS/C5hL+LL3ko9hs6T5lQ0= github.com/opencontainers/selinux v1.11.0/go.mod h1:E5dMC3VPuVvVHDYmi78qvhJp8+M586T4DlDRYpFkyec= +github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf/go.mod h1:1OhaNsaU9vuy/dlYZLEve7bgE2Ed+yTV5VSbYvGXt4s= github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f/go.mod h1:Shkl4HanLwDiiBzakv+con/aMGnVE2MAGvoKp5oyYUo= github.com/openshift/apiserver-library-go v0.0.0-20241001175710-6064b62894a6/go.mod h1:9Anrq7+DZmmw1Brchx4zmh26hAZbe6Dv7bGXRclnhYI= github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f/go.mod h1:KiZi2mJRH1TOJ3FtBDYS6YvUL30s/iIXaGSUrSa36mo= github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2/go.mod h1:s6Fqwb2KiWG6leCegVhw4KW40tf9f7m+SF1aXiE8Wsk= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= @@ -217,15 +218,15 @@ golang.org/x/crypto v0.27.0/go.mod h1:1Xngt8kV6Dvbssa53Ziq6Eqn0HqbZi5Z6R0ZpwQzt7 golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc/go.mod h1:V1LtkGg67GoY2N1AnLN78QLrzxkLyJw7RJb1gzOOz9w= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -247,17 +248,17 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNqU5K3HdILfMLy1f3aNYFI/wnl100a8= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/controller-manager/go.mod b/staging/src/k8s.io/controller-manager/go.mod index 8f5d182927cca..1970ae775a079 100644 --- a/staging/src/k8s.io/controller-manager/go.mod +++ b/staging/src/k8s.io/controller-manager/go.mod @@ -7,7 +7,7 @@ go 1.22.0 require ( github.com/spf13/pflag v1.0.5 github.com/stretchr/testify v1.9.0 - golang.org/x/oauth2 v0.21.0 + golang.org/x/oauth2 v0.22.0 k8s.io/api v0.31.1 k8s.io/apimachinery v0.31.1 k8s.io/apiserver v0.31.1 @@ -88,9 +88,9 @@ require ( golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect - google.golang.org/grpc v1.65.0 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect + google.golang.org/grpc v1.67.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/evanphx/json-patch.v4 v4.12.0 // indirect gopkg.in/inf.v0 v0.9.1 // indirect @@ -104,7 +104,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/apiserver => ../apiserver diff --git a/staging/src/k8s.io/controller-manager/go.sum b/staging/src/k8s.io/controller-manager/go.sum index 38c94467c20aa..faa84a58563df 100644 --- a/staging/src/k8s.io/controller-manager/go.sum +++ b/staging/src/k8s.io/controller-manager/go.sum @@ -1,5 +1,5 @@ -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1/go.mod h1:xomTg63KZ2rFqZQzSB4Vz2SUXa1BpHTVz9L5PTmPC4E= github.com/NYTimes/gziphandler v1.1.1 h1:ZUDjpQae29j0ryrS0u/B8HZfJBtBQHjqw2rQ2cqUQ3I= github.com/NYTimes/gziphandler v1.1.1/go.mod h1:n/CVRwUEOgIxrgPvAQhUUr9oeUtvrhMomdKFjzJNB0c= @@ -20,7 +20,7 @@ github.com/cenkalti/backoff/v4 v4.3.0/go.mod h1:Y3VNntkOUPxTVeUxJ/G5vcM//AlwfmyY github.com/census-instrumentation/opencensus-proto v0.4.1/go.mod h1:4T9NM4+4Vw91VeyqjLS6ao50K5bOcLKN6Q42XnYaRYw= github.com/cespare/xxhash/v2 v2.3.0 h1:UL815xU9SqsFlibzuggzjXhog7bL6oX9BbNZnL2UFvs= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/coreos/go-oidc v2.2.1+incompatible/go.mod h1:CgnwVTmzoESiwO9qyAFEMiHoZ1nMCKZlZ9V6mm3/LKc= github.com/coreos/go-semver v0.3.1 h1:yi21YpKnrx1gt5R+la8n5WgS0kCrsPp33dmEyHReZr4= github.com/coreos/go-semver v0.3.1/go.mod h1:irMmmIw/7yzSRPWryHsK7EYSg09caPQL03VsM8rvUec= @@ -36,8 +36,8 @@ github.com/dustin/go-humanize v1.0.1 h1:GzkhY7T5VNhEkwH0PVJgjz+fX1rhBrR7pRT3mDkp github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto= github.com/emicklei/go-restful/v3 v3.11.0 h1:rAQeMHw1c7zTmncogyy8VvRZwtkmkZ4FxERmMY4rD+g= github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRry+4bhvbpWn3mrbc= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/felixge/httpsnoop v1.0.4 h1:NFTV2Zj1bL4mc9sqWACXbQFVBBg2W3GPvqp8/ESS2Wg= github.com/felixge/httpsnoop v1.0.4/go.mod h1:m8KPJKqk1gH5J9DgRY2ASl2lWCfGKXixSwevea8zH2U= github.com/fsnotify/fsnotify v1.7.0 h1:8JEhPFa5W2WU7YfeZzPNqzMP6Lwt7L2715Ggo0nosvA= @@ -67,7 +67,7 @@ github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= @@ -84,8 +84,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/websocket v1.5.0 h1:PPwGk2jz7EePpoHN/+ClbZu8SPxiqlu12wZP/3sWmnc= @@ -137,11 +137,12 @@ github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+ github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -246,8 +247,8 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -270,8 +271,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -280,12 +281,12 @@ golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNq google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d h1:VBu5YqKPv6XiJ199exd8Br+Aetz+o08F+PLMnwJQHAY= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/cri-api/go.mod b/staging/src/k8s.io/cri-api/go.mod index 945fc5d5221f4..53628b4343644 100644 --- a/staging/src/k8s.io/cri-api/go.mod +++ b/staging/src/k8s.io/cri-api/go.mod @@ -7,7 +7,7 @@ go 1.22.0 require ( github.com/gogo/protobuf v1.3.2 github.com/stretchr/testify v1.9.0 - google.golang.org/grpc v1.65.0 + google.golang.org/grpc v1.67.0 ) require ( @@ -18,7 +18,7 @@ require ( golang.org/x/net v0.29.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/text v0.18.0 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c // indirect gopkg.in/yaml.v3 v3.0.1 // indirect diff --git a/staging/src/k8s.io/cri-api/go.sum b/staging/src/k8s.io/cri-api/go.sum index 46714cf88ecb3..2fe0e2234a8c5 100644 --- a/staging/src/k8s.io/cri-api/go.sum +++ b/staging/src/k8s.io/cri-api/go.sum @@ -1,17 +1,17 @@ -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/census-instrumentation/opencensus-proto v0.4.1/go.mod h1:4T9NM4+4Vw91VeyqjLS6ao50K5bOcLKN6Q42XnYaRYw= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/creack/pty v1.1.9/go.mod h1:oKZEueFk5CKHvIhNR5MUki03XCEU+Q6VDXinZuGJ33E= github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc h1:U9qPSI2PIWSS1VwoXQT9A3Wy9MM3WgvqSxFWenqJduM= github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= -github.com/golang/protobuf v1.5.4/go.mod h1:lnTiLA8Wa4RWRcIUkrtSVa5nRhsEGBg48fD6rSs7xps= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/protobuf v1.5.0/go.mod h1:FsONVRAS9T7sI+LIUmWTfcYkHO4aIWwzhcaSAoJOfIk= github.com/google/go-cmp v0.6.0 h1:ofyhxvXcZhMsU5ulbFiLKl/XBFqE1GSq7atu8tAmTRI= github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeNGIjoY= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= @@ -25,6 +25,7 @@ github.com/kr/text v0.1.0/go.mod h1:4Jbv+DJW3UT/LiOwJeYQe1efqtUx/iVham/4vfdArNI= github.com/kr/text v0.2.0 h1:5Nx0Ya0ZqY2ygV366QzturHI13Jq95ApcVaJBhpS+AY= github.com/kr/text v0.2.0/go.mod h1:eLer722TekiGuMkidMxC/pM04lWEeraHUUmBw8l2grE= github.com/pkg/diff v0.0.0-20210226163009-20ebb0f2a09e/go.mod h1:pJLUxLENpZxwdsKMEsNbx1VGcRFpLqf3715MtcvvzbA= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/rogpeppe/go-internal v1.9.0/go.mod h1:WtVeX8xhTBvf0smdhujwtBcq4Qrzq/fJaraNFVN+nFs= @@ -48,7 +49,7 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.20.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -72,11 +73,11 @@ golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8T golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240814211410-ddb44dafa142/go.mod h1:d6be+8HhtEtucleCbxpPW9PA9XwISACu8nvpPqF0BVo= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/cri-client/go.mod b/staging/src/k8s.io/cri-client/go.mod index 51bf0d73f225f..aa89a26d7eb02 100644 --- a/staging/src/k8s.io/cri-client/go.mod +++ b/staging/src/k8s.io/cri-client/go.mod @@ -12,7 +12,7 @@ require ( go.opentelemetry.io/otel/sdk v1.28.0 go.opentelemetry.io/otel/trace v1.28.0 golang.org/x/sys v0.25.0 - google.golang.org/grpc v1.65.0 + google.golang.org/grpc v1.67.0 k8s.io/api v0.31.1 k8s.io/apimachinery v0.31.1 k8s.io/client-go v0.31.1 @@ -61,15 +61,15 @@ require ( go.opentelemetry.io/otel/exporters/otlp/otlptrace/otlptracegrpc v1.27.0 // indirect go.opentelemetry.io/otel/metric v1.28.0 // indirect go.opentelemetry.io/proto/otlp v1.3.1 // indirect - golang.org/x/mod v0.17.0 // indirect + golang.org/x/mod v0.20.0 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect + golang.org/x/tools v0.24.0 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/inf.v0 v0.9.1 // indirect gopkg.in/yaml.v2 v2.4.0 // indirect @@ -80,7 +80,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/client-go => ../client-go diff --git a/staging/src/k8s.io/cri-client/go.sum b/staging/src/k8s.io/cri-client/go.sum index 445b11f420693..6c565622342b1 100644 --- a/staging/src/k8s.io/cri-client/go.sum +++ b/staging/src/k8s.io/cri-client/go.sum @@ -1,5 +1,5 @@ -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1/go.mod h1:xomTg63KZ2rFqZQzSB4Vz2SUXa1BpHTVz9L5PTmPC4E= github.com/Microsoft/go-winio v0.6.0 h1:slsWYD/zyx7lCXoZVlvQrj0hPTM1HI4+v1sIda2yDvg= github.com/Microsoft/go-winio v0.6.0/go.mod h1:cTAf44im0RAYeL23bpB+fzCyDH2MJiz2BO69KH/soAE= @@ -18,7 +18,7 @@ github.com/cenkalti/backoff/v4 v4.3.0/go.mod h1:Y3VNntkOUPxTVeUxJ/G5vcM//AlwfmyY github.com/census-instrumentation/opencensus-proto v0.4.1/go.mod h1:4T9NM4+4Vw91VeyqjLS6ao50K5bOcLKN6Q42XnYaRYw= github.com/cespare/xxhash/v2 v2.3.0 h1:UL815xU9SqsFlibzuggzjXhog7bL6oX9BbNZnL2UFvs= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/creack/pty v1.1.9/go.mod h1:oKZEueFk5CKHvIhNR5MUki03XCEU+Q6VDXinZuGJ33E= github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= @@ -26,8 +26,8 @@ github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc h1:U9qPSI2PIWSS1 github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= github.com/emicklei/go-restful/v3 v3.11.0 h1:rAQeMHw1c7zTmncogyy8VvRZwtkmkZ4FxERmMY4rD+g= github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRry+4bhvbpWn3mrbc= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/felixge/httpsnoop v1.0.4 h1:NFTV2Zj1bL4mc9sqWACXbQFVBBg2W3GPvqp8/ESS2Wg= github.com/felixge/httpsnoop v1.0.4/go.mod h1:m8KPJKqk1gH5J9DgRY2ASl2lWCfGKXixSwevea8zH2U= github.com/fsnotify/fsnotify v1.7.0 h1:8JEhPFa5W2WU7YfeZzPNqzMP6Lwt7L2715Ggo0nosvA= @@ -52,7 +52,7 @@ github.com/go-openapi/swag v0.22.4/go.mod h1:UzaqsxGiab7freDnrUUra0MwWfN/q7tE4j+ github.com/go-task/slim-sprig/v3 v3.0.0/go.mod h1:W848ghGpv3Qj3dhTPRyJypKRiqCdHZiAzKg9hl15HA8= github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= github.com/golang/protobuf v1.5.4/go.mod h1:lnTiLA8Wa4RWRcIUkrtSVa5nRhsEGBg48fD6rSs7xps= @@ -65,7 +65,7 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/websocket v1.5.0/go.mod h1:YR8l580nyteQvAITg2hZ9XVh4b55+EU/adAjf1fMHhE= @@ -103,9 +103,10 @@ github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8m github.com/mwitkow/go-conntrack v0.0.0-20190716064945-2f068394615f/go.mod h1:qRWi+5nqEBWmkhHvq77mSJWrCKwh8bxhgT7d/eI7P4U= github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.19.0/go.mod h1:LY+I3pBVzYsTBU1AnDwOSxaYi9WoWiqgwooUqq9yPro= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -168,16 +169,16 @@ golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9/go.mod h1:LzIPMQfyMNhhGPh golang.org/x/crypto v0.27.0/go.mod h1:1Xngt8kV6Dvbssa53Ziq6Eqn0HqbZi5Z6R0ZpwQzt70= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0 h1:zY54UmvipHiNd+pm+m0x9KhZ9hl1/7QNMyxXbc6ICqA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0 h1:utOm6MM3R3dnawAiJgn0y+xvuYRsm1RKM/4giyfDgV0= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -188,7 +189,7 @@ golang.org/x/sys v0.0.0-20190412213103-97732733099d/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20200930185726-fdedc70b468f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.24.0/go.mod h1:lOBK/LVxemqiMij05LGJ0tzNr8xlmwBRJ81PX6wVLH8= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= golang.org/x/text v0.3.3/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ= @@ -200,20 +201,20 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNqU5K3HdILfMLy1f3aNYFI/wnl100a8= google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/csi-translation-lib/go.mod b/staging/src/k8s.io/csi-translation-lib/go.mod index 7dbb2bb5e4a73..ff5e84a077c87 100644 --- a/staging/src/k8s.io/csi-translation-lib/go.mod +++ b/staging/src/k8s.io/csi-translation-lib/go.mod @@ -35,7 +35,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/csi-translation-lib => ../csi-translation-lib diff --git a/staging/src/k8s.io/csi-translation-lib/go.sum b/staging/src/k8s.io/csi-translation-lib/go.sum index ca7d5f0c52364..789ebe3db7597 100644 --- a/staging/src/k8s.io/csi-translation-lib/go.sum +++ b/staging/src/k8s.io/csi-translation-lib/go.sum @@ -22,7 +22,7 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/josharian/intern v1.0.0/go.mod h1:5DoeVV0s6jJacbCEi61lwdGj/aVlrQvzHFFd8Hwg//Y= @@ -42,7 +42,7 @@ github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd/go.mod h1:6dJ github.com/modern-go/reflect2 v1.0.2 h1:xBagoLtFs94CBntxluKeaWgTMpvLxC4ur3nMaC9Gz0M= github.com/modern-go/reflect2 v1.0.2/go.mod h1:yWuevngMOJpCy52FWWMvUC8ws7m/LJsjYzDa0/r8luk= github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= @@ -91,7 +91,7 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/staging/src/k8s.io/dynamic-resource-allocation/go.mod b/staging/src/k8s.io/dynamic-resource-allocation/go.mod index c91443e7b729f..09a91ac799789 100644 --- a/staging/src/k8s.io/dynamic-resource-allocation/go.mod +++ b/staging/src/k8s.io/dynamic-resource-allocation/go.mod @@ -11,7 +11,7 @@ require ( github.com/google/go-cmp v0.6.0 github.com/onsi/gomega v1.33.1 github.com/stretchr/testify v1.9.0 - google.golang.org/grpc v1.65.0 + google.golang.org/grpc v1.67.0 k8s.io/api v0.31.1 k8s.io/apimachinery v0.31.1 k8s.io/apiserver v0.31.1 @@ -58,14 +58,14 @@ require ( github.com/x448/float16 v0.8.4 // indirect golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/evanphx/json-patch.v4 v4.12.0 // indirect gopkg.in/inf.v0 v0.9.1 // indirect @@ -79,7 +79,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/apiserver => ../apiserver diff --git a/staging/src/k8s.io/dynamic-resource-allocation/go.sum b/staging/src/k8s.io/dynamic-resource-allocation/go.sum index f3b757cfd05d3..f8894884de46f 100644 --- a/staging/src/k8s.io/dynamic-resource-allocation/go.sum +++ b/staging/src/k8s.io/dynamic-resource-allocation/go.sum @@ -1,5 +1,5 @@ -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1/go.mod h1:xomTg63KZ2rFqZQzSB4Vz2SUXa1BpHTVz9L5PTmPC4E= github.com/NYTimes/gziphandler v1.1.1/go.mod h1:n/CVRwUEOgIxrgPvAQhUUr9oeUtvrhMomdKFjzJNB0c= github.com/alecthomas/kingpin/v2 v2.4.0/go.mod h1:0gyi0zQnjuFk8xrkNKamJoyUo382HRL7ATRpFZCw6tE= @@ -17,7 +17,7 @@ github.com/cenkalti/backoff/v4 v4.3.0/go.mod h1:Y3VNntkOUPxTVeUxJ/G5vcM//AlwfmyY github.com/census-instrumentation/opencensus-proto v0.4.1/go.mod h1:4T9NM4+4Vw91VeyqjLS6ao50K5bOcLKN6Q42XnYaRYw= github.com/cespare/xxhash/v2 v2.3.0 h1:UL815xU9SqsFlibzuggzjXhog7bL6oX9BbNZnL2UFvs= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/coreos/go-oidc v2.2.1+incompatible/go.mod h1:CgnwVTmzoESiwO9qyAFEMiHoZ1nMCKZlZ9V6mm3/LKc= github.com/coreos/go-semver v0.3.1/go.mod h1:irMmmIw/7yzSRPWryHsK7EYSg09caPQL03VsM8rvUec= github.com/coreos/go-systemd/v22 v22.5.0/go.mod h1:Y58oyj3AT4RCenI/lSvhwexgC+NSVTIJ3seZv2GcEnc= @@ -30,8 +30,8 @@ github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc/go.mod h1:J7Y8Yc github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto= github.com/emicklei/go-restful/v3 v3.11.0 h1:rAQeMHw1c7zTmncogyy8VvRZwtkmkZ4FxERmMY4rD+g= github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRry+4bhvbpWn3mrbc= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/felixge/httpsnoop v1.0.4/go.mod h1:m8KPJKqk1gH5J9DgRY2ASl2lWCfGKXixSwevea8zH2U= github.com/fsnotify/fsnotify v1.7.0/go.mod h1:40Bi/Hjc2AVfZrqy+aj+yEI+/bRxZnMJyTJwOpGvigM= github.com/fxamacker/cbor/v2 v2.7.0 h1:iM5WgngdRBanHcxugY4JySA0nk1wZorNOpTgCMedv5E= @@ -54,7 +54,7 @@ github.com/go-task/slim-sprig/v3 v3.0.0/go.mod h1:W848ghGpv3Qj3dhTPRyJypKRiqCdHZ github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= @@ -70,8 +70,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/websocket v1.5.0/go.mod h1:YR8l580nyteQvAITg2hZ9XVh4b55+EU/adAjf1fMHhE= @@ -115,11 +115,12 @@ github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+ github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -198,8 +199,8 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -222,8 +223,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -231,12 +232,12 @@ golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8T golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNqU5K3HdILfMLy1f3aNYFI/wnl100a8= google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/endpointslice/go.mod b/staging/src/k8s.io/endpointslice/go.mod index ca7726cfaeb6c..acd07bcf12617 100644 --- a/staging/src/k8s.io/endpointslice/go.mod +++ b/staging/src/k8s.io/endpointslice/go.mod @@ -48,7 +48,7 @@ require ( github.com/spf13/pflag v1.0.5 // indirect github.com/x448/float16 v0.8.4 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect @@ -65,7 +65,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/client-go => ../client-go diff --git a/staging/src/k8s.io/endpointslice/go.sum b/staging/src/k8s.io/endpointslice/go.sum index 5c24ae619811b..b466f6b78a612 100644 --- a/staging/src/k8s.io/endpointslice/go.sum +++ b/staging/src/k8s.io/endpointslice/go.sum @@ -52,8 +52,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/websocket v1.5.0/go.mod h1:YR8l580nyteQvAITg2hZ9XVh4b55+EU/adAjf1fMHhE= @@ -91,8 +91,8 @@ github.com/mwitkow/go-conntrack v0.0.0-20190716064945-2f068394615f/go.mod h1:qRW github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= @@ -152,8 +152,8 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -175,17 +175,17 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNqU5K3HdILfMLy1f3aNYFI/wnl100a8= google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/kms/go.mod b/staging/src/k8s.io/kms/go.mod index 9e491fe066f47..71a93629621ca 100644 --- a/staging/src/k8s.io/kms/go.mod +++ b/staging/src/k8s.io/kms/go.mod @@ -6,14 +6,14 @@ go 1.22.0 require ( github.com/gogo/protobuf v1.3.2 - google.golang.org/grpc v1.65.0 + google.golang.org/grpc v1.67.0 ) require ( golang.org/x/net v0.29.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/text v0.18.0 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect google.golang.org/protobuf v1.34.2 // indirect ) diff --git a/staging/src/k8s.io/kms/go.sum b/staging/src/k8s.io/kms/go.sum index 41cb09b8108a3..d8ecb005595aa 100644 --- a/staging/src/k8s.io/kms/go.sum +++ b/staging/src/k8s.io/kms/go.sum @@ -1,19 +1,20 @@ -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/census-instrumentation/opencensus-proto v0.4.1/go.mod h1:4T9NM4+4Vw91VeyqjLS6ao50K5bOcLKN6Q42XnYaRYw= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= -github.com/golang/protobuf v1.5.4/go.mod h1:lnTiLA8Wa4RWRcIUkrtSVa5nRhsEGBg48fD6rSs7xps= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/protobuf v1.5.0/go.mod h1:FsONVRAS9T7sI+LIUmWTfcYkHO4aIWwzhcaSAoJOfIk= github.com/google/go-cmp v0.6.0 h1:ofyhxvXcZhMsU5ulbFiLKl/XBFqE1GSq7atu8tAmTRI= github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeNGIjoY= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/kisielk/errcheck v1.5.0/go.mod h1:pFxgyoBC7bSaBwPgfKdkLd5X25qrDl4LWUI2bnpBCr8= github.com/kisielk/gotool v1.0.0/go.mod h1:XhKaO+MFFWcvkIS/tQcRk01m1F5IRFswLeQ+oQHNcck= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/yuin/goldmark v1.1.27/go.mod h1:3hX8gzYuyVAZsxl0MRgGTJEmQBFcNTphYh9decYSb74= github.com/yuin/goldmark v1.2.1/go.mod h1:3hX8gzYuyVAZsxl0MRgGTJEmQBFcNTphYh9decYSb74= golang.org/x/crypto v0.0.0-20190308221718-c2843e01d9a2/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w= @@ -29,7 +30,7 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.20.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -53,10 +54,10 @@ golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8T golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240814211410-ddb44dafa142/go.mod h1:d6be+8HhtEtucleCbxpPW9PA9XwISACu8nvpPqF0BVo= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= diff --git a/staging/src/k8s.io/kms/internal/plugins/_mock/go.mod b/staging/src/k8s.io/kms/internal/plugins/_mock/go.mod index 690494697dbf9..9b7fa972c90b1 100644 --- a/staging/src/k8s.io/kms/internal/plugins/_mock/go.mod +++ b/staging/src/k8s.io/kms/internal/plugins/_mock/go.mod @@ -15,8 +15,8 @@ require ( golang.org/x/net v0.29.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/text v0.18.0 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect - google.golang.org/grpc v1.65.0 // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect + google.golang.org/grpc v1.67.0 // indirect google.golang.org/protobuf v1.34.2 // indirect ) @@ -24,4 +24,4 @@ replace k8s.io/kms => ../../../../kms replace github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 -replace github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 +replace github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 diff --git a/staging/src/k8s.io/kms/internal/plugins/_mock/go.sum b/staging/src/k8s.io/kms/internal/plugins/_mock/go.sum index 25bd685d4a6b3..1dcb854add8a5 100644 --- a/staging/src/k8s.io/kms/internal/plugins/_mock/go.sum +++ b/staging/src/k8s.io/kms/internal/plugins/_mock/go.sum @@ -53,9 +53,9 @@ golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8T golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= diff --git a/staging/src/k8s.io/kube-aggregator/go.mod b/staging/src/k8s.io/kube-aggregator/go.mod index 85a299445b5a5..dc80c18e1a018 100644 --- a/staging/src/k8s.io/kube-aggregator/go.mod +++ b/staging/src/k8s.io/kube-aggregator/go.mod @@ -88,17 +88,17 @@ require ( go.uber.org/zap v1.26.0 // indirect golang.org/x/crypto v0.27.0 // indirect golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc // indirect - golang.org/x/mod v0.17.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/mod v0.20.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect - google.golang.org/grpc v1.65.0 // indirect + golang.org/x/tools v0.24.0 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect + google.golang.org/grpc v1.67.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/evanphx/json-patch.v4 v4.12.0 // indirect gopkg.in/inf.v0 v0.9.1 // indirect @@ -114,7 +114,7 @@ require ( replace ( github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apiextensions-apiserver => ../apiextensions-apiserver k8s.io/apimachinery => ../apimachinery diff --git a/staging/src/k8s.io/kube-aggregator/go.sum b/staging/src/k8s.io/kube-aggregator/go.sum index 122143a771b88..eb55290a99702 100644 --- a/staging/src/k8s.io/kube-aggregator/go.sum +++ b/staging/src/k8s.io/kube-aggregator/go.sum @@ -1,6 +1,6 @@ bitbucket.org/bertimus9/systemstat v0.5.0/go.mod h1:EkUWPp8lKFPMXP8vnbpT5JDI0W/sTiLZAvN8ONWErHY= -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1/go.mod h1:xomTg63KZ2rFqZQzSB4Vz2SUXa1BpHTVz9L5PTmPC4E= github.com/Azure/go-ntlmssp v0.0.0-20211209120228-48547f28849e/go.mod h1:chxPXzSsl7ZWRAuOIE23GDNzjWuZquvFlgA8xmpunjU= github.com/JeffAshton/win_pdh v0.0.0-20161109143554-76bb4ee9f0ab/go.mod h1:3VYc5hodBMJ5+l/7J4xAyMeuM2PNuepvHlGs8yilUCA= @@ -32,7 +32,7 @@ github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XL github.com/chai2010/gettext-go v1.0.2/go.mod h1:y+wnP2cHYaVj19NZhYKAwEMH2CI1gNHeQQ+5AjwawxA= github.com/checkpoint-restore/go-criu/v5 v5.3.0/go.mod h1:E/eQpaFtUKGOOSEBZgmKAcn+zUUwWxqcaKZlF54wK8E= github.com/cilium/ebpf v0.9.1/go.mod h1:+OhNOIXx/Fnu1IE8bJz2dzOA+VSfyTfdNUVdlQnxUFY= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/container-storage-interface/spec v1.9.0/go.mod h1:ZfDu+3ZRyeVqxZM0Ds19MVLkN2d1XJ5MAfi1L3VjlT0= github.com/containerd/cgroups v1.1.0/go.mod h1:6ppBcbh/NOOUU+dMKrykgaBnK9lCIBxHqJDGwsa1mIw= github.com/containerd/console v1.0.3/go.mod h1:7LqA/THxQ86k76b8c/EMSiaJ3h1eZkMkXar0TQ1gf3U= @@ -61,8 +61,8 @@ github.com/dustin/go-humanize v1.0.1 h1:GzkhY7T5VNhEkwH0PVJgjz+fX1rhBrR7pRT3mDkp github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto= github.com/emicklei/go-restful/v3 v3.11.0 h1:rAQeMHw1c7zTmncogyy8VvRZwtkmkZ4FxERmMY4rD+g= github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRry+4bhvbpWn3mrbc= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/euank/go-kmsg-parser v2.0.0+incompatible/go.mod h1:MhmAMZ8V4CYH4ybgdRwPr2TU5ThnS43puaKEMpja1uw= github.com/evanphx/json-patch v4.12.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= github.com/exponent-io/jsonpath v0.0.0-20151013193312-d6023ce2651d/go.mod h1:ZZMPRZwes7CROmyNKgQzC3XPs6L/G2EJLHddWejkmf4= @@ -102,7 +102,7 @@ github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= @@ -125,8 +125,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= @@ -199,6 +199,7 @@ github.com/opencontainers/image-spec v1.0.2/go.mod h1:BtxoFyWECRxE4U/7sNtV5W15zM github.com/opencontainers/runc v1.1.13/go.mod h1:R016aXacfp/gwQBYw2FDGa9m+n6atbLWrYY8hNMT/sA= github.com/opencontainers/runtime-spec v1.0.3-0.20220909204839-494a5a6aca78/go.mod h1:jwyrGlmzljRJv/Fgzds9SsS/C5hL+LL3ko9hs6T5lQ0= github.com/opencontainers/selinux v1.11.0/go.mod h1:E5dMC3VPuVvVHDYmi78qvhJp8+M586T4DlDRYpFkyec= +github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf/go.mod h1:1OhaNsaU9vuy/dlYZLEve7bgE2Ed+yTV5VSbYvGXt4s= github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f/go.mod h1:Shkl4HanLwDiiBzakv+con/aMGnVE2MAGvoKp5oyYUo= github.com/openshift/apiserver-library-go v0.0.0-20241001175710-6064b62894a6/go.mod h1:9Anrq7+DZmmw1Brchx4zmh26hAZbe6Dv7bGXRclnhYI= github.com/openshift/build-machinery-go v0.0.0-20240613134303-8359781da660/go.mod h1:8jcm8UPtg2mCAsxfqKil1xrmRMI3a+XU2TZ9fF8A7TE= @@ -206,12 +207,13 @@ github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f/go.mod h1:KiZi github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2/go.mod h1:s6Fqwb2KiWG6leCegVhw4KW40tf9f7m+SF1aXiE8Wsk= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060 h1:jiDC7d8d+jmjv2WfiMY0+Uf55q11MGyYkGGqXnfqWTU= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/profile v1.7.0/go.mod h1:8Uer0jas47ZQMJ7VD+OHknK4YDY07LPUC6dEvqDjvNo= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -321,16 +323,16 @@ golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc h1:mCRnTeVUjcrhlRmO0VK8a6k6R golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc/go.mod h1:V1LtkGg67GoY2N1AnLN78QLrzxkLyJw7RJb1gzOOz9w= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0 h1:zY54UmvipHiNd+pm+m0x9KhZ9hl1/7QNMyxXbc6ICqA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0 h1:utOm6MM3R3dnawAiJgn0y+xvuYRsm1RKM/4giyfDgV0= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -341,7 +343,7 @@ golang.org/x/sys v0.0.0-20190412213103-97732733099d/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20200930185726-fdedc70b468f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.24.0 h1:Mh5cbb+Zk2hqqXNO7S1iTjEphVL+jb8ZWaqh/g+JWkM= golang.org/x/term v0.24.0/go.mod h1:lOBK/LVxemqiMij05LGJ0tzNr8xlmwBRJ81PX6wVLH8= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= @@ -354,8 +356,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -364,12 +366,12 @@ golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNq google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d h1:VBu5YqKPv6XiJ199exd8Br+Aetz+o08F+PLMnwJQHAY= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/kube-controller-manager/go.mod b/staging/src/k8s.io/kube-controller-manager/go.mod index 83ccc330606e4..d676cf6934d83 100644 --- a/staging/src/k8s.io/kube-controller-manager/go.mod +++ b/staging/src/k8s.io/kube-controller-manager/go.mod @@ -33,7 +33,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/apiserver => ../apiserver diff --git a/staging/src/k8s.io/kube-controller-manager/go.sum b/staging/src/k8s.io/kube-controller-manager/go.sum index 97a9c34890507..9bcab2268eeca 100644 --- a/staging/src/k8s.io/kube-controller-manager/go.sum +++ b/staging/src/k8s.io/kube-controller-manager/go.sum @@ -39,7 +39,7 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0/go.mod h1:8NvIoxWQoOIhqOTXgfV/d3M/q6VIi02HzZEHgUlZvzk= github.com/grpc-ecosystem/grpc-gateway/v2 v2.20.0/go.mod h1:P+Lt/0by1T8bfcF3z737NnSbmxQAppXMRziHUxPOC8k= @@ -69,7 +69,7 @@ github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8m github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= @@ -121,7 +121,7 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -140,14 +140,14 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c h1:Hei/4ADfdWqJk1ZMxUNpqntNwaWcugrBjAiHlqqRiVk= diff --git a/staging/src/k8s.io/kube-proxy/go.mod b/staging/src/k8s.io/kube-proxy/go.mod index e0a0feccc3ade..9ab406426ca7f 100644 --- a/staging/src/k8s.io/kube-proxy/go.mod +++ b/staging/src/k8s.io/kube-proxy/go.mod @@ -45,7 +45,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/client-go => ../client-go diff --git a/staging/src/k8s.io/kube-proxy/go.sum b/staging/src/k8s.io/kube-proxy/go.sum index 84fc6ce5ee29b..cf4284a626a0a 100644 --- a/staging/src/k8s.io/kube-proxy/go.sum +++ b/staging/src/k8s.io/kube-proxy/go.sum @@ -39,7 +39,7 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/grpc-ecosystem/grpc-gateway/v2 v2.20.0/go.mod h1:P+Lt/0by1T8bfcF3z737NnSbmxQAppXMRziHUxPOC8k= github.com/inconshreveable/mousetrap v1.1.0 h1:wN+x4NVGpMsO7ErUn/mUI3vEoE6Jt13X2s0bqwp9tc8= @@ -67,7 +67,7 @@ github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822 h1:C3w9PqII01/Oq github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8mK8HuQTcFwEeznm/DIxMOiR9yIdICNftLE1DvQ= github.com/mwitkow/go-conntrack v0.0.0-20190716064945-2f068394615f/go.mod h1:qRWi+5nqEBWmkhHvq77mSJWrCKwh8bxhgT7d/eI7P4U= github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= @@ -119,7 +119,7 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -139,15 +139,15 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/kube-scheduler/go.mod b/staging/src/k8s.io/kube-scheduler/go.mod index 4d3e5d091c9ae..3ea833ab13090 100644 --- a/staging/src/k8s.io/kube-scheduler/go.mod +++ b/staging/src/k8s.io/kube-scheduler/go.mod @@ -33,7 +33,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/client-go => ../client-go diff --git a/staging/src/k8s.io/kube-scheduler/go.sum b/staging/src/k8s.io/kube-scheduler/go.sum index 399a8eeb65c7d..16c62619b75f8 100644 --- a/staging/src/k8s.io/kube-scheduler/go.sum +++ b/staging/src/k8s.io/kube-scheduler/go.sum @@ -31,7 +31,7 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/grpc-ecosystem/grpc-gateway/v2 v2.20.0/go.mod h1:P+Lt/0by1T8bfcF3z737NnSbmxQAppXMRziHUxPOC8k= github.com/inconshreveable/mousetrap v1.1.0/go.mod h1:vpF70FUmC8bwa3OWnCshd2FqLfsEA9PFc4w1p2J65bw= @@ -54,7 +54,7 @@ github.com/modern-go/reflect2 v1.0.2 h1:xBagoLtFs94CBntxluKeaWgTMpvLxC4ur3nMaC9G github.com/modern-go/reflect2 v1.0.2/go.mod h1:yWuevngMOJpCy52FWWMvUC8ws7m/LJsjYzDa0/r8luk= github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8mK8HuQTcFwEeznm/DIxMOiR9yIdICNftLE1DvQ= github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= @@ -99,7 +99,7 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -118,14 +118,14 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c h1:Hei/4ADfdWqJk1ZMxUNpqntNwaWcugrBjAiHlqqRiVk= diff --git a/staging/src/k8s.io/kubectl/go.mod b/staging/src/k8s.io/kubectl/go.mod index c787e16ea5693..0591a610de25b 100644 --- a/staging/src/k8s.io/kubectl/go.mod +++ b/staging/src/k8s.io/kubectl/go.mod @@ -19,7 +19,7 @@ require ( github.com/lithammer/dedent v1.1.0 github.com/mitchellh/go-wordwrap v1.0.1 github.com/moby/term v0.5.0 - github.com/onsi/ginkgo/v2 v2.17.2 + github.com/onsi/ginkgo/v2 v2.20.2 github.com/onsi/gomega v1.33.1 github.com/pkg/errors v0.9.1 github.com/russross/blackfriday/v2 v2.1.0 @@ -61,7 +61,7 @@ require ( github.com/golang/protobuf v1.5.4 // indirect github.com/google/btree v1.0.1 // indirect github.com/google/gofuzz v1.2.0 // indirect - github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 // indirect + github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 // indirect github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510 // indirect github.com/google/uuid v1.6.0 // indirect github.com/gorilla/websocket v1.5.0 // indirect @@ -84,12 +84,12 @@ require ( github.com/xlab/treeprint v1.2.0 // indirect go.starlark.net v0.0.0-20230525235612-a134d8f9ddca // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect + golang.org/x/tools v0.24.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/inf.v0 v0.9.1 // indirect gopkg.in/yaml.v3 v3.0.1 // indirect @@ -97,7 +97,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/cli-runtime => ../cli-runtime diff --git a/staging/src/k8s.io/kubectl/go.sum b/staging/src/k8s.io/kubectl/go.sum index de9f655d697d7..3dbb203f20f0e 100644 --- a/staging/src/k8s.io/kubectl/go.sum +++ b/staging/src/k8s.io/kubectl/go.sum @@ -96,8 +96,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510 h1:El6M4kTTCOh6aBiKaUGG7oYTSPP8MxqL4YI3kZKwcP4= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= @@ -154,8 +154,8 @@ github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= github.com/opencontainers/go-digest v1.0.0 h1:apOUWs51W5PlhuyGyz9FCeeBIOUDA/6nW8Oi/yOhh5U= github.com/opencontainers/go-digest v1.0.0/go.mod h1:0JzlMkj0TRzQZfJkVvzbP0HBR3IKzErnv2BNG4W4MAM= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible h1:UBdAOUP5p4RWqPBg048CAvpKN+vxiaj6gdUUzhl4XmI= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= @@ -221,7 +221,7 @@ golang.org/x/lint v0.0.0-20190227174305-5b3e6a55c961/go.mod h1:wehouNa3lNwaWXcvx golang.org/x/lint v0.0.0-20190313153728-d0100b6bd8b3/go.mod h1:6SW0HCj/g11FgYtHlgUYUwCkIfeOF89ocIRzGO/8vkc= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190213061140-3a22650c66bd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= @@ -233,8 +233,8 @@ golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwY golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181108010431-42b317875d0f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -251,7 +251,7 @@ golang.org/x/sys v0.0.0-20210616094352-59db8d763f22/go.mod h1:oPkhp1MJrh7nUepCBc golang.org/x/sys v0.0.0-20220715151400-c0bba94af5f8/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.0.0-20220526004731-065cf7ba2467/go.mod h1:jbD1KX2456YbFQfuXm/mYQcufACuNUgVhRMnK/tPxf8= golang.org/x/term v0.24.0 h1:Mh5cbb+Zk2hqqXNO7S1iTjEphVL+jb8ZWaqh/g+JWkM= golang.org/x/term v0.24.0/go.mod h1:lOBK/LVxemqiMij05LGJ0tzNr8xlmwBRJ81PX6wVLH8= @@ -269,8 +269,8 @@ golang.org/x/tools v0.0.0-20190524140312-2c0ae7006135/go.mod h1:RgjU9mgBXZiqYHBn golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -281,12 +281,12 @@ google.golang.org/appengine v1.4.0/go.mod h1:xpcJRLb0r/rnEns0DIKYYv+WjYCduHsrkT7 google.golang.org/genproto v0.0.0-20180817151627-c66870c02cf8/go.mod h1:JiN7NxoALGmiZfu7CAH4rXhgtRTLTxftemlI0sWmxmc= google.golang.org/genproto v0.0.0-20190819201941-24fa4b261c55/go.mod h1:DMBHOl98Agz4BDEuKkezgsaosCRResVns1a3J2ZsMNc= google.golang.org/genproto v0.0.0-20200526211855-cb27e3aa2013/go.mod h1:NbSheEEYHJ7i3ixzK3sjbqSGDJWnxyFXZblF3eUsNvo= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= google.golang.org/grpc v1.19.0/go.mod h1:mqu4LbDTu4XGKhr4mRzUsmM4RtVoemTSY81AxZiDr8c= google.golang.org/grpc v1.23.0/go.mod h1:Y5yQAOtifL1yxbo5wqy6BxZv8vAUGQwXBOALyacEbxg= google.golang.org/grpc v1.27.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v0.0.0-20200109180630-ec00e32a8dfd/go.mod h1:DFci5gLYBciE7Vtevhsrf46CRTquxDuWsQurQQe4oz8= google.golang.org/protobuf v0.0.0-20200221191635-4d8936d0db64/go.mod h1:kwYJMbMJ01Woi6D6+Kah6886xMZcty6N08ah7+eCXa0= google.golang.org/protobuf v0.0.0-20200228230310-ab0ca4ff8a60/go.mod h1:cfTl7dwQJ+fmap5saPgwCLgHXTUD7jkjRqWcaiX5VyM= diff --git a/staging/src/k8s.io/kubelet/go.mod b/staging/src/k8s.io/kubelet/go.mod index 9b394509cf08c..a7f9ecc381f91 100644 --- a/staging/src/k8s.io/kubelet/go.mod +++ b/staging/src/k8s.io/kubelet/go.mod @@ -8,7 +8,7 @@ require ( github.com/emicklei/go-restful/v3 v3.11.0 github.com/gogo/protobuf v1.3.2 github.com/stretchr/testify v1.9.0 - google.golang.org/grpc v1.65.0 + google.golang.org/grpc v1.67.0 k8s.io/api v0.31.1 k8s.io/apimachinery v0.31.1 k8s.io/apiserver v0.31.1 @@ -46,12 +46,12 @@ require ( github.com/spf13/pflag v1.0.5 // indirect github.com/x448/float16 v0.8.4 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/inf.v0 v0.9.1 // indirect gopkg.in/yaml.v2 v2.4.0 // indirect @@ -62,7 +62,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/apiserver => ../apiserver diff --git a/staging/src/k8s.io/kubelet/go.sum b/staging/src/k8s.io/kubelet/go.sum index 6555e7067a3f8..8aa58dd051467 100644 --- a/staging/src/k8s.io/kubelet/go.sum +++ b/staging/src/k8s.io/kubelet/go.sum @@ -1,5 +1,5 @@ -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1/go.mod h1:xomTg63KZ2rFqZQzSB4Vz2SUXa1BpHTVz9L5PTmPC4E= github.com/NYTimes/gziphandler v1.1.1/go.mod h1:n/CVRwUEOgIxrgPvAQhUUr9oeUtvrhMomdKFjzJNB0c= github.com/alecthomas/kingpin/v2 v2.4.0/go.mod h1:0gyi0zQnjuFk8xrkNKamJoyUo382HRL7ATRpFZCw6tE= @@ -16,7 +16,7 @@ github.com/cenkalti/backoff/v4 v4.3.0/go.mod h1:Y3VNntkOUPxTVeUxJ/G5vcM//AlwfmyY github.com/census-instrumentation/opencensus-proto v0.4.1/go.mod h1:4T9NM4+4Vw91VeyqjLS6ao50K5bOcLKN6Q42XnYaRYw= github.com/cespare/xxhash/v2 v2.3.0 h1:UL815xU9SqsFlibzuggzjXhog7bL6oX9BbNZnL2UFvs= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/coreos/go-oidc v2.2.1+incompatible/go.mod h1:CgnwVTmzoESiwO9qyAFEMiHoZ1nMCKZlZ9V6mm3/LKc= github.com/coreos/go-semver v0.3.1/go.mod h1:irMmmIw/7yzSRPWryHsK7EYSg09caPQL03VsM8rvUec= github.com/coreos/go-systemd/v22 v22.5.0/go.mod h1:Y58oyj3AT4RCenI/lSvhwexgC+NSVTIJ3seZv2GcEnc= @@ -29,8 +29,8 @@ github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc/go.mod h1:J7Y8Yc github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto= github.com/emicklei/go-restful/v3 v3.11.0 h1:rAQeMHw1c7zTmncogyy8VvRZwtkmkZ4FxERmMY4rD+g= github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRry+4bhvbpWn3mrbc= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/felixge/httpsnoop v1.0.4/go.mod h1:m8KPJKqk1gH5J9DgRY2ASl2lWCfGKXixSwevea8zH2U= github.com/fsnotify/fsnotify v1.7.0/go.mod h1:40Bi/Hjc2AVfZrqy+aj+yEI+/bRxZnMJyTJwOpGvigM= github.com/fxamacker/cbor/v2 v2.7.0 h1:iM5WgngdRBanHcxugY4JySA0nk1wZorNOpTgCMedv5E= @@ -52,7 +52,7 @@ github.com/go-task/slim-sprig/v3 v3.0.0/go.mod h1:W848ghGpv3Qj3dhTPRyJypKRiqCdHZ github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= github.com/golang/protobuf v1.5.4/go.mod h1:lnTiLA8Wa4RWRcIUkrtSVa5nRhsEGBg48fD6rSs7xps= @@ -66,8 +66,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/websocket v1.5.0 h1:PPwGk2jz7EePpoHN/+ClbZu8SPxiqlu12wZP/3sWmnc= @@ -109,10 +109,11 @@ github.com/mwitkow/go-conntrack v0.0.0-20190716064945-2f068394615f/go.mod h1:qRW github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f h1:y5//uYreIhSUg3J1GEMiLbxo1LJaP8RfCpH6pymGZus= github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -182,8 +183,8 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -205,19 +206,19 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/metrics/go.mod b/staging/src/k8s.io/metrics/go.mod index 30b0e9da4d07a..57ab747548a22 100644 --- a/staging/src/k8s.io/metrics/go.mod +++ b/staging/src/k8s.io/metrics/go.mod @@ -36,15 +36,15 @@ require ( github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 // indirect github.com/spf13/pflag v1.0.5 // indirect github.com/x448/float16 v0.8.4 // indirect - golang.org/x/mod v0.17.0 // indirect + golang.org/x/mod v0.20.0 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect + golang.org/x/tools v0.24.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/evanphx/json-patch.v4 v4.12.0 // indirect gopkg.in/inf.v0 v0.9.1 // indirect @@ -60,7 +60,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/client-go => ../client-go diff --git a/staging/src/k8s.io/metrics/go.sum b/staging/src/k8s.io/metrics/go.sum index 2c707c3058066..d4bd059996c38 100644 --- a/staging/src/k8s.io/metrics/go.sum +++ b/staging/src/k8s.io/metrics/go.sum @@ -36,8 +36,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/websocket v1.5.0/go.mod h1:YR8l580nyteQvAITg2hZ9XVh4b55+EU/adAjf1fMHhE= @@ -69,8 +69,8 @@ github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8m github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= @@ -103,16 +103,16 @@ golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9/go.mod h1:LzIPMQfyMNhhGPh golang.org/x/crypto v0.27.0/go.mod h1:1Xngt8kV6Dvbssa53Ziq6Eqn0HqbZi5Z6R0ZpwQzt70= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0 h1:zY54UmvipHiNd+pm+m0x9KhZ9hl1/7QNMyxXbc6ICqA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0 h1:utOm6MM3R3dnawAiJgn0y+xvuYRsm1RKM/4giyfDgV0= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -123,7 +123,7 @@ golang.org/x/sys v0.0.0-20190412213103-97732733099d/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20200930185726-fdedc70b468f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.24.0 h1:Mh5cbb+Zk2hqqXNO7S1iTjEphVL+jb8ZWaqh/g+JWkM= golang.org/x/term v0.24.0/go.mod h1:lOBK/LVxemqiMij05LGJ0tzNr8xlmwBRJ81PX6wVLH8= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= @@ -136,8 +136,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/staging/src/k8s.io/pod-security-admission/go.mod b/staging/src/k8s.io/pod-security-admission/go.mod index 8d6cc1fa28179..bbbc9ebb6dc5f 100644 --- a/staging/src/k8s.io/pod-security-admission/go.mod +++ b/staging/src/k8s.io/pod-security-admission/go.mod @@ -85,15 +85,15 @@ require ( golang.org/x/crypto v0.27.0 // indirect golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect - google.golang.org/grpc v1.65.0 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect + google.golang.org/grpc v1.67.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/evanphx/json-patch.v4 v4.12.0 // indirect gopkg.in/inf.v0 v0.9.1 // indirect @@ -108,7 +108,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apiextensions-apiserver => ../apiextensions-apiserver k8s.io/apimachinery => ../apimachinery diff --git a/staging/src/k8s.io/pod-security-admission/go.sum b/staging/src/k8s.io/pod-security-admission/go.sum index 41b9f66bc800b..34857f98649af 100644 --- a/staging/src/k8s.io/pod-security-admission/go.sum +++ b/staging/src/k8s.io/pod-security-admission/go.sum @@ -1,5 +1,5 @@ -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1/go.mod h1:xomTg63KZ2rFqZQzSB4Vz2SUXa1BpHTVz9L5PTmPC4E= github.com/Azure/go-ntlmssp v0.0.0-20211209120228-48547f28849e/go.mod h1:chxPXzSsl7ZWRAuOIE23GDNzjWuZquvFlgA8xmpunjU= github.com/NYTimes/gziphandler v1.1.1 h1:ZUDjpQae29j0ryrS0u/B8HZfJBtBQHjqw2rQ2cqUQ3I= @@ -22,7 +22,7 @@ github.com/cenkalti/backoff/v4 v4.3.0/go.mod h1:Y3VNntkOUPxTVeUxJ/G5vcM//AlwfmyY github.com/census-instrumentation/opencensus-proto v0.4.1/go.mod h1:4T9NM4+4Vw91VeyqjLS6ao50K5bOcLKN6Q42XnYaRYw= github.com/cespare/xxhash/v2 v2.3.0 h1:UL815xU9SqsFlibzuggzjXhog7bL6oX9BbNZnL2UFvs= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/coreos/go-oidc v2.2.1+incompatible/go.mod h1:CgnwVTmzoESiwO9qyAFEMiHoZ1nMCKZlZ9V6mm3/LKc= github.com/coreos/go-semver v0.3.1 h1:yi21YpKnrx1gt5R+la8n5WgS0kCrsPp33dmEyHReZr4= github.com/coreos/go-semver v0.3.1/go.mod h1:irMmmIw/7yzSRPWryHsK7EYSg09caPQL03VsM8rvUec= @@ -41,8 +41,8 @@ github.com/dustin/go-humanize v1.0.1 h1:GzkhY7T5VNhEkwH0PVJgjz+fX1rhBrR7pRT3mDkp github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto= github.com/emicklei/go-restful/v3 v3.11.0 h1:rAQeMHw1c7zTmncogyy8VvRZwtkmkZ4FxERmMY4rD+g= github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRry+4bhvbpWn3mrbc= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/evanphx/json-patch v4.12.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= github.com/felixge/fgprof v0.9.4/go.mod h1:yKl+ERSa++RYOs32d8K6WEXCB4uXdLls4ZaZPpayhMM= github.com/felixge/httpsnoop v1.0.4 h1:NFTV2Zj1bL4mc9sqWACXbQFVBBg2W3GPvqp8/ESS2Wg= @@ -77,7 +77,7 @@ github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= @@ -100,8 +100,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/mux v1.8.1/go.mod h1:AKf9I4AEqPTmMytcMc0KkNouC66V3BtZ4qD5fmWSiMQ= @@ -163,12 +163,13 @@ github.com/openshift/build-machinery-go v0.0.0-20240613134303-8359781da660/go.mo github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f/go.mod h1:KiZi2mJRH1TOJ3FtBDYS6YvUL30s/iIXaGSUrSa36mo= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060 h1:jiDC7d8d+jmjv2WfiMY0+Uf55q11MGyYkGGqXnfqWTU= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/profile v1.7.0/go.mod h1:8Uer0jas47ZQMJ7VD+OHknK4YDY07LPUC6dEvqDjvNo= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -275,8 +276,8 @@ golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -299,8 +300,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -309,12 +310,12 @@ golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNq google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d h1:VBu5YqKPv6XiJ199exd8Br+Aetz+o08F+PLMnwJQHAY= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/sample-apiserver/go.mod b/staging/src/k8s.io/sample-apiserver/go.mod index 6bea345c4b9ab..d792c9ac4b3bb 100644 --- a/staging/src/k8s.io/sample-apiserver/go.mod +++ b/staging/src/k8s.io/sample-apiserver/go.mod @@ -82,18 +82,18 @@ require ( go.uber.org/zap v1.26.0 // indirect golang.org/x/crypto v0.27.0 // indirect golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc // indirect - golang.org/x/mod v0.17.0 // indirect + golang.org/x/mod v0.20.0 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect golang.org/x/time v0.3.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect - google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 // indirect - google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 // indirect - google.golang.org/grpc v1.65.0 // indirect + golang.org/x/tools v0.24.0 // indirect + google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f // indirect + google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 // indirect + google.golang.org/grpc v1.67.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/evanphx/json-patch.v4 v4.12.0 // indirect gopkg.in/inf.v0 v0.9.1 // indirect @@ -110,7 +110,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apiextensions-apiserver => ../apiextensions-apiserver k8s.io/apimachinery => ../apimachinery diff --git a/staging/src/k8s.io/sample-apiserver/go.sum b/staging/src/k8s.io/sample-apiserver/go.sum index dd568e90028e3..22be54db2ca5c 100644 --- a/staging/src/k8s.io/sample-apiserver/go.sum +++ b/staging/src/k8s.io/sample-apiserver/go.sum @@ -1,5 +1,5 @@ -cel.dev/expr v0.15.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= -cloud.google.com/go/compute/metadata v0.3.0/go.mod h1:zFmK7XCadkQkj6TtorcaGlCW1hT1fIilQDwofLpJ20k= +cel.dev/expr v0.16.0/go.mod h1:TRSuuV7DlVCE/uwv5QbAiW/v8l5O8C4eEPHeu7gf7Sg= +cloud.google.com/go/compute/metadata v0.5.0/go.mod h1:aHnloV2TPI38yx4s9+wAZhHykWvVCfu7hQbF+9CWoiY= github.com/Azure/go-ansiterm v0.0.0-20210617225240-d185dfc1b5a1/go.mod h1:xomTg63KZ2rFqZQzSB4Vz2SUXa1BpHTVz9L5PTmPC4E= github.com/Azure/go-ntlmssp v0.0.0-20211209120228-48547f28849e/go.mod h1:chxPXzSsl7ZWRAuOIE23GDNzjWuZquvFlgA8xmpunjU= github.com/NYTimes/gziphandler v1.1.1 h1:ZUDjpQae29j0ryrS0u/B8HZfJBtBQHjqw2rQ2cqUQ3I= @@ -22,7 +22,7 @@ github.com/cenkalti/backoff/v4 v4.3.0/go.mod h1:Y3VNntkOUPxTVeUxJ/G5vcM//AlwfmyY github.com/census-instrumentation/opencensus-proto v0.4.1/go.mod h1:4T9NM4+4Vw91VeyqjLS6ao50K5bOcLKN6Q42XnYaRYw= github.com/cespare/xxhash/v2 v2.3.0 h1:UL815xU9SqsFlibzuggzjXhog7bL6oX9BbNZnL2UFvs= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= -github.com/cncf/xds/go v0.0.0-20240423153145-555b57ec207b/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= +github.com/cncf/xds/go v0.0.0-20240723142845-024c85f92f20/go.mod h1:W+zGtBO5Y1IgJhy4+A9GOqVhqLpfZi+vwmdNXUehLA8= github.com/coreos/go-oidc v2.2.1+incompatible/go.mod h1:CgnwVTmzoESiwO9qyAFEMiHoZ1nMCKZlZ9V6mm3/LKc= github.com/coreos/go-semver v0.3.1 h1:yi21YpKnrx1gt5R+la8n5WgS0kCrsPp33dmEyHReZr4= github.com/coreos/go-semver v0.3.1/go.mod h1:irMmmIw/7yzSRPWryHsK7EYSg09caPQL03VsM8rvUec= @@ -41,8 +41,8 @@ github.com/dustin/go-humanize v1.0.1 h1:GzkhY7T5VNhEkwH0PVJgjz+fX1rhBrR7pRT3mDkp github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto= github.com/emicklei/go-restful/v3 v3.11.0 h1:rAQeMHw1c7zTmncogyy8VvRZwtkmkZ4FxERmMY4rD+g= github.com/emicklei/go-restful/v3 v3.11.0/go.mod h1:6n3XBCmQQb25CM2LCACGz8ukIrRry+4bhvbpWn3mrbc= -github.com/envoyproxy/go-control-plane v0.12.0/go.mod h1:ZBTaoJ23lqITozF0M6G4/IragXCQKCnYbmlmtHvwRG0= -github.com/envoyproxy/protoc-gen-validate v1.0.4/go.mod h1:qys6tmnRsYrQqIhm2bvKZH4Blx/1gTIZ2UKVY1M+Yew= +github.com/envoyproxy/go-control-plane v0.13.0/go.mod h1:GRaKG3dwvFoTg4nj7aXdZnvMg4d7nvT/wl9WgVXn3Q8= +github.com/envoyproxy/protoc-gen-validate v1.1.0/go.mod h1:sXRDRVmzEbkM7CVcM06s9shE/m23dg3wzjl0UWqJ2q4= github.com/evanphx/json-patch v4.12.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= github.com/felixge/fgprof v0.9.4/go.mod h1:yKl+ERSa++RYOs32d8K6WEXCB4uXdLls4ZaZPpayhMM= github.com/felixge/httpsnoop v1.0.4 h1:NFTV2Zj1bL4mc9sqWACXbQFVBBg2W3GPvqp8/ESS2Wg= @@ -77,7 +77,7 @@ github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg= github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= -github.com/golang/glog v1.2.1/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= +github.com/golang/glog v1.2.2/go.mod h1:6AhwSGph0fcJtXVM/PEHPqZlFeoLxhs7/t5UDAwmO+w= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE= github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= @@ -100,8 +100,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/mux v1.8.1/go.mod h1:AKf9I4AEqPTmMytcMc0KkNouC66V3BtZ4qD5fmWSiMQ= @@ -163,12 +163,13 @@ github.com/openshift/build-machinery-go v0.0.0-20240613134303-8359781da660/go.mo github.com/openshift/client-go v0.0.0-20241001162912-da6d55e4611f/go.mod h1:KiZi2mJRH1TOJ3FtBDYS6YvUL30s/iIXaGSUrSa36mo= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060 h1:jiDC7d8d+jmjv2WfiMY0+Uf55q11MGyYkGGqXnfqWTU= github.com/openshift/library-go v0.0.0-20241107160307-0064ad7bd060/go.mod h1:9B1MYPoLtP9tqjWxcbUNVpwxy68zOH/3EIP6c31dAM0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/profile v1.7.0/go.mod h1:8Uer0jas47ZQMJ7VD+OHknK4YDY07LPUC6dEvqDjvNo= +github.com/planetscale/vtprotobuf v0.6.1-0.20240319094008-0393e58bdf10/go.mod h1:t/avpk3KcrXxUnYOhZhMXJlSEyie6gQbtLq5NM3loB8= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= @@ -269,16 +270,16 @@ golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc h1:mCRnTeVUjcrhlRmO0VK8a6k6R golang.org/x/exp v0.0.0-20230515195305-f3d0a9c9a5cc/go.mod h1:V1LtkGg67GoY2N1AnLN78QLrzxkLyJw7RJb1gzOOz9w= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0 h1:zY54UmvipHiNd+pm+m0x9KhZ9hl1/7QNMyxXbc6ICqA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0 h1:utOm6MM3R3dnawAiJgn0y+xvuYRsm1RKM/4giyfDgV0= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -289,7 +290,7 @@ golang.org/x/sys v0.0.0-20190412213103-97732733099d/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20200930185726-fdedc70b468f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.24.0 h1:Mh5cbb+Zk2hqqXNO7S1iTjEphVL+jb8ZWaqh/g+JWkM= golang.org/x/term v0.24.0/go.mod h1:lOBK/LVxemqiMij05LGJ0tzNr8xlmwBRJ81PX6wVLH8= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= @@ -302,8 +303,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -312,12 +313,12 @@ golang.org/x/xerrors v0.0.0-20220907171357-04be3eba64a2/go.mod h1:K8+ghG5WaK9qNq google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d h1:VBu5YqKPv6XiJ199exd8Br+Aetz+o08F+PLMnwJQHAY= google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d/go.mod h1:yZTlhN0tQnXo3h00fuXNCxJdLdIdnVFVBaRJ5LWBbw4= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 h1:7whR9kGa5LUwFtpLm2ArCEejtnxlGeLbAyjFY8sGNFw= -google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157/go.mod h1:99sLkeliLXfdj2J75X3Ho+rrVCaJze0uwN7zDDkjPVU= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 h1:BwIjyKYGsK9dMCBOorzRri8MQwmi7mT9rGHsCEinZkA= -google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094/go.mod h1:Ue6ibwXGpU+dqIcODieyLOcgj7z8+IcskoNIgZxtrFY= -google.golang.org/grpc v1.65.0 h1:bs/cUb4lp1G5iImFFd3u5ixQzweKizoZJAwBNLR42lc= -google.golang.org/grpc v1.65.0/go.mod h1:WgYC2ypjlB0EiQi6wdKixMqukr6lBc0Vo+oOgjrM5ZQ= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f h1:jTm13A2itBi3La6yTGqn8bVSrc3ZZ1r8ENHlIXBfnRA= +google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f/go.mod h1:CLGoBuH1VHxAUXVPP8FfPwPEVJB6lz3URE5mY2SuayE= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 h1:N9BgCIAUvn/M+p4NJccWPWb3BWh88+zyL0ll9HgbEeM= +google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61/go.mod h1:UqMtugtsSgubUsoxbuAoiCXvqvErP7Gf0so0mK9tHxU= +google.golang.org/grpc v1.67.0 h1:IdH9y6PF5MPSdAntIcpjQ+tXO41pcQsfZV2RxtQgVcw= +google.golang.org/grpc v1.67.0/go.mod h1:1gLDyUQU7CTLJI90u3nXZ9ekeghjeM7pTDZlqFNg2AA= google.golang.org/protobuf v1.34.2 h1:6xV6lTsCfpGD21XK49h7MhtcApnLqkfYgPcdHftf6hg= google.golang.org/protobuf v1.34.2/go.mod h1:qYOHts0dSfpeUzUFpOMr/WGzszTmLH+DiWniOlNbLDw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= diff --git a/staging/src/k8s.io/sample-cli-plugin/go.mod b/staging/src/k8s.io/sample-cli-plugin/go.mod index 237efd7a4a37c..6e114ca51fd0c 100644 --- a/staging/src/k8s.io/sample-cli-plugin/go.mod +++ b/staging/src/k8s.io/sample-cli-plugin/go.mod @@ -49,7 +49,7 @@ require ( github.com/xlab/treeprint v1.2.0 // indirect go.starlark.net v0.0.0-20230525235612-a134d8f9ddca // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect @@ -73,7 +73,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/cli-runtime => ../cli-runtime diff --git a/staging/src/k8s.io/sample-cli-plugin/go.sum b/staging/src/k8s.io/sample-cli-plugin/go.sum index a7002f7088d0b..fadf48a0b0f1a 100644 --- a/staging/src/k8s.io/sample-cli-plugin/go.sum +++ b/staging/src/k8s.io/sample-cli-plugin/go.sum @@ -71,8 +71,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510 h1:El6M4kTTCOh6aBiKaUGG7oYTSPP8MxqL4YI3kZKwcP4= github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510/go.mod h1:pupxD2MaaD3pAXIBCelhxNneeOaAeabZDe5s4K6zSpQ= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= @@ -116,8 +116,8 @@ github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8m github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible h1:UBdAOUP5p4RWqPBg048CAvpKN+vxiaj6gdUUzhl4XmI= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= @@ -179,8 +179,8 @@ golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwY golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181108010431-42b317875d0f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -214,8 +214,8 @@ golang.org/x/tools v0.0.0-20190524140312-2c0ae7006135/go.mod h1:RgjU9mgBXZiqYHBn golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/staging/src/k8s.io/sample-controller/go.mod b/staging/src/k8s.io/sample-controller/go.mod index c5bdef493d809..60338272a5731 100644 --- a/staging/src/k8s.io/sample-controller/go.mod +++ b/staging/src/k8s.io/sample-controller/go.mod @@ -38,14 +38,14 @@ require ( github.com/pkg/errors v0.9.1 // indirect github.com/spf13/pflag v1.0.5 // indirect github.com/x448/float16 v0.8.4 // indirect - golang.org/x/mod v0.17.0 // indirect + golang.org/x/mod v0.20.0 // indirect golang.org/x/net v0.29.0 // indirect - golang.org/x/oauth2 v0.21.0 // indirect + golang.org/x/oauth2 v0.22.0 // indirect golang.org/x/sync v0.8.0 // indirect golang.org/x/sys v0.25.0 // indirect golang.org/x/term v0.24.0 // indirect golang.org/x/text v0.18.0 // indirect - golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d // indirect + golang.org/x/tools v0.24.0 // indirect google.golang.org/protobuf v1.34.2 // indirect gopkg.in/evanphx/json-patch.v4 v4.12.0 // indirect gopkg.in/inf.v0 v0.9.1 // indirect @@ -60,7 +60,7 @@ require ( ) replace ( - github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 + github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 k8s.io/api => ../api k8s.io/apimachinery => ../apimachinery k8s.io/client-go => ../client-go diff --git a/staging/src/k8s.io/sample-controller/go.sum b/staging/src/k8s.io/sample-controller/go.sum index f1e537124477d..c6e3d06632dcd 100644 --- a/staging/src/k8s.io/sample-controller/go.sum +++ b/staging/src/k8s.io/sample-controller/go.sum @@ -37,8 +37,8 @@ github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeN github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.2.0 h1:xRy4A+RhZaiKjJ1bPfwQ8sedCA+YS2YcCHW6ec7JMi0= github.com/google/gofuzz v1.2.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 h1:FKHo8hFI3A+7w0aUQuYXQ+6EN5stWmeY/AZqtM8xk9k= -github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8/go.mod h1:K1liHPHnj73Fdn/EKuT8nrFqBihUSKXoLYU0BuatOYo= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 h1:5iH8iuqE5apketRbSFBy+X1V0o+l+8NF1avt4HWl7cA= +github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5/go.mod h1:vavhavw2zAxS5dIdcRluK6cSGGPlZynqzFM8NdvU144= github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/gorilla/websocket v1.5.0/go.mod h1:YR8l580nyteQvAITg2hZ9XVh4b55+EU/adAjf1fMHhE= @@ -71,8 +71,8 @@ github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8m github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/onsi/gomega v1.33.1 h1:dsYjIxxSR755MDmKVsaFQTE22ChNBcuuTWgkUDSubOk= github.com/onsi/gomega v1.33.1/go.mod h1:U4R44UsT+9eLIaYRB2a5qajjtQYn0hauxvRm16AVYg0= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 h1:HJvLw9nNfoqCs16h505eP8E1kVmq6KNdzGm5csPlYsQ= -github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 h1:dKUAGq29JBk2oRn955gnnz3sG/UhlkUKh1ISleLiews= +github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44/go.mod h1:rlwLi9PilAFJ8jCg9UE1QP6VBpd6/xj3SRC0d6TU0To= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= @@ -105,16 +105,16 @@ golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9/go.mod h1:LzIPMQfyMNhhGPh golang.org/x/crypto v0.27.0/go.mod h1:1Xngt8kV6Dvbssa53Ziq6Eqn0HqbZi5Z6R0ZpwQzt70= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.17.0 h1:zY54UmvipHiNd+pm+m0x9KhZ9hl1/7QNMyxXbc6ICqA= -golang.org/x/mod v0.17.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= +golang.org/x/mod v0.20.0 h1:utOm6MM3R3dnawAiJgn0y+xvuYRsm1RKM/4giyfDgV0= +golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.29.0 h1:5ORfpBpCs4HzDYoodCDBbwHzdR5UrLBZ3sOnUJmFoHo= golang.org/x/net v0.29.0/go.mod h1:gLkgy8jTGERgjzMic6DS9+SP0ajcu6Xu3Orq/SpETg0= -golang.org/x/oauth2 v0.21.0 h1:tsimM75w1tF/uws5rbeHzIWxEqElMehnc+iW793zsZs= -golang.org/x/oauth2 v0.21.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= +golang.org/x/oauth2 v0.22.0 h1:BzDx2FehcG7jJwgWLELCdmLuxk2i+x9UDpSiss2u0ZA= +golang.org/x/oauth2 v0.22.0/go.mod h1:XYTD2NtWslqkgxebSiOHnXEap4TF09sJSc7H1sXbhtI= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -125,7 +125,7 @@ golang.org/x/sys v0.0.0-20190412213103-97732733099d/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20200930185726-fdedc70b468f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.25.0 h1:r+8e+loiHxRqhXVl6ML1nO3l1+oFoWbnlu2Ehimmi34= golang.org/x/sys v0.25.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= -golang.org/x/telemetry v0.0.0-20240228155512-f48c80bd79b2/go.mod h1:TeRTkGYfJXctD9OcfyVLyj2J3IxLnKwHJR8f4D8a3YE= +golang.org/x/telemetry v0.0.0-20240521205824-bda55230c457/go.mod h1:pRgIJT+bRLFKnoM1ldnzKoxTIn14Yxz928LQRYYgIN0= golang.org/x/term v0.24.0 h1:Mh5cbb+Zk2hqqXNO7S1iTjEphVL+jb8ZWaqh/g+JWkM= golang.org/x/term v0.24.0/go.mod h1:lOBK/LVxemqiMij05LGJ0tzNr8xlmwBRJ81PX6wVLH8= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= @@ -138,8 +138,8 @@ golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d h1:vU5i/LfpvrRCpgM/VPfJLg5KjxD3E+hfT1SH+d9zLwg= -golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d/go.mod h1:aiJjzUbINMkxbQROHiO6hDPo2LHcIPhhQsa9DLh0yGk= +golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24= +golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= diff --git a/vendor/github.com/onsi/ginkgo/v2/OWNERS b/vendor/github.com/onsi/ginkgo/v2/OWNERS index a7df319007e89..5c9e0b087fbdf 100644 --- a/vendor/github.com/onsi/ginkgo/v2/OWNERS +++ b/vendor/github.com/onsi/ginkgo/v2/OWNERS @@ -1,3 +1,6 @@ +reviewers: approvers: + - bertinatto - bparees - soltysh + - stbenjam diff --git a/vendor/github.com/onsi/ginkgo/v2/core_dsl_patch.go b/vendor/github.com/onsi/ginkgo/v2/core_dsl_patch.go index 0260e4a53a74b..bf60ceb522dd2 100644 --- a/vendor/github.com/onsi/ginkgo/v2/core_dsl_patch.go +++ b/vendor/github.com/onsi/ginkgo/v2/core_dsl_patch.go @@ -1,6 +1,8 @@ package ginkgo import ( + "io" + "github.com/onsi/ginkgo/v2/internal" "github.com/onsi/ginkgo/v2/internal/global" "github.com/onsi/ginkgo/v2/types" @@ -18,6 +20,10 @@ func GetFailer() *internal.Failer { return global.Failer } +func NewWriter(w io.Writer) *internal.Writer { + return internal.NewWriter(w) +} + func GetWriter() *internal.Writer { return GinkgoWriter.(*internal.Writer) } diff --git a/vendor/github.com/onsi/ginkgo/v2/internal/spec_patch.go b/vendor/github.com/onsi/ginkgo/v2/internal/spec_patch.go index f486f3849e926..2d0bcc914dfff 100644 --- a/vendor/github.com/onsi/ginkgo/v2/internal/spec_patch.go +++ b/vendor/github.com/onsi/ginkgo/v2/internal/spec_patch.go @@ -11,3 +11,12 @@ func (s Spec) CodeLocations() []types.CodeLocation { func (s Spec) AppendText(text string) { s.Nodes[len(s.Nodes)-1].Text += text } + +func (s Spec) Labels() []string { + var labels []string + for _, n := range s.Nodes { + labels = append(labels, n.Labels...) + } + + return labels +} diff --git a/vendor/github.com/onsi/ginkgo/v2/types/types_patch.go b/vendor/github.com/onsi/ginkgo/v2/types/types_patch.go index c17b3af223781..02d319bba041f 100644 --- a/vendor/github.com/onsi/ginkgo/v2/types/types_patch.go +++ b/vendor/github.com/onsi/ginkgo/v2/types/types_patch.go @@ -4,4 +4,5 @@ type TestSpec interface { CodeLocations() []CodeLocation Text() string AppendText(text string) + Labels() []string } diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/LICENSE b/vendor/github.com/openshift-eng/openshift-tests-extension/LICENSE new file mode 100644 index 0000000000000..261eeb9e9f8b2 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/LICENSE @@ -0,0 +1,201 @@ + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "[]" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright [yyyy] [name of copyright owner] + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmd.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmd.go new file mode 100644 index 0000000000000..5db1741be74ba --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmd.go @@ -0,0 +1,21 @@ +package cmd + +import ( + "github.com/spf13/cobra" + + "github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdinfo" + "github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdlist" + "github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdrun" + "github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdupdate" + "github.com/openshift-eng/openshift-tests-extension/pkg/extension" +) + +func DefaultExtensionCommands(registry *extension.Registry) []*cobra.Command { + return []*cobra.Command{ + cmdrun.NewRunSuiteCommand(registry), + cmdrun.NewRunTestCommand(registry), + cmdlist.NewListCommand(registry), + cmdinfo.NewInfoCommand(registry), + cmdupdate.NewUpdateCommand(registry), + } +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdinfo/info.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdinfo/info.go new file mode 100644 index 0000000000000..1d4237876d539 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdinfo/info.go @@ -0,0 +1,38 @@ +package cmdinfo + +import ( + "encoding/json" + "fmt" + "os" + + "github.com/spf13/cobra" + + "github.com/openshift-eng/openshift-tests-extension/pkg/extension" + "github.com/openshift-eng/openshift-tests-extension/pkg/flags" +) + +func NewInfoCommand(registry *extension.Registry) *cobra.Command { + componentFlags := flags.NewComponentFlags() + + cmd := &cobra.Command{ + Use: "info", + Short: "Display extension metadata", + SilenceUsage: true, + RunE: func(cmd *cobra.Command, args []string) error { + extension := registry.Get(componentFlags.Component) + if extension == nil { + return fmt.Errorf("couldn't find the component %q", componentFlags.Component) + } + + info, err := json.MarshalIndent(extension, "", " ") + if err != nil { + return err + } + + fmt.Fprintf(os.Stdout, "%s\n", string(info)) + return nil + }, + } + componentFlags.BindFlags(cmd.Flags()) + return cmd +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdlist/list.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdlist/list.go new file mode 100644 index 0000000000000..69f02660a0588 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdlist/list.go @@ -0,0 +1,124 @@ +package cmdlist + +import ( + "fmt" + "os" + + "github.com/spf13/cobra" + + "github.com/openshift-eng/openshift-tests-extension/pkg/extension" + "github.com/openshift-eng/openshift-tests-extension/pkg/flags" +) + +func NewListCommand(registry *extension.Registry) *cobra.Command { + opts := struct { + componentFlags *flags.ComponentFlags + suiteFlags *flags.SuiteFlags + outputFlags *flags.OutputFlags + }{ + suiteFlags: flags.NewSuiteFlags(), + componentFlags: flags.NewComponentFlags(), + outputFlags: flags.NewOutputFlags(), + } + + // Tests + listTestsCmd := &cobra.Command{ + Use: "tests", + Short: "List available tests", + SilenceUsage: true, + RunE: func(cmd *cobra.Command, args []string) error { + ext := registry.Get(opts.componentFlags.Component) + if ext == nil { + return fmt.Errorf("component not found: %s", opts.componentFlags.Component) + } + + // Find suite, if specified + var foundSuite *extension.Suite + var err error + if opts.suiteFlags.Suite != "" { + foundSuite, err = ext.GetSuite(opts.suiteFlags.Suite) + if err != nil { + return err + } + } + + // Filter for suite + specs := ext.GetSpecs() + if foundSuite != nil { + specs, err = specs.Filter(foundSuite.Qualifiers) + if err != nil { + return err + } + } + + data, err := opts.outputFlags.Marshal(specs) + if err != nil { + return err + } + fmt.Fprintf(os.Stdout, "%s\n", string(data)) + return nil + }, + } + opts.suiteFlags.BindFlags(listTestsCmd.Flags()) + opts.componentFlags.BindFlags(listTestsCmd.Flags()) + opts.outputFlags.BindFlags(listTestsCmd.Flags()) + + // Suites + listSuitesCommand := &cobra.Command{ + Use: "suites", + Short: "List available suites", + SilenceUsage: true, + RunE: func(cmd *cobra.Command, args []string) error { + ext := registry.Get(opts.componentFlags.Component) + if ext == nil { + return fmt.Errorf("component not found: %s", opts.componentFlags.Component) + } + + suites := ext.Suites + + data, err := opts.outputFlags.Marshal(suites) + if err != nil { + return err + } + fmt.Fprintf(os.Stdout, "%s\n", string(data)) + return nil + }, + } + opts.componentFlags.BindFlags(listSuitesCommand.Flags()) + opts.outputFlags.BindFlags(listSuitesCommand.Flags()) + + // Components + listComponentsCmd := &cobra.Command{ + Use: "components", + Short: "List available components", + SilenceUsage: true, + RunE: func(cmd *cobra.Command, args []string) error { + var components []*extension.Component + registry.Walk(func(e *extension.Extension) { + components = append(components, &e.Component) + }) + + data, err := opts.outputFlags.Marshal(components) + if err != nil { + return err + } + fmt.Fprintf(os.Stdout, "%s\n", string(data)) + return nil + }, + } + opts.outputFlags.BindFlags(listComponentsCmd.Flags()) + + var listCmd = &cobra.Command{ + Use: "list [subcommand]", + Short: "List items", + RunE: func(cmd *cobra.Command, args []string) error { + return listTestsCmd.RunE(cmd, args) + }, + } + opts.suiteFlags.BindFlags(listCmd.Flags()) + opts.componentFlags.BindFlags(listCmd.Flags()) + opts.outputFlags.BindFlags(listCmd.Flags()) + listCmd.AddCommand(listTestsCmd, listComponentsCmd, listSuitesCommand) + + return listCmd +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdrun/runsuite.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdrun/runsuite.go new file mode 100644 index 0000000000000..c79b3ff786b40 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdrun/runsuite.go @@ -0,0 +1,63 @@ +package cmdrun + +import ( + "fmt" + "os" + + "github.com/pkg/errors" + "github.com/spf13/cobra" + + "github.com/openshift-eng/openshift-tests-extension/pkg/extension" + "github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests" + "github.com/openshift-eng/openshift-tests-extension/pkg/flags" +) + +func NewRunSuiteCommand(registry *extension.Registry) *cobra.Command { + opts := struct { + componentFlags *flags.ComponentFlags + outputFlags *flags.OutputFlags + concurrencyFlags *flags.ConcurrencyFlags + }{ + componentFlags: flags.NewComponentFlags(), + outputFlags: flags.NewOutputFlags(), + concurrencyFlags: flags.NewConcurrencyFlags(), + } + + cmd := &cobra.Command{ + Use: "run-suite NAME", + Short: "Run a group of tests by suite", + SilenceUsage: true, + RunE: func(cmd *cobra.Command, args []string) error { + ext := registry.Get(opts.componentFlags.Component) + if ext == nil { + return fmt.Errorf("component not found: %s", opts.componentFlags.Component) + } + if len(args) != 1 { + return fmt.Errorf("must specify one suite name") + } + + w, err := extensiontests.NewResultWriter(os.Stdout, extensiontests.ResultFormat(opts.outputFlags.Output)) + if err != nil { + return err + } + defer w.Flush() + + suite, err := ext.GetSuite(args[0]) + if err != nil { + return errors.Wrapf(err, "couldn't find suite: %s", args[0]) + } + + specs, err := ext.GetSpecs().Filter(suite.Qualifiers) + if err != nil { + return errors.Wrap(err, "couldn't filter specs") + } + + return specs.Run(w, opts.concurrencyFlags.MaxConcurency) + }, + } + opts.componentFlags.BindFlags(cmd.Flags()) + opts.outputFlags.BindFlags(cmd.Flags()) + opts.concurrencyFlags.BindFlags(cmd.Flags()) + + return cmd +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdrun/runtest.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdrun/runtest.go new file mode 100644 index 0000000000000..ea4b62cb6bfba --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdrun/runtest.go @@ -0,0 +1,81 @@ +package cmdrun + +import ( + "bufio" + "fmt" + "os" + + "github.com/spf13/cobra" + + "github.com/openshift-eng/openshift-tests-extension/pkg/extension" + "github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests" + "github.com/openshift-eng/openshift-tests-extension/pkg/flags" +) + +func NewRunTestCommand(registry *extension.Registry) *cobra.Command { + opts := struct { + componentFlags *flags.ComponentFlags + concurrencyFlags *flags.ConcurrencyFlags + nameFlags *flags.NamesFlags + outputFlags *flags.OutputFlags + }{ + componentFlags: flags.NewComponentFlags(), + nameFlags: flags.NewNamesFlags(), + outputFlags: flags.NewOutputFlags(), + concurrencyFlags: flags.NewConcurrencyFlags(), + } + + cmd := &cobra.Command{ + Use: "run-test [-n NAME...] [NAME]", + Short: "Runs tests by name", + SilenceUsage: true, + RunE: func(cmd *cobra.Command, args []string) error { + ext := registry.Get(opts.componentFlags.Component) + if ext == nil { + return fmt.Errorf("component not found: %s", opts.componentFlags.Component) + } + if len(args) > 1 { + return fmt.Errorf("use --names to specify more than one test") + } + opts.nameFlags.Names = append(opts.nameFlags.Names, args...) + + // allow reading tests from an stdin pipe + info, err := os.Stdin.Stat() + if err != nil { + return err + } + if info.Mode()&os.ModeCharDevice == 0 { // Check if input is from a pipe + scanner := bufio.NewScanner(os.Stdin) + for scanner.Scan() { + opts.nameFlags.Names = append(opts.nameFlags.Names, scanner.Text()) + } + if err := scanner.Err(); err != nil { + return fmt.Errorf("error reading from stdin: %v", err) + } + } + + if len(opts.nameFlags.Names) == 0 { + return fmt.Errorf("must specify at least one test") + } + + specs, err := ext.FindSpecsByName(opts.nameFlags.Names...) + if err != nil { + return err + } + + w, err := extensiontests.NewResultWriter(os.Stdout, extensiontests.ResultFormat(opts.outputFlags.Output)) + if err != nil { + return err + } + defer w.Flush() + + return specs.Run(w, opts.concurrencyFlags.MaxConcurency) + }, + } + opts.componentFlags.BindFlags(cmd.Flags()) + opts.nameFlags.BindFlags(cmd.Flags()) + opts.outputFlags.BindFlags(cmd.Flags()) + opts.concurrencyFlags.BindFlags(cmd.Flags()) + + return cmd +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdupdate/update.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdupdate/update.go new file mode 100644 index 0000000000000..5d847308e59e9 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdupdate/update.go @@ -0,0 +1,84 @@ +package cmdupdate + +import ( + "encoding/json" + "fmt" + "os" + "path/filepath" + "strings" + + "github.com/spf13/cobra" + + "github.com/openshift-eng/openshift-tests-extension/pkg/extension" + "github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests" + "github.com/openshift-eng/openshift-tests-extension/pkg/flags" +) + +const metadataDirectory = ".openshift-tests-extension" + +// NewUpdateCommand adds an "update" command used to generate and verify the metadata we keep track of. This should +// be a black box to end users, i.e. we can add more criteria later they'll consume when revendoring. For now, +// we prevent a test to be renamed without updating other names, or a test to be deleted. +func NewUpdateCommand(registry *extension.Registry) *cobra.Command { + componentFlags := flags.NewComponentFlags() + + cmd := &cobra.Command{ + Use: "update", + Short: "Update test metadata", + SilenceUsage: true, + RunE: func(cmd *cobra.Command, args []string) error { + ext := registry.Get(componentFlags.Component) + if ext == nil { + return fmt.Errorf("couldn't find the component %q", componentFlags.Component) + } + + // Create the metadata directory if it doesn't exist + if err := os.MkdirAll(metadataDirectory, 0755); err != nil { + return fmt.Errorf("failed to create directory %s: %w", metadataDirectory, err) + } + + // Read existing specs + metadataPath := filepath.Join(metadataDirectory, fmt.Sprintf("%s.json", strings.ReplaceAll(ext.Component.Identifier(), ":", "_"))) + var oldSpecs extensiontests.ExtensionTestSpecs + source, err := os.Open(metadataPath) + if err != nil { + if !os.IsNotExist(err) { + return fmt.Errorf("failed to open file: %s: %+w", metadataPath, err) + } + } else { + if err := json.NewDecoder(source).Decode(&oldSpecs); err != nil { + return fmt.Errorf("failed to decode file: %s: %+w", metadataPath, err) + } + + missing, err := ext.FindRemovedTestsWithoutRename(oldSpecs) + if err != nil && len(missing) > 0 { + fmt.Fprintf(os.Stderr, "Missing Tests:\n") + for _, name := range missing { + fmt.Fprintf(os.Stdout, " * %s\n", name) + } + fmt.Fprintf(os.Stderr, "\n") + + return fmt.Errorf("missing tests, if you've renamed tests you must add their names to OriginalName, " + + "or mark them obsolete") + } + } + + // no missing tests, write the results + newSpecs := ext.GetSpecs() + data, err := json.MarshalIndent(newSpecs, "", " ") + if err != nil { + return fmt.Errorf("failed to marshal specs to JSON: %w", err) + } + + // Write the JSON data to the file + if err := os.WriteFile(metadataPath, data, 0644); err != nil { + return fmt.Errorf("failed to write file %s: %w", metadataPath, err) + } + + fmt.Printf("successfully updated metadata\n") + return nil + }, + } + componentFlags.BindFlags(cmd.Flags()) + return cmd +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/dbtime/time.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/dbtime/time.go new file mode 100644 index 0000000000000..b7651ba0220dd --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/dbtime/time.go @@ -0,0 +1,26 @@ +package dbtime + +import "time" + +// DBTime is a type suitable for direct importing into databases like BigQuery, +// formatted like 2006-01-02 15:04:05.000000 UTC. +type DBTime time.Time + +func Ptr(t time.Time) *DBTime { + return (*DBTime)(&t) +} + +func (dbt *DBTime) MarshalJSON() ([]byte, error) { + formattedTime := time.Time(*dbt).Format(`"2006-01-02 15:04:05.000000 UTC"`) + return []byte(formattedTime), nil +} + +func (dbt *DBTime) UnmarshalJSON(b []byte) error { + timeStr := string(b[1 : len(b)-1]) + parsedTime, err := time.Parse("2006-01-02 15:04:05.000000 UTC", timeStr) + if err != nil { + return err + } + *dbt = (DBTime)(parsedTime) + return nil +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extension.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extension.go new file mode 100644 index 0000000000000..747871ec1c272 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extension.go @@ -0,0 +1,155 @@ +package extension + +import ( + "fmt" + "strings" + + "k8s.io/apimachinery/pkg/util/sets" + + et "github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests" + "github.com/openshift-eng/openshift-tests-extension/pkg/version" +) + +func NewExtension(product, kind, name string) *Extension { + return &Extension{ + APIVersion: CurrentExtensionAPIVersion, + Source: Source{ + Commit: version.CommitFromGit, + BuildDate: version.BuildDate, + GitTreeState: version.GitTreeState, + }, + Component: Component{ + Product: product, + Kind: kind, + Name: name, + }, + } +} + +func (e *Extension) GetSuite(name string) (*Suite, error) { + var suite *Suite + + for _, s := range e.Suites { + if s.Name == name { + suite = &s + break + } + } + + if suite == nil { + return nil, fmt.Errorf("no such suite: %s", name) + } + + return suite, nil +} + +func (e *Extension) GetSpecs() et.ExtensionTestSpecs { + return e.specs +} + +func (e *Extension) AddSpecs(specs et.ExtensionTestSpecs) { + specs.Walk(func(spec *et.ExtensionTestSpec) { + spec.Source = e.Component.Identifier() + }) + + e.specs = append(e.specs, specs...) +} + +// IgnoreObsoleteTests allows removal of a test. +func (e *Extension) IgnoreObsoleteTests(testNames ...string) { + if e.obsoleteTests == nil { + e.obsoleteTests = sets.New[string](testNames...) + } else { + e.obsoleteTests.Insert(testNames...) + } +} + +// FindRemovedTestsWithoutRename compares the current set of test specs against oldSpecs, including consideration of the original name, +// we return an error. Can be used to detect test renames or removals. +func (e *Extension) FindRemovedTestsWithoutRename(oldSpecs et.ExtensionTestSpecs) ([]string, error) { + currentSpecs := e.GetSpecs() + currentMap := make(map[string]bool) + + // Populate current specs into a map for quick lookup by both Name and OriginalName. + for _, spec := range currentSpecs { + currentMap[spec.Name] = true + if spec.OriginalName != "" { + currentMap[spec.OriginalName] = true + } + } + + var removedTests []string + + // Check oldSpecs against current specs. + for _, oldSpec := range oldSpecs { + // Skip if the test is marked as obsolete. + if e.obsoleteTests.Has(oldSpec.Name) { + continue + } + + // Check if oldSpec is missing in currentSpecs by both Name and OriginalName. + if !currentMap[oldSpec.Name] && (oldSpec.OriginalName == "" || !currentMap[oldSpec.OriginalName]) { + removedTests = append(removedTests, oldSpec.Name) + } + } + + // Return error if any removed tests were found. + if len(removedTests) > 0 { + return removedTests, fmt.Errorf("tests removed without rename: %v", removedTests) + } + + return nil, nil +} + +// AddGlobalSuite adds a suite whose qualifiers will apply to all tests, +// not just this one. Allowing a developer to create a composed suite of +// tests from many sources. +func (e *Extension) AddGlobalSuite(suite Suite) *Extension { + if e.Suites == nil { + e.Suites = []Suite{suite} + } else { + e.Suites = append(e.Suites, suite) + } + + return e +} + +// AddSuite adds a suite whose qualifiers will only apply to tests present +// in its own extension. +func (e *Extension) AddSuite(suite Suite) *Extension { + expr := fmt.Sprintf("source == %q", e.Component.Identifier()) + for i := range suite.Qualifiers { + suite.Qualifiers[i] = fmt.Sprintf("(%s) && (%s)", expr, suite.Qualifiers[i]) + } + e.AddGlobalSuite(suite) + return e +} + +func (e *Extension) FindSpecsByName(names ...string) (et.ExtensionTestSpecs, error) { + var specs et.ExtensionTestSpecs + var notFound []string + + for _, name := range names { + found := false + for i := range e.specs { + if e.specs[i].Name == name { + specs = append(specs, e.specs[i]) + found = true + break + } + } + if !found { + notFound = append(notFound, name) + } + } + + if len(notFound) > 0 { + return nil, fmt.Errorf("no such tests: %s", strings.Join(notFound, ", ")) + } + + return specs, nil +} + +func (e *Component) Identifier() string { + return fmt.Sprintf("%s:%s:%s", e.Product, e.Kind, e.Name) +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/result.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/result.go new file mode 100644 index 0000000000000..f33fb5c2745b2 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/result.go @@ -0,0 +1,12 @@ +package extensiontests + +func (results ExtensionTestResults) Walk(walkFn func(*ExtensionTestResult)) { + for i := range results { + walkFn(results[i]) + } +} + +// AddDetails adds additional information to an ExtensionTestResult. Value must marshal to JSON. +func (result *ExtensionTestResult) AddDetails(name string, value interface{}) { + result.Details = append(result.Details, Details{Name: name, Value: value}) +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/result_writer.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/result_writer.go new file mode 100644 index 0000000000000..821599791352c --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/result_writer.go @@ -0,0 +1,71 @@ +package extensiontests + +import ( + "encoding/json" + "fmt" + "io" +) + +type ResultWriter interface { + Write(result *ExtensionTestResult) + Flush() +} + +type NullResultWriter struct{} + +func (NullResultWriter) Write(*ExtensionTestResult) {} +func (NullResultWriter) Flush() {} + +type ResultFormat string + +var ( + JSON ResultFormat = "json" + JSONL ResultFormat = "jsonl" +) + +type JSONResultWriter struct { + out io.Writer + format ResultFormat + results ExtensionTestResults +} + +func NewResultWriter(out io.Writer, format ResultFormat) (*JSONResultWriter, error) { + switch format { + case JSON, JSONL: + // do nothing + default: + return nil, fmt.Errorf("unsupported result format: %s", format) + } + + return &JSONResultWriter{ + out: out, + format: format, + }, nil +} + +func (w *JSONResultWriter) Write(result *ExtensionTestResult) { + switch w.format { + case JSONL: + // JSONL gets written to out as we get the items + data, err := json.Marshal(result) + if err != nil { + panic(err) + } + fmt.Fprintf(w.out, "%s\n", string(data)) + case JSON: + w.results = append(w.results, result) + } +} + +func (w *JSONResultWriter) Flush() { + switch w.format { + case JSONL: + // we already wrote it out + case JSON: + data, err := json.MarshalIndent(w.results, "", " ") + if err != nil { + panic(err) + } + fmt.Fprintf(w.out, "%s\n", string(data)) + } +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/spec.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/spec.go new file mode 100644 index 0000000000000..f77ab9f383132 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/spec.go @@ -0,0 +1,274 @@ +package extensiontests + +import ( + "fmt" + "sync" + "sync/atomic" + "time" + + "github.com/google/cel-go/cel" + "github.com/google/cel-go/checker/decls" + "github.com/google/cel-go/common/types" + + "github.com/openshift-eng/openshift-tests-extension/pkg/dbtime" +) + +// Walk iterates over all test specs, and executions the function provided. The test spec can be mutated. +func (specs ExtensionTestSpecs) Walk(walkFn func(*ExtensionTestSpec)) ExtensionTestSpecs { + for i := range specs { + walkFn(specs[i]) + } + + return specs +} + +func (specs ExtensionTestSpecs) Names() []string { + var names []string + for _, spec := range specs { + names = append(names, spec.Name) + } + return names +} + +// Run executes all the specs in parallel, up to maxConcurrent at the same time. Results +// are written to the given ResultWriter after each spec has completed execution. BeforeEach, +// BeforeAll, AfterEach, AfterAll hooks are executed when specified. "Each" hooks must be thread +// safe. Returns an error if any test spec failed, indicating the quantity of failures. +func (specs ExtensionTestSpecs) Run(w ResultWriter, maxConcurrent int) error { + queue := make(chan *ExtensionTestSpec) + failures := atomic.Int64{} + + // Execute beforeAll + for _, spec := range specs { + for _, beforeAllTask := range spec.beforeAll { + beforeAllTask.Run() + } + } + + // Feed the queue + go func() { + specs.Walk(func(spec *ExtensionTestSpec) { + queue <- spec + }) + close(queue) + }() + + // Start consumers + var wg sync.WaitGroup + for i := 0; i < maxConcurrent; i++ { + wg.Add(1) + go func() { + defer wg.Done() + for spec := range queue { + for _, beforeEachTask := range spec.beforeEach { + beforeEachTask.Run(*spec) + } + + res := runSpec(spec) + if res.Result == ResultFailed { + failures.Add(1) + } + + for _, afterEachTask := range spec.afterEach { + afterEachTask.Run(res) + } + w.Write(res) + } + }() + } + + // Wait for all consumers to finish + wg.Wait() + + // Execute afterAll + for _, spec := range specs { + for _, afterAllTask := range spec.afterAll { + afterAllTask.Run() + } + } + + failCount := failures.Load() + if failCount > 0 { + return fmt.Errorf("%d tests failed", failCount) + } + return nil +} + +// AddBeforeAll adds a function to be run once before all tests start executing. +func (specs ExtensionTestSpecs) AddBeforeAll(fn func()) { + task := &OneTimeTask{fn: fn} + specs.Walk(func(spec *ExtensionTestSpec) { + spec.beforeAll = append(spec.beforeAll, task) + }) +} + +// AddAfterAll adds a function to be run once after all tests have finished. +func (specs ExtensionTestSpecs) AddAfterAll(fn func()) { + task := &OneTimeTask{fn: fn} + specs.Walk(func(spec *ExtensionTestSpec) { + spec.afterAll = append(spec.afterAll, task) + }) +} + +// AddBeforeEach adds a function that runs before each test starts executing. The ExtensionTestSpec is +// passed in for contextual information, but must not be modified. The provided function must be thread +// safe. +func (specs ExtensionTestSpecs) AddBeforeEach(fn func(spec ExtensionTestSpec)) { + task := &SpecTask{fn: fn} + specs.Walk(func(spec *ExtensionTestSpec) { + spec.beforeEach = append(spec.beforeEach, task) + }) +} + +// AddAfterEach adds a function that runs after each test has finished executing. The ExtensionTestResult +// can be modified if needed. The provided function must be thread safe. +func (specs ExtensionTestSpecs) AddAfterEach(fn func(task *ExtensionTestResult)) { + task := &TestResultTask{fn: fn} + specs.Walk(func(spec *ExtensionTestSpec) { + spec.afterEach = append(spec.afterEach, task) + }) +} + +// MustFilter filters specs using the given celExprs. Each celExpr is OR'd together, if any +// match the spec is included in the filtered set. If your CEL expression is invalid or filtering +// otherwise fails, this function panics. +func (specs ExtensionTestSpecs) MustFilter(celExprs []string) ExtensionTestSpecs { + specs, err := specs.Filter(celExprs) + if err != nil { + panic(fmt.Sprintf("filter did not succeed: %s", err.Error())) + } + + return specs +} + +// Filter filters specs using the given celExprs. Each celExpr is OR'd together, if any +// match the spec is included in the filtered set. +func (specs ExtensionTestSpecs) Filter(celExprs []string) (ExtensionTestSpecs, error) { + var filteredSpecs ExtensionTestSpecs + + // Empty filters returns all + if len(celExprs) == 0 { + return specs, nil + } + + env, err := cel.NewEnv( + cel.Declarations( + decls.NewVar("source", decls.String), + decls.NewVar("name", decls.String), + decls.NewVar("originalName", decls.String), + decls.NewVar("labels", decls.NewListType(decls.String)), + decls.NewVar("tags", decls.NewMapType(decls.String, decls.String)), + ), + ) + if err != nil { + return nil, fmt.Errorf("failed to create CEL environment: %w", err) + } + + // OR all expressions together + for _, spec := range specs { + include := false + for _, celExpr := range celExprs { + // Parse CEL expression + ast, iss := env.Parse(celExpr) + if iss.Err() != nil { + return nil, fmt.Errorf("error parsing CEL expression '%s': %v", celExpr, iss.Err()) + } + + // Check the AST + checked, iss := env.Check(ast) + if iss.Err() != nil { + return nil, fmt.Errorf("error checking CEL expression '%s': %v", celExpr, iss.Err()) + } + + // Create a CEL program from the checked AST + prg, err := env.Program(checked) + if err != nil { + return nil, fmt.Errorf("error creating CEL program: %v", err) + } + + out, _, err := prg.Eval(map[string]interface{}{ + "name": spec.Name, + "source": spec.Source, + "originalName": spec.OriginalName, + "labels": spec.Labels.UnsortedList(), + "tags": spec.Tags, + }) + if err != nil { + return nil, fmt.Errorf("error evaluating CEL expression: %v", err) + } + + // If any CEL expression evaluates to true, include the TestSpec + if out == types.True { + include = true + break + } + } + if include { + filteredSpecs = append(filteredSpecs, spec) + } + } + + return filteredSpecs, nil +} + +// AddLabel adds the labels to each spec. +func (specs ExtensionTestSpecs) AddLabel(labels ...string) ExtensionTestSpecs { + for i := range specs { + specs[i].Labels.Insert(labels...) + } + + return specs +} + +// RemoveLabel removes the labels from each spec. +func (specs ExtensionTestSpecs) RemoveLabel(labels ...string) ExtensionTestSpecs { + for i := range specs { + specs[i].Labels.Delete(labels...) + } + + return specs +} + +// SetTag specifies a key/value pair for each spec. +func (specs ExtensionTestSpecs) SetTag(key, value string) ExtensionTestSpecs { + for i := range specs { + specs[i].Tags[key] = value + } + + return specs +} + +// UnsetTag removes the specified key from each spec. +func (specs ExtensionTestSpecs) UnsetTag(key string) ExtensionTestSpecs { + for i := range specs { + delete(specs[i].Tags, key) + } + + return specs +} + +func runSpec(spec *ExtensionTestSpec) *ExtensionTestResult { + startTime := time.Now().UTC() + res := spec.Run() + duration := time.Since(startTime) + endTime := startTime.Add(duration).UTC() + if res == nil { + // this shouldn't happen + panic(fmt.Sprintf("test produced no result: %s", spec.Name)) + } + + res.Lifecycle = spec.Lifecycle + + // If the runner doesn't populate this info, we should set it + if res.StartTime == nil { + res.StartTime = dbtime.Ptr(startTime) + } + if res.EndTime == nil { + res.EndTime = dbtime.Ptr(endTime) + } + if res.Duration == 0 { + res.Duration = duration.Milliseconds() + } + + return res +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/task.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/task.go new file mode 100644 index 0000000000000..e808bea87bb8b --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/task.go @@ -0,0 +1,31 @@ +package extensiontests + +import "sync/atomic" + +type SpecTask struct { + fn func(spec ExtensionTestSpec) +} + +func (t *SpecTask) Run(spec ExtensionTestSpec) { + t.fn(spec) +} + +type TestResultTask struct { + fn func(result *ExtensionTestResult) +} + +func (t *TestResultTask) Run(result *ExtensionTestResult) { + t.fn(result) +} + +type OneTimeTask struct { + fn func() + executed int32 // Atomic boolean to indicate whether the function has been run +} + +func (t *OneTimeTask) Run() { + // Ensure one-time tasks are only run once + if atomic.CompareAndSwapInt32(&t.executed, 0, 1) { + t.fn() + } +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/types.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/types.go new file mode 100644 index 0000000000000..2d507d6feeef8 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests/types.go @@ -0,0 +1,90 @@ +package extensiontests + +import ( + "k8s.io/apimachinery/pkg/util/sets" + + "github.com/openshift-eng/openshift-tests-extension/pkg/dbtime" +) + +type Lifecycle string + +var LifecycleInforming Lifecycle = "informing" +var LifecycleBlocking Lifecycle = "blocking" + +type ExtensionTestSpecs []*ExtensionTestSpec + +type ExtensionTestSpec struct { + Name string `json:"name"` + + // OriginalName contains the very first name this test was ever known as, used to preserve + // history across all names. + OriginalName string `json:"originalName,omitempty"` + + // Labels are single string values to apply to the test spec + Labels sets.Set[string] `json:"labels"` + + // Tags are key:value pairs + Tags map[string]string `json:"tags,omitempty"` + + // Resources gives optional information about what's required to run this test. + Resources Resources `json:"resources"` + + // Source is the origin of the test. + Source string `json:"source"` + + // Lifecycle informs the executor whether the test is informing only, and should not cause the + // overall job run to fail, or if it's blocking where a failure of the test is fatal. + // Informing lifecycle tests can be used temporarily to gather information about a test's stability. + // Tests must not remain informing forever. + Lifecycle Lifecycle `json:"lifecycle"` + + // Run invokes a test + Run func() *ExtensionTestResult `json:"-"` + + // Hook functions + afterAll []*OneTimeTask + beforeAll []*OneTimeTask + afterEach []*TestResultTask + beforeEach []*SpecTask +} + +type Resources struct { + Isolation Isolation `json:"isolation"` + Memory string `json:"memory,omitempty"` + Duration string `json:"duration,omitempty"` + Timeout string `json:"timeout,omitempty"` +} + +type Isolation struct { + Mode string `json:"mode,omitempty"` + Conflict []string `json:"conflict,omitempty"` +} + +type ExtensionTestResults []*ExtensionTestResult + +type Result string + +var ResultPassed Result = "passed" +var ResultSkipped Result = "skipped" +var ResultFailed Result = "failed" + +type ExtensionTestResult struct { + Name string `json:"name"` + Lifecycle Lifecycle `json:"lifecycle"` + Duration int64 `json:"duration"` + StartTime *dbtime.DBTime `json:"startTime"` + EndTime *dbtime.DBTime `json:"endTime"` + Result Result `json:"result"` + Output string `json:"output"` + Error string `json:"error,omitempty"` + Details []Details `json:"details,omitempty"` +} + +// Details are human-readable messages to further explain skips, timeouts, etc. +// It can also be used to provide contemporaneous information about failures +// that may not be easily returned by must-gather. For larger artifacts (greater than +// 10KB, write them to $EXTENSION_ARTIFACTS_DIR. +type Details struct { + Name string `json:"name"` + Value interface{} `json:"value"` +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/registry.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/registry.go new file mode 100644 index 0000000000000..bbae421df7747 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/registry.go @@ -0,0 +1,39 @@ +package extension + +const DefaultExtension = "default" + +type Registry struct { + extensions map[string]*Extension +} + +func NewRegistry() *Registry { + var r Registry + return &r +} + +func (r *Registry) Walk(walkFn func(*Extension)) { + for k := range r.extensions { + if k == DefaultExtension { + continue + } + walkFn(r.extensions[k]) + } +} + +func (r *Registry) Get(name string) *Extension { + return r.extensions[name] +} + +func (r *Registry) Register(extension *Extension) { + if r.extensions == nil { + r.extensions = make(map[string]*Extension) + // first extension is default + r.extensions[DefaultExtension] = extension + } + + r.extensions[extension.Component.Identifier()] = extension +} + +func (r *Registry) Deregister(name string) { + delete(r.extensions, name) +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/types.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/types.go new file mode 100644 index 0000000000000..2e840c9e88df9 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/extension/types.go @@ -0,0 +1,55 @@ +package extension + +import ( + "k8s.io/apimachinery/pkg/util/sets" + + "github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests" +) + +const CurrentExtensionAPIVersion = "v1.0" + +// Extension represents an extension to openshift-tests. +type Extension struct { + APIVersion string `json:"apiVersion"` + Source Source `json:"source"` + Component Component `json:"component"` + + // Suites that the extension wants to advertise/participate in. + Suites []Suite `json:"suites"` + + // Private data + specs extensiontests.ExtensionTestSpecs + obsoleteTests sets.Set[string] +} + +// Source contains the details of the commit and source URL. +type Source struct { + // Commit from which this binary was compiled. + Commit string `json:"commit"` + // BuildDate ISO8601 string of when the binary was built + BuildDate string `json:"build_date"` + // GitTreeState lets you know the status of the git tree (clean/dirty) + GitTreeState string `json:"git_tree_state"` + // SourceURL contains the url of the git repository (if known) that this extension was built from. + SourceURL string `json:"source_url,omitempty"` +} + +// Component represents the component the binary acts on. +type Component struct { + // The product this component is part of. + Product string `json:"product"` + // The type of the component. + Kind string `json:"type"` + // The name of the component. + Name string `json:"name"` +} + +// Suite represents additional suites the extension wants to advertise. +type Suite struct { + // The name of the suite. + Name string `json:"name"` + // Parent suites this suite is part of. + Parents []string `json:"parents,omitempty"` + // Qualifiers are CEL expressions that are OR'd together for test selection that are members of the suite. + Qualifiers []string `json:"qualifiers,omitempty"` +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/component.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/component.go new file mode 100644 index 0000000000000..55e91819ba9ef --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/component.go @@ -0,0 +1,25 @@ +package flags + +import ( + "github.com/spf13/pflag" + + "github.com/openshift-eng/openshift-tests-extension/pkg/extension" +) + +// ComponentFlags contains information for specifying the component. +type ComponentFlags struct { + Component string +} + +func NewComponentFlags() *ComponentFlags { + return &ComponentFlags{ + Component: extension.DefaultExtension, + } +} + +func (f *ComponentFlags) BindFlags(fs *pflag.FlagSet) { + fs.StringVar(&f.Component, + "component", + f.Component, + "specify the component to enable") +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/concurrency.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/concurrency.go new file mode 100644 index 0000000000000..2db07c7654757 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/concurrency.go @@ -0,0 +1,23 @@ +package flags + +import "github.com/spf13/pflag" + +// ConcurrencyFlags contains information for configuring concurrency +type ConcurrencyFlags struct { + MaxConcurency int +} + +func NewConcurrencyFlags() *ConcurrencyFlags { + return &ConcurrencyFlags{ + MaxConcurency: 10, + } +} + +func (f *ConcurrencyFlags) BindFlags(fs *pflag.FlagSet) { + fs.IntVarP(&f.MaxConcurency, + "max-concurrency", + "c", + f.MaxConcurency, + "maximum number of tests to run in parallel", + ) +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/names.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/names.go new file mode 100644 index 0000000000000..b073f1a26b629 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/names.go @@ -0,0 +1,24 @@ +package flags + +import ( + "github.com/spf13/pflag" +) + +// NamesFlags contains information for specifying multiple test names. +type NamesFlags struct { + Names []string +} + +func NewNamesFlags() *NamesFlags { + return &NamesFlags{ + Names: []string{}, + } +} + +func (f *NamesFlags) BindFlags(fs *pflag.FlagSet) { + fs.StringSliceVarP(&f.Names, + "names", + "n", + f.Names, + "specify test name (can be specified multiple times)") +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/output.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/output.go new file mode 100644 index 0000000000000..24f49f6387b16 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/output.go @@ -0,0 +1,95 @@ +package flags + +import ( + "encoding/json" + "reflect" + "strings" + + "github.com/pkg/errors" + "github.com/spf13/pflag" +) + +// OutputFlags contains information for specifying multiple test names. +type OutputFlags struct { + Output string +} + +func NewOutputFlags() *OutputFlags { + return &OutputFlags{ + Output: "json", + } +} + +func (f *OutputFlags) BindFlags(fs *pflag.FlagSet) { + fs.StringVarP(&f.Output, + "output", + "o", + f.Output, + "output mode") +} + +func (o *OutputFlags) Marshal(v interface{}) ([]byte, error) { + switch o.Output { + case "", "json": + j, err := json.MarshalIndent(&v, "", " ") + if err != nil { + return nil, err + } + return j, nil + case "jsonl": + // Check if v is a slice or array + val := reflect.ValueOf(v) + if val.Kind() == reflect.Slice || val.Kind() == reflect.Array { + var result []byte + for i := 0; i < val.Len(); i++ { + item := val.Index(i).Interface() + j, err := json.Marshal(item) + if err != nil { + return nil, err + } + result = append(result, j...) + result = append(result, '\n') // Append newline after each item + } + return result, nil + } + return nil, errors.New("jsonl format requires a slice or array") + case "names": + val := reflect.ValueOf(v) + if val.Kind() == reflect.Slice || val.Kind() == reflect.Array { + var names []string + outerLoop: + for i := 0; i < val.Len(); i++ { + item := val.Index(i) + // Check for Name() or Identifier() methods + itemInterface := item.Interface() + nameFuncs := []string{"Name", "Identifier"} + for _, fn := range nameFuncs { + method := reflect.ValueOf(itemInterface).MethodByName(fn) + if method.IsValid() && method.Kind() == reflect.Func && method.Type().NumIn() == 0 && method.Type().NumOut() == 1 && method.Type().Out(0).Kind() == reflect.String { + name := method.Call(nil)[0].String() + names = append(names, name) + continue outerLoop + } + } + + // Dereference pointer if needed + if item.Kind() == reflect.Ptr { + item = item.Elem() + } + // Check for struct with Name field + if item.Kind() == reflect.Struct { + nameField := item.FieldByName("Name") + if nameField.IsValid() && nameField.Kind() == reflect.String { + names = append(names, nameField.String()) + } + } else { + return nil, errors.New("items must have a Name field or a Name() method") + } + } + return []byte(strings.Join(names, "\n")), nil + } + return nil, errors.New("names format requires an array of structs") + default: + return nil, errors.Errorf("invalid output format: %s", o.Output) + } +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/suite.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/suite.go new file mode 100644 index 0000000000000..23de832a850ab --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/flags/suite.go @@ -0,0 +1,21 @@ +package flags + +import ( + "github.com/spf13/pflag" +) + +// SuiteFlags contains information for specifying the suite. +type SuiteFlags struct { + Suite string +} + +func NewSuiteFlags() *SuiteFlags { + return &SuiteFlags{} +} + +func (f *SuiteFlags) BindFlags(fs *pflag.FlagSet) { + fs.StringVar(&f.Suite, + "suite", + f.Suite, + "specify the suite to use") +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/ginkgo/util.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/ginkgo/util.go new file mode 100644 index 0000000000000..9faea67f849f9 --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/ginkgo/util.go @@ -0,0 +1,153 @@ +package ginkgo + +import ( + "fmt" + "os" + "strings" + "sync" + "time" + + "github.com/onsi/ginkgo/v2" + "github.com/onsi/ginkgo/v2/types" + "github.com/onsi/gomega" + "github.com/pkg/errors" + "k8s.io/apimachinery/pkg/util/sets" + + ext "github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests" +) + +func configureGinkgo() (*types.SuiteConfig, *types.ReporterConfig, error) { + if !ginkgo.GetSuite().InPhaseBuildTree() { + if err := ginkgo.GetSuite().BuildTree(); err != nil { + return nil, nil, errors.Wrapf(err, "couldn't build ginkgo tree") + } + } + + // Ginkgo initialization + ginkgo.GetSuite().ClearBeforeAndAfterSuiteNodes() + suiteConfig, reporterConfig := ginkgo.GinkgoConfiguration() + suiteConfig.RandomizeAllSpecs = true + suiteConfig.Timeout = 24 * time.Hour + reporterConfig.NoColor = true + reporterConfig.Verbose = true + ginkgo.SetReporterConfig(reporterConfig) + + // Write output to Stderr + ginkgo.GinkgoWriter = ginkgo.NewWriter(os.Stderr) + + gomega.RegisterFailHandler(ginkgo.Fail) + + return &suiteConfig, &reporterConfig, nil +} + +func BuildExtensionTestSpecsFromOpenShiftGinkgoSuite() (ext.ExtensionTestSpecs, error) { + var tests []*ext.ExtensionTestSpec + var enforceSerialExecutionForGinkgo sync.Mutex // in-process parallelization for ginkgo is impossible so far + + if _, _, err := configureGinkgo(); err != nil { + return nil, err + } + + cwd, err := os.Getwd() + if err != nil { + return nil, errors.Wrap(err, "couldn't get current working directory") + } + + ginkgo.GetSuite().WalkTests(func(name string, spec types.TestSpec) { + testCase := &ext.ExtensionTestSpec{ + Name: spec.Text(), + Labels: sets.New[string](spec.Labels()...), + Lifecycle: GetLifecycle(spec.Labels()), + Run: func() *ext.ExtensionTestResult { + enforceSerialExecutionForGinkgo.Lock() + defer enforceSerialExecutionForGinkgo.Unlock() + + suiteConfig, reporterConfig, _ := configureGinkgo() + + result := &ext.ExtensionTestResult{ + Name: spec.Text(), + } + + var summary types.SpecReport + ginkgo.GetSuite().RunSpec(spec, ginkgo.Labels{}, "", cwd, ginkgo.GetFailer(), ginkgo.GetWriter(), *suiteConfig, + *reporterConfig) + for _, report := range ginkgo.GetSuite().GetReport().SpecReports { + if report.NumAttempts > 0 { + summary = report + } + } + + result.Output = summary.CapturedGinkgoWriterOutput + result.Error = summary.CapturedStdOutErr + + switch { + case summary.State == types.SpecStatePassed: + result.Result = ext.ResultPassed + case summary.State == types.SpecStateSkipped: + result.Result = ext.ResultSkipped + case summary.State == types.SpecStateFailed, summary.State == types.SpecStatePanicked, summary.State == types.SpecStateInterrupted: + result.Result = ext.ResultFailed + var errors []string + if len(summary.Failure.ForwardedPanic) > 0 { + if len(summary.Failure.Location.FullStackTrace) > 0 { + errors = append(errors, fmt.Sprintf("\n%s\n", summary.Failure.Location.FullStackTrace)) + } + errors = append(errors, fmt.Sprintf("fail [%s:%d]: Test Panicked: %s", lastFilenameSegment(summary.Failure.Location.FileName), summary.Failure.Location.LineNumber, summary.Failure.ForwardedPanic)) + } + errors = append(errors, fmt.Sprintf("fail [%s:%d]: %s", lastFilenameSegment(summary.Failure.Location.FileName), summary.Failure.Location.LineNumber, summary.Failure.Message)) + result.Error = strings.Join(errors, "\n") + default: + panic(fmt.Sprintf("test produced unknown outcome: %#v", summary)) + } + + return result + }, + } + tests = append(tests, testCase) + }) + + return tests, nil +} + +func Informing() ginkgo.Labels { + return ginkgo.Label(fmt.Sprintf("Lifecycle:%s", ext.LifecycleInforming)) +} + +func Slow() ginkgo.Labels { + return ginkgo.Label("SLOW") +} + +func Blocking() ginkgo.Labels { + return ginkgo.Label(fmt.Sprintf("Lifecycle:%s", ext.LifecycleBlocking)) +} + +func GetLifecycle(labels ginkgo.Labels) ext.Lifecycle { + for _, label := range labels { + res := strings.Split(label, ":") + if len(res) != 2 || !strings.EqualFold(res[0], "lifecycle") { + continue + } + return MustLifecycle(res[1]) // this panics if unsupported lifecycle is used + } + + return ext.LifecycleBlocking +} + +func MustLifecycle(l string) ext.Lifecycle { + switch ext.Lifecycle(l) { + case ext.LifecycleInforming, ext.LifecycleBlocking: + return ext.Lifecycle(l) + default: + panic(fmt.Sprintf("unknown test lifecycle: %s", l)) + } +} + +func lastFilenameSegment(filename string) string { + if parts := strings.Split(filename, "/vendor/"); len(parts) > 1 { + return parts[len(parts)-1] + } + if parts := strings.Split(filename, "/src/"); len(parts) > 1 { + return parts[len(parts)-1] + } + return filename +} diff --git a/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/version/version.go b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/version/version.go new file mode 100644 index 0000000000000..7d6a3309b312b --- /dev/null +++ b/vendor/github.com/openshift-eng/openshift-tests-extension/pkg/version/version.go @@ -0,0 +1,11 @@ +package version + +var ( + // CommitFromGit is a constant representing the source version that + // generated this build. It should be set during build via -ldflags. + CommitFromGit string + // BuildDate in ISO8601 format, output of $(date -u +'%Y-%m-%dT%H:%M:%SZ') + BuildDate string + // GitTreeState has the state of git tree, either "clean" or "dirty" + GitTreeState string +) diff --git a/vendor/golang.org/x/mod/LICENSE b/vendor/golang.org/x/mod/LICENSE index 6a66aea5eafe0..2a7cf70da6e49 100644 --- a/vendor/golang.org/x/mod/LICENSE +++ b/vendor/golang.org/x/mod/LICENSE @@ -1,4 +1,4 @@ -Copyright (c) 2009 The Go Authors. All rights reserved. +Copyright 2009 The Go Authors. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are @@ -10,7 +10,7 @@ notice, this list of conditions and the following disclaimer. copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. - * Neither the name of Google Inc. nor the names of its + * Neither the name of Google LLC nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. diff --git a/vendor/golang.org/x/oauth2/LICENSE b/vendor/golang.org/x/oauth2/LICENSE index 6a66aea5eafe0..2a7cf70da6e49 100644 --- a/vendor/golang.org/x/oauth2/LICENSE +++ b/vendor/golang.org/x/oauth2/LICENSE @@ -1,4 +1,4 @@ -Copyright (c) 2009 The Go Authors. All rights reserved. +Copyright 2009 The Go Authors. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are @@ -10,7 +10,7 @@ notice, this list of conditions and the following disclaimer. copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. - * Neither the name of Google Inc. nor the names of its + * Neither the name of Google LLC nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. diff --git a/vendor/golang.org/x/tools/LICENSE b/vendor/golang.org/x/tools/LICENSE index 6a66aea5eafe0..2a7cf70da6e49 100644 --- a/vendor/golang.org/x/tools/LICENSE +++ b/vendor/golang.org/x/tools/LICENSE @@ -1,4 +1,4 @@ -Copyright (c) 2009 The Go Authors. All rights reserved. +Copyright 2009 The Go Authors. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are @@ -10,7 +10,7 @@ notice, this list of conditions and the following disclaimer. copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. - * Neither the name of Google Inc. nor the names of its + * Neither the name of Google LLC nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. diff --git a/vendor/golang.org/x/tools/go/ast/astutil/enclosing.go b/vendor/golang.org/x/tools/go/ast/astutil/enclosing.go index 2c4c4e232890a..6e34df46130b7 100644 --- a/vendor/golang.org/x/tools/go/ast/astutil/enclosing.go +++ b/vendor/golang.org/x/tools/go/ast/astutil/enclosing.go @@ -106,8 +106,21 @@ func PathEnclosingInterval(root *ast.File, start, end token.Pos) (path []ast.Nod // Does augmented child strictly contain [start, end)? if augPos <= start && end <= augEnd { - _, isToken := child.(tokenNode) - return isToken || visit(child) + if is[tokenNode](child) { + return true + } + + // childrenOf elides the FuncType node beneath FuncDecl. + // Add it back here for TypeParams, Params, Results, + // all FieldLists). But we don't add it back for the "func" token + // even though it is is the tree at FuncDecl.Type.Func. + if decl, ok := node.(*ast.FuncDecl); ok { + if fields, ok := child.(*ast.FieldList); ok && fields != decl.Recv { + path = append(path, decl.Type) + } + } + + return visit(child) } // Does [start, end) overlap multiple children? @@ -313,6 +326,8 @@ func childrenOf(n ast.Node) []ast.Node { // // As a workaround, we inline the case for FuncType // here and order things correctly. + // We also need to insert the elided FuncType just + // before the 'visit' recursion. // children = nil // discard ast.Walk(FuncDecl) info subtrees children = append(children, tok(n.Type.Func, len("func"))) @@ -632,3 +647,8 @@ func NodeDescription(n ast.Node) string { } panic(fmt.Sprintf("unexpected node type: %T", n)) } + +func is[T any](x any) bool { + _, ok := x.(T) + return ok +} diff --git a/vendor/golang.org/x/tools/go/ast/astutil/util.go b/vendor/golang.org/x/tools/go/ast/astutil/util.go index 919d5305ab422..6bdcf70ac2750 100644 --- a/vendor/golang.org/x/tools/go/ast/astutil/util.go +++ b/vendor/golang.org/x/tools/go/ast/astutil/util.go @@ -7,6 +7,7 @@ package astutil import "go/ast" // Unparen returns e with any enclosing parentheses stripped. +// TODO(adonovan): use go1.22's ast.Unparen. func Unparen(e ast.Expr) ast.Expr { for { p, ok := e.(*ast.ParenExpr) diff --git a/vendor/golang.org/x/tools/go/internal/packagesdriver/sizes.go b/vendor/golang.org/x/tools/go/internal/packagesdriver/sizes.go deleted file mode 100644 index 333676b7cfccf..0000000000000 --- a/vendor/golang.org/x/tools/go/internal/packagesdriver/sizes.go +++ /dev/null @@ -1,53 +0,0 @@ -// Copyright 2018 The Go Authors. All rights reserved. -// Use of this source code is governed by a BSD-style -// license that can be found in the LICENSE file. - -// Package packagesdriver fetches type sizes for go/packages and go/analysis. -package packagesdriver - -import ( - "context" - "fmt" - "strings" - - "golang.org/x/tools/internal/gocommand" -) - -func GetSizesForArgsGolist(ctx context.Context, inv gocommand.Invocation, gocmdRunner *gocommand.Runner) (string, string, error) { - inv.Verb = "list" - inv.Args = []string{"-f", "{{context.GOARCH}} {{context.Compiler}}", "--", "unsafe"} - stdout, stderr, friendlyErr, rawErr := gocmdRunner.RunRaw(ctx, inv) - var goarch, compiler string - if rawErr != nil { - rawErrMsg := rawErr.Error() - if strings.Contains(rawErrMsg, "cannot find main module") || - strings.Contains(rawErrMsg, "go.mod file not found") { - // User's running outside of a module. - // All bets are off. Get GOARCH and guess compiler is gc. - // TODO(matloob): Is this a problem in practice? - inv.Verb = "env" - inv.Args = []string{"GOARCH"} - envout, enverr := gocmdRunner.Run(ctx, inv) - if enverr != nil { - return "", "", enverr - } - goarch = strings.TrimSpace(envout.String()) - compiler = "gc" - } else if friendlyErr != nil { - return "", "", friendlyErr - } else { - // This should be unreachable, but be defensive - // in case RunRaw's error results are inconsistent. - return "", "", rawErr - } - } else { - fields := strings.Fields(stdout.String()) - if len(fields) < 2 { - return "", "", fmt.Errorf("could not parse GOARCH and Go compiler in format \" \":\nstdout: <<%s>>\nstderr: <<%s>>", - stdout.String(), stderr.String()) - } - goarch = fields[0] - compiler = fields[1] - } - return compiler, goarch, nil -} diff --git a/vendor/golang.org/x/tools/go/packages/doc.go b/vendor/golang.org/x/tools/go/packages/doc.go index a8d7b06ac0900..3531ac8f5fcf3 100644 --- a/vendor/golang.org/x/tools/go/packages/doc.go +++ b/vendor/golang.org/x/tools/go/packages/doc.go @@ -198,14 +198,6 @@ Instead, ssadump no longer requests the runtime package, but seeks it among the dependencies of the user-specified packages, and emits an error if it is not found. -Overlays: The Overlay field in the Config allows providing alternate contents -for Go source files, by providing a mapping from file path to contents. -go/packages will pull in new imports added in overlay files when go/packages -is run in LoadImports mode or greater. -Overlay support for the go list driver isn't complete yet: if the file doesn't -exist on disk, it will only be recognized in an overlay if it is a non-test file -and the package would be reported even without the overlay. - Questions & Tasks - Add GOARCH/GOOS? diff --git a/vendor/golang.org/x/tools/go/packages/external.go b/vendor/golang.org/x/tools/go/packages/external.go index 4335c1eb14c79..c2b4b711b59df 100644 --- a/vendor/golang.org/x/tools/go/packages/external.go +++ b/vendor/golang.org/x/tools/go/packages/external.go @@ -34,8 +34,8 @@ type DriverRequest struct { // Tests specifies whether the patterns should also return test packages. Tests bool `json:"tests"` - // Overlay maps file paths (relative to the driver's working directory) to the byte contents - // of overlay files. + // Overlay maps file paths (relative to the driver's working directory) + // to the contents of overlay files (see Config.Overlay). Overlay map[string][]byte `json:"overlay"` } @@ -119,7 +119,19 @@ func findExternalDriver(cfg *Config) driver { stderr := new(bytes.Buffer) cmd := exec.CommandContext(cfg.Context, tool, words...) cmd.Dir = cfg.Dir - cmd.Env = cfg.Env + // The cwd gets resolved to the real path. On Darwin, where + // /tmp is a symlink, this breaks anything that expects the + // working directory to keep the original path, including the + // go command when dealing with modules. + // + // os.Getwd stdlib has a special feature where if the + // cwd and the PWD are the same node then it trusts + // the PWD, so by setting it in the env for the child + // process we fix up all the paths returned by the go + // command. + // + // (See similar trick in Invocation.run in ../../internal/gocommand/invoke.go) + cmd.Env = append(slicesClip(cfg.Env), "PWD="+cfg.Dir) cmd.Stdin = bytes.NewReader(req) cmd.Stdout = buf cmd.Stderr = stderr @@ -138,3 +150,7 @@ func findExternalDriver(cfg *Config) driver { return &response, nil } } + +// slicesClip removes unused capacity from the slice, returning s[:len(s):len(s)]. +// TODO(adonovan): use go1.21 slices.Clip. +func slicesClip[S ~[]E, E any](s S) S { return s[:len(s):len(s)] } diff --git a/vendor/golang.org/x/tools/go/packages/golist.go b/vendor/golang.org/x/tools/go/packages/golist.go index 22305d9c90a1c..1a3a5b44f5c5f 100644 --- a/vendor/golang.org/x/tools/go/packages/golist.go +++ b/vendor/golang.org/x/tools/go/packages/golist.go @@ -21,7 +21,6 @@ import ( "sync" "unicode" - "golang.org/x/tools/go/internal/packagesdriver" "golang.org/x/tools/internal/gocommand" "golang.org/x/tools/internal/packagesinternal" ) @@ -149,7 +148,7 @@ func goListDriver(cfg *Config, patterns ...string) (_ *DriverResponse, err error if cfg.Mode&NeedTypesSizes != 0 || cfg.Mode&NeedTypes != 0 { errCh := make(chan error) go func() { - compiler, arch, err := packagesdriver.GetSizesForArgsGolist(ctx, state.cfgInvocation(), cfg.gocmdRunner) + compiler, arch, err := getSizesForArgs(ctx, state.cfgInvocation(), cfg.gocmdRunner) response.dr.Compiler = compiler response.dr.Arch = arch errCh <- err @@ -841,6 +840,7 @@ func (state *golistState) cfgInvocation() gocommand.Invocation { Env: cfg.Env, Logf: cfg.Logf, WorkingDir: cfg.Dir, + Overlay: cfg.goListOverlayFile, } } @@ -849,26 +849,6 @@ func (state *golistState) invokeGo(verb string, args ...string) (*bytes.Buffer, cfg := state.cfg inv := state.cfgInvocation() - - // For Go versions 1.16 and above, `go list` accepts overlays directly via - // the -overlay flag. Set it, if it's available. - // - // The check for "list" is not necessarily required, but we should avoid - // getting the go version if possible. - if verb == "list" { - goVersion, err := state.getGoVersion() - if err != nil { - return nil, err - } - if goVersion >= 16 { - filename, cleanup, err := state.writeOverlays() - if err != nil { - return nil, err - } - defer cleanup() - inv.Overlay = filename - } - } inv.Verb = verb inv.Args = args gocmdRunner := cfg.gocmdRunner @@ -1015,67 +995,6 @@ func (state *golistState) invokeGo(verb string, args ...string) (*bytes.Buffer, return stdout, nil } -// OverlayJSON is the format overlay files are expected to be in. -// The Replace map maps from overlaid paths to replacement paths: -// the Go command will forward all reads trying to open -// each overlaid path to its replacement path, or consider the overlaid -// path not to exist if the replacement path is empty. -// -// From golang/go#39958. -type OverlayJSON struct { - Replace map[string]string `json:"replace,omitempty"` -} - -// writeOverlays writes out files for go list's -overlay flag, as described -// above. -func (state *golistState) writeOverlays() (filename string, cleanup func(), err error) { - // Do nothing if there are no overlays in the config. - if len(state.cfg.Overlay) == 0 { - return "", func() {}, nil - } - dir, err := os.MkdirTemp("", "gopackages-*") - if err != nil { - return "", nil, err - } - // The caller must clean up this directory, unless this function returns an - // error. - cleanup = func() { - os.RemoveAll(dir) - } - defer func() { - if err != nil { - cleanup() - } - }() - overlays := map[string]string{} - for k, v := range state.cfg.Overlay { - // Create a unique filename for the overlaid files, to avoid - // creating nested directories. - noSeparator := strings.Join(strings.Split(filepath.ToSlash(k), "/"), "") - f, err := os.CreateTemp(dir, fmt.Sprintf("*-%s", noSeparator)) - if err != nil { - return "", func() {}, err - } - if _, err := f.Write(v); err != nil { - return "", func() {}, err - } - if err := f.Close(); err != nil { - return "", func() {}, err - } - overlays[k] = f.Name() - } - b, err := json.Marshal(OverlayJSON{Replace: overlays}) - if err != nil { - return "", func() {}, err - } - // Write out the overlay file that contains the filepath mappings. - filename = filepath.Join(dir, "overlay.json") - if err := os.WriteFile(filename, b, 0665); err != nil { - return "", func() {}, err - } - return filename, cleanup, nil -} - func containsGoFile(s []string) bool { for _, f := range s { if strings.HasSuffix(f, ".go") { @@ -1104,3 +1023,44 @@ func cmdDebugStr(cmd *exec.Cmd) string { } return fmt.Sprintf("GOROOT=%v GOPATH=%v GO111MODULE=%v GOPROXY=%v PWD=%v %v", env["GOROOT"], env["GOPATH"], env["GO111MODULE"], env["GOPROXY"], env["PWD"], strings.Join(args, " ")) } + +// getSizesForArgs queries 'go list' for the appropriate +// Compiler and GOARCH arguments to pass to [types.SizesFor]. +func getSizesForArgs(ctx context.Context, inv gocommand.Invocation, gocmdRunner *gocommand.Runner) (string, string, error) { + inv.Verb = "list" + inv.Args = []string{"-f", "{{context.GOARCH}} {{context.Compiler}}", "--", "unsafe"} + stdout, stderr, friendlyErr, rawErr := gocmdRunner.RunRaw(ctx, inv) + var goarch, compiler string + if rawErr != nil { + rawErrMsg := rawErr.Error() + if strings.Contains(rawErrMsg, "cannot find main module") || + strings.Contains(rawErrMsg, "go.mod file not found") { + // User's running outside of a module. + // All bets are off. Get GOARCH and guess compiler is gc. + // TODO(matloob): Is this a problem in practice? + inv.Verb = "env" + inv.Args = []string{"GOARCH"} + envout, enverr := gocmdRunner.Run(ctx, inv) + if enverr != nil { + return "", "", enverr + } + goarch = strings.TrimSpace(envout.String()) + compiler = "gc" + } else if friendlyErr != nil { + return "", "", friendlyErr + } else { + // This should be unreachable, but be defensive + // in case RunRaw's error results are inconsistent. + return "", "", rawErr + } + } else { + fields := strings.Fields(stdout.String()) + if len(fields) < 2 { + return "", "", fmt.Errorf("could not parse GOARCH and Go compiler in format \" \":\nstdout: <<%s>>\nstderr: <<%s>>", + stdout.String(), stderr.String()) + } + goarch = fields[0] + compiler = fields[1] + } + return compiler, goarch, nil +} diff --git a/vendor/golang.org/x/tools/go/packages/packages.go b/vendor/golang.org/x/tools/go/packages/packages.go index 3ea1b3fa46d0a..0b6bfaff8087f 100644 --- a/vendor/golang.org/x/tools/go/packages/packages.go +++ b/vendor/golang.org/x/tools/go/packages/packages.go @@ -37,10 +37,19 @@ import ( // A LoadMode controls the amount of detail to return when loading. // The bits below can be combined to specify which fields should be // filled in the result packages. +// // The zero value is a special case, equivalent to combining // the NeedName, NeedFiles, and NeedCompiledGoFiles bits. +// // ID and Errors (if present) will always be filled. -// Load may return more information than requested. +// [Load] may return more information than requested. +// +// Unfortunately there are a number of open bugs related to +// interactions among the LoadMode bits: +// - https://github.com/golang/go/issues/56633 +// - https://github.com/golang/go/issues/56677 +// - https://github.com/golang/go/issues/58726 +// - https://github.com/golang/go/issues/63517 type LoadMode int const ( @@ -66,7 +75,7 @@ const ( // NeedTypes adds Types, Fset, and IllTyped. NeedTypes - // NeedSyntax adds Syntax. + // NeedSyntax adds Syntax and Fset. NeedSyntax // NeedTypesInfo adds TypesInfo. @@ -123,7 +132,14 @@ const ( // A Config specifies details about how packages should be loaded. // The zero value is a valid configuration. +// // Calls to Load do not modify this struct. +// +// TODO(adonovan): #67702: this is currently false: in fact, +// calls to [Load] do not modify the public fields of this struct, but +// may modify hidden fields, so concurrent calls to [Load] must not +// use the same Config. But perhaps we should reestablish the +// documented invariant. type Config struct { // Mode controls the level of information returned for each package. Mode LoadMode @@ -199,13 +215,23 @@ type Config struct { // setting Tests may have no effect. Tests bool - // Overlay provides a mapping of absolute file paths to file contents. - // If the file with the given path already exists, the parser will use the - // alternative file contents provided by the map. + // Overlay is a mapping from absolute file paths to file contents. + // + // For each map entry, [Load] uses the alternative file + // contents provided by the overlay mapping instead of reading + // from the file system. This mechanism can be used to enable + // editor-integrated tools to correctly analyze the contents + // of modified but unsaved buffers, for example. // - // Overlays provide incomplete support for when a given file doesn't - // already exist on disk. See the package doc above for more details. + // The overlay mapping is passed to the build system's driver + // (see "The driver protocol") so that it too can report + // consistent package metadata about unsaved files. However, + // drivers may vary in their level of support for overlays. Overlay map[string][]byte + + // goListOverlayFile is the JSON file that encodes the Overlay + // mapping, used by 'go list -overlay=...' + goListOverlayFile string } // Load loads and returns the Go packages named by the given patterns. @@ -213,6 +239,20 @@ type Config struct { // Config specifies loading options; // nil behaves the same as an empty Config. // +// The [Config.Mode] field is a set of bits that determine what kinds +// of information should be computed and returned. Modes that require +// more information tend to be slower. See [LoadMode] for details +// and important caveats. Its zero value is equivalent to +// NeedName | NeedFiles | NeedCompiledGoFiles. +// +// Each call to Load returns a new set of [Package] instances. +// The Packages and their Imports form a directed acyclic graph. +// +// If the [NeedTypes] mode flag was set, each call to Load uses a new +// [types.Importer], so [types.Object] and [types.Type] values from +// different calls to Load must not be mixed as they will have +// inconsistent notions of type identity. +// // If any of the patterns was invalid as defined by the // underlying build system, Load returns an error. // It may return an empty list of packages without an error, @@ -286,6 +326,17 @@ func defaultDriver(cfg *Config, patterns ...string) (*DriverResponse, bool, erro // (fall through) } + // go list fallback + // + // Write overlays once, as there are many calls + // to 'go list' (one per chunk plus others too). + overlay, cleanupOverlay, err := gocommand.WriteOverlays(cfg.Overlay) + if err != nil { + return nil, false, err + } + defer cleanupOverlay() + cfg.goListOverlayFile = overlay + response, err := callDriverOnChunks(goListDriver, cfg, chunks) if err != nil { return nil, false, err @@ -365,6 +416,9 @@ func mergeResponses(responses ...*DriverResponse) *DriverResponse { } // A Package describes a loaded Go package. +// +// It also defines part of the JSON schema of [DriverResponse]. +// See the package documentation for an overview. type Package struct { // ID is a unique identifier for a package, // in a syntax provided by the underlying build system. @@ -423,6 +477,13 @@ type Package struct { // to corresponding loaded Packages. Imports map[string]*Package + // Module is the module information for the package if it exists. + // + // Note: it may be missing for std and cmd; see Go issue #65816. + Module *Module + + // -- The following fields are not part of the driver JSON schema. -- + // Types provides type information for the package. // The NeedTypes LoadMode bit sets this field for packages matching the // patterns; type information for dependencies may be missing or incomplete, @@ -431,15 +492,15 @@ type Package struct { // Each call to [Load] returns a consistent set of type // symbols, as defined by the comment at [types.Identical]. // Avoid mixing type information from two or more calls to [Load]. - Types *types.Package + Types *types.Package `json:"-"` // Fset provides position information for Types, TypesInfo, and Syntax. // It is set only when Types is set. - Fset *token.FileSet + Fset *token.FileSet `json:"-"` // IllTyped indicates whether the package or any dependency contains errors. // It is set only when Types is set. - IllTyped bool + IllTyped bool `json:"-"` // Syntax is the package's syntax trees, for the files listed in CompiledGoFiles. // @@ -449,26 +510,28 @@ type Package struct { // // Syntax is kept in the same order as CompiledGoFiles, with the caveat that nils are // removed. If parsing returned nil, Syntax may be shorter than CompiledGoFiles. - Syntax []*ast.File + Syntax []*ast.File `json:"-"` // TypesInfo provides type information about the package's syntax trees. // It is set only when Syntax is set. - TypesInfo *types.Info + TypesInfo *types.Info `json:"-"` // TypesSizes provides the effective size function for types in TypesInfo. - TypesSizes types.Sizes + TypesSizes types.Sizes `json:"-"` + + // -- internal -- // forTest is the package under test, if any. forTest string // depsErrors is the DepsErrors field from the go list response, if any. depsErrors []*packagesinternal.PackageError - - // module is the module information for the package if it exists. - Module *Module } // Module provides module information for a package. +// +// It also defines part of the JSON schema of [DriverResponse]. +// See the package documentation for an overview. type Module struct { Path string // module path Version string // module version @@ -601,6 +664,7 @@ func (p *Package) UnmarshalJSON(b []byte) error { OtherFiles: flat.OtherFiles, EmbedFiles: flat.EmbedFiles, EmbedPatterns: flat.EmbedPatterns, + IgnoredFiles: flat.IgnoredFiles, ExportFile: flat.ExportFile, } if len(flat.Imports) > 0 { @@ -896,12 +960,14 @@ func (ld *loader) refine(response *DriverResponse) ([]*Package, error) { } if ld.requestedMode&NeedTypes == 0 { ld.pkgs[i].Types = nil - ld.pkgs[i].Fset = nil ld.pkgs[i].IllTyped = false } if ld.requestedMode&NeedSyntax == 0 { ld.pkgs[i].Syntax = nil } + if ld.requestedMode&NeedTypes == 0 && ld.requestedMode&NeedSyntax == 0 { + ld.pkgs[i].Fset = nil + } if ld.requestedMode&NeedTypesInfo == 0 { ld.pkgs[i].TypesInfo = nil } @@ -1434,6 +1500,10 @@ func impliedLoadMode(loadMode LoadMode) LoadMode { // All these things require knowing the import graph. loadMode |= NeedImports } + if loadMode&NeedTypes != 0 { + // Types require the GoVersion from Module. + loadMode |= NeedModule + } return loadMode } diff --git a/vendor/golang.org/x/tools/go/packages/visit.go b/vendor/golang.org/x/tools/go/packages/visit.go index a1dcc40b7270d..df14ffd94dc3e 100644 --- a/vendor/golang.org/x/tools/go/packages/visit.go +++ b/vendor/golang.org/x/tools/go/packages/visit.go @@ -49,11 +49,20 @@ func Visit(pkgs []*Package, pre func(*Package) bool, post func(*Package)) { // PrintErrors returns the number of errors printed. func PrintErrors(pkgs []*Package) int { var n int + errModules := make(map[*Module]bool) Visit(pkgs, nil, func(pkg *Package) { for _, err := range pkg.Errors { fmt.Fprintln(os.Stderr, err) n++ } + + // Print pkg.Module.Error once if present. + mod := pkg.Module + if mod != nil && mod.Error != nil && !errModules[mod] { + errModules[mod] = true + fmt.Fprintln(os.Stderr, mod.Error.Err) + n++ + } }) return n } diff --git a/vendor/golang.org/x/tools/go/types/objectpath/objectpath.go b/vendor/golang.org/x/tools/go/types/objectpath/objectpath.go index a2386c347a250..9ada177758ff7 100644 --- a/vendor/golang.org/x/tools/go/types/objectpath/objectpath.go +++ b/vendor/golang.org/x/tools/go/types/objectpath/objectpath.go @@ -51,7 +51,7 @@ type Path string // // PO package->object Package.Scope.Lookup // OT object->type Object.Type -// TT type->type Type.{Elem,Key,Params,Results,Underlying} [EKPRU] +// TT type->type Type.{Elem,Key,{,{,Recv}Type}Params,Results,Underlying,Rhs} [EKPRUTrCa] // TO type->object Type.{At,Field,Method,Obj} [AFMO] // // All valid paths start with a package and end at an object @@ -63,8 +63,8 @@ type Path string // - The only PO operator is Package.Scope.Lookup, which requires an identifier. // - The only OT operator is Object.Type, // which we encode as '.' because dot cannot appear in an identifier. -// - The TT operators are encoded as [EKPRUTC]; -// one of these (TypeParam) requires an integer operand, +// - The TT operators are encoded as [EKPRUTrCa]; +// two of these ({,Recv}TypeParams) require an integer operand, // which is encoded as a string of decimal digits. // - The TO operators are encoded as [AFMO]; // three of these (At,Field,Method) require an integer operand, @@ -98,19 +98,21 @@ const ( opType = '.' // .Type() (Object) // type->type operators - opElem = 'E' // .Elem() (Pointer, Slice, Array, Chan, Map) - opKey = 'K' // .Key() (Map) - opParams = 'P' // .Params() (Signature) - opResults = 'R' // .Results() (Signature) - opUnderlying = 'U' // .Underlying() (Named) - opTypeParam = 'T' // .TypeParams.At(i) (Named, Signature) - opConstraint = 'C' // .Constraint() (TypeParam) + opElem = 'E' // .Elem() (Pointer, Slice, Array, Chan, Map) + opKey = 'K' // .Key() (Map) + opParams = 'P' // .Params() (Signature) + opResults = 'R' // .Results() (Signature) + opUnderlying = 'U' // .Underlying() (Named) + opTypeParam = 'T' // .TypeParams.At(i) (Named, Signature) + opRecvTypeParam = 'r' // .RecvTypeParams.At(i) (Signature) + opConstraint = 'C' // .Constraint() (TypeParam) + opRhs = 'a' // .Rhs() (Alias) // type->object operators - opAt = 'A' // .At(i) (Tuple) - opField = 'F' // .Field(i) (Struct) - opMethod = 'M' // .Method(i) (Named or Interface; not Struct: "promoted" names are ignored) - opObj = 'O' // .Obj() (Named, TypeParam) + opAt = 'A' // .At(i) (Tuple) + opField = 'F' // .Field(i) (Struct) + opMethod = 'M' // .Method(i) (Named or Interface; not Struct: "promoted" names are ignored) + opObj = 'O' // .Obj() (Named, TypeParam) ) // For is equivalent to new(Encoder).For(obj). @@ -278,21 +280,26 @@ func (enc *Encoder) For(obj types.Object) (Path, error) { path = append(path, opType) T := o.Type() + if alias, ok := T.(*aliases.Alias); ok { + if r := findTypeParam(obj, aliases.TypeParams(alias), path, opTypeParam, nil); r != nil { + return Path(r), nil + } + if r := find(obj, aliases.Rhs(alias), append(path, opRhs), nil); r != nil { + return Path(r), nil + } - if tname.IsAlias() { - // type alias + } else if tname.IsAlias() { + // legacy alias if r := find(obj, T, path, nil); r != nil { return Path(r), nil } - } else { - if named, _ := T.(*types.Named); named != nil { - if r := findTypeParam(obj, named.TypeParams(), path, nil); r != nil { - // generic named type - return Path(r), nil - } - } + + } else if named, ok := T.(*types.Named); ok { // defined (named) type - if r := find(obj, T.Underlying(), append(path, opUnderlying), nil); r != nil { + if r := findTypeParam(obj, named.TypeParams(), path, opTypeParam, nil); r != nil { + return Path(r), nil + } + if r := find(obj, named.Underlying(), append(path, opUnderlying), nil); r != nil { return Path(r), nil } } @@ -462,7 +469,10 @@ func find(obj types.Object, T types.Type, path []byte, seen map[*types.TypeName] } return find(obj, T.Elem(), append(path, opElem), seen) case *types.Signature: - if r := findTypeParam(obj, T.TypeParams(), path, seen); r != nil { + if r := findTypeParam(obj, T.RecvTypeParams(), path, opRecvTypeParam, nil); r != nil { + return r + } + if r := findTypeParam(obj, T.TypeParams(), path, opTypeParam, seen); r != nil { return r } if r := find(obj, T.Params(), append(path, opParams), seen); r != nil { @@ -525,10 +535,10 @@ func find(obj types.Object, T types.Type, path []byte, seen map[*types.TypeName] panic(T) } -func findTypeParam(obj types.Object, list *types.TypeParamList, path []byte, seen map[*types.TypeName]bool) []byte { +func findTypeParam(obj types.Object, list *types.TypeParamList, path []byte, op byte, seen map[*types.TypeName]bool) []byte { for i := 0; i < list.Len(); i++ { tparam := list.At(i) - path2 := appendOpArg(path, opTypeParam, i) + path2 := appendOpArg(path, op, i) if r := find(obj, tparam, path2, seen); r != nil { return r } @@ -580,10 +590,10 @@ func Object(pkg *types.Package, p Path) (types.Object, error) { code := suffix[0] suffix = suffix[1:] - // Codes [AFM] have an integer operand. + // Codes [AFMTr] have an integer operand. var index int switch code { - case opAt, opField, opMethod, opTypeParam: + case opAt, opField, opMethod, opTypeParam, opRecvTypeParam: rest := strings.TrimLeft(suffix, "0123456789") numerals := suffix[:len(suffix)-len(rest)] suffix = rest @@ -653,6 +663,16 @@ func Object(pkg *types.Package, p Path) (types.Object, error) { } t = named.Underlying() + case opRhs: + if alias, ok := t.(*aliases.Alias); ok { + t = aliases.Rhs(alias) + } else if false && aliases.Enabled() { + // The Enabled check is too expensive, so for now we + // simply assume that aliases are not enabled. + // TODO(adonovan): replace with "if true {" when go1.24 is assured. + return nil, fmt.Errorf("cannot apply %q to %s (got %T, want alias)", code, t, t) + } + case opTypeParam: hasTypeParams, ok := t.(hasTypeParams) // Named, Signature if !ok { @@ -664,6 +684,17 @@ func Object(pkg *types.Package, p Path) (types.Object, error) { } t = tparams.At(index) + case opRecvTypeParam: + sig, ok := t.(*types.Signature) // Signature + if !ok { + return nil, fmt.Errorf("cannot apply %q to %s (got %T, want signature)", code, t, t) + } + rtparams := sig.RecvTypeParams() + if n := rtparams.Len(); index >= n { + return nil, fmt.Errorf("tuple index %d out of range [0-%d)", index, n) + } + t = rtparams.At(index) + case opConstraint: tparam, ok := t.(*types.TypeParam) if !ok { @@ -725,6 +756,10 @@ func Object(pkg *types.Package, p Path) (types.Object, error) { } } + if obj == nil { + panic(p) // path does not end in an object-valued operator + } + if obj.Pkg() != pkg { return nil, fmt.Errorf("path denotes %s, which belongs to a different package", obj) } diff --git a/vendor/golang.org/x/tools/internal/aliases/aliases_go121.go b/vendor/golang.org/x/tools/internal/aliases/aliases_go121.go index c027b9f315f68..6652f7db0fbd4 100644 --- a/vendor/golang.org/x/tools/internal/aliases/aliases_go121.go +++ b/vendor/golang.org/x/tools/internal/aliases/aliases_go121.go @@ -15,10 +15,14 @@ import ( // It will never be created by go/types. type Alias struct{} -func (*Alias) String() string { panic("unreachable") } -func (*Alias) Underlying() types.Type { panic("unreachable") } -func (*Alias) Obj() *types.TypeName { panic("unreachable") } -func Rhs(alias *Alias) types.Type { panic("unreachable") } +func (*Alias) String() string { panic("unreachable") } +func (*Alias) Underlying() types.Type { panic("unreachable") } +func (*Alias) Obj() *types.TypeName { panic("unreachable") } +func Rhs(alias *Alias) types.Type { panic("unreachable") } +func TypeParams(alias *Alias) *types.TypeParamList { panic("unreachable") } +func SetTypeParams(alias *Alias, tparams []*types.TypeParam) { panic("unreachable") } +func TypeArgs(alias *Alias) *types.TypeList { panic("unreachable") } +func Origin(alias *Alias) *Alias { panic("unreachable") } // Unalias returns the type t for go <=1.21. func Unalias(t types.Type) types.Type { return t } diff --git a/vendor/golang.org/x/tools/internal/aliases/aliases_go122.go b/vendor/golang.org/x/tools/internal/aliases/aliases_go122.go index b32995484190a..3ef1afeb403dd 100644 --- a/vendor/golang.org/x/tools/internal/aliases/aliases_go122.go +++ b/vendor/golang.org/x/tools/internal/aliases/aliases_go122.go @@ -28,6 +28,42 @@ func Rhs(alias *Alias) types.Type { return Unalias(alias) } +// TypeParams returns the type parameter list of the alias. +func TypeParams(alias *Alias) *types.TypeParamList { + if alias, ok := any(alias).(interface{ TypeParams() *types.TypeParamList }); ok { + return alias.TypeParams() // go1.23+ + } + return nil +} + +// SetTypeParams sets the type parameters of the alias type. +func SetTypeParams(alias *Alias, tparams []*types.TypeParam) { + if alias, ok := any(alias).(interface { + SetTypeParams(tparams []*types.TypeParam) + }); ok { + alias.SetTypeParams(tparams) // go1.23+ + } else if len(tparams) > 0 { + panic("cannot set type parameters of an Alias type in go1.22") + } +} + +// TypeArgs returns the type arguments used to instantiate the Alias type. +func TypeArgs(alias *Alias) *types.TypeList { + if alias, ok := any(alias).(interface{ TypeArgs() *types.TypeList }); ok { + return alias.TypeArgs() // go1.23+ + } + return nil // empty (go1.22) +} + +// Origin returns the generic Alias type of which alias is an instance. +// If alias is not an instance of a generic alias, Origin returns alias. +func Origin(alias *Alias) *Alias { + if alias, ok := any(alias).(interface{ Origin() *types.Alias }); ok { + return alias.Origin() // go1.23+ + } + return alias // not an instance of a generic alias (go1.22) +} + // Unalias is a wrapper of types.Unalias. func Unalias(t types.Type) types.Type { return types.Unalias(t) } diff --git a/vendor/golang.org/x/tools/internal/gocommand/invoke.go b/vendor/golang.org/x/tools/internal/gocommand/invoke.go index eb7a8282f9e7d..2e59ff8558c7e 100644 --- a/vendor/golang.org/x/tools/internal/gocommand/invoke.go +++ b/vendor/golang.org/x/tools/internal/gocommand/invoke.go @@ -8,12 +8,14 @@ package gocommand import ( "bytes" "context" + "encoding/json" "errors" "fmt" "io" "log" "os" "os/exec" + "path/filepath" "reflect" "regexp" "runtime" @@ -167,7 +169,9 @@ type Invocation struct { // TODO(rfindley): remove, in favor of Args. ModFile string - // If Overlay is set, the go command is invoked with -overlay=Overlay. + // Overlay is the name of the JSON overlay file that describes + // unsaved editor buffers; see [WriteOverlays]. + // If set, the go command is invoked with -overlay=Overlay. // TODO(rfindley): remove, in favor of Args. Overlay string @@ -196,12 +200,14 @@ func (i *Invocation) runWithFriendlyError(ctx context.Context, stdout, stderr io return } -func (i *Invocation) run(ctx context.Context, stdout, stderr io.Writer) error { - log := i.Logf - if log == nil { - log = func(string, ...interface{}) {} +// logf logs if i.Logf is non-nil. +func (i *Invocation) logf(format string, args ...any) { + if i.Logf != nil { + i.Logf(format, args...) } +} +func (i *Invocation) run(ctx context.Context, stdout, stderr io.Writer) error { goArgs := []string{i.Verb} appendModFile := func() { @@ -255,12 +261,15 @@ func (i *Invocation) run(ctx context.Context, stdout, stderr io.Writer) error { waitDelay.Set(reflect.ValueOf(30 * time.Second)) } - // On darwin the cwd gets resolved to the real path, which breaks anything that - // expects the working directory to keep the original path, including the + // The cwd gets resolved to the real path. On Darwin, where + // /tmp is a symlink, this breaks anything that expects the + // working directory to keep the original path, including the // go command when dealing with modules. - // The Go stdlib has a special feature where if the cwd and the PWD are the - // same node then it trusts the PWD, so by setting it in the env for the child - // process we fix up all the paths returned by the go command. + // + // os.Getwd has a special feature where if the cwd and the PWD + // are the same node then it trusts the PWD, so by setting it + // in the env for the child process we fix up all the paths + // returned by the go command. if !i.CleanEnv { cmd.Env = os.Environ() } @@ -270,7 +279,12 @@ func (i *Invocation) run(ctx context.Context, stdout, stderr io.Writer) error { cmd.Dir = i.WorkingDir } - defer func(start time.Time) { log("%s for %v", time.Since(start), cmdDebugStr(cmd)) }(time.Now()) + debugStr := cmdDebugStr(cmd) + i.logf("starting %v", debugStr) + start := time.Now() + defer func() { + i.logf("%s for %v", time.Since(start), debugStr) + }() return runCmdContext(ctx, cmd) } @@ -351,6 +365,7 @@ func runCmdContext(ctx context.Context, cmd *exec.Cmd) (err error) { } } + startTime := time.Now() err = cmd.Start() if stdoutW != nil { // The child process has inherited the pipe file, @@ -377,7 +392,7 @@ func runCmdContext(ctx context.Context, cmd *exec.Cmd) (err error) { case err := <-resChan: return err case <-timer.C: - HandleHangingGoCommand(cmd.Process) + HandleHangingGoCommand(startTime, cmd) case <-ctx.Done(): } } else { @@ -411,7 +426,7 @@ func runCmdContext(ctx context.Context, cmd *exec.Cmd) (err error) { return <-resChan } -func HandleHangingGoCommand(proc *os.Process) { +func HandleHangingGoCommand(start time.Time, cmd *exec.Cmd) { switch runtime.GOOS { case "linux", "darwin", "freebsd", "netbsd": fmt.Fprintln(os.Stderr, `DETECTED A HANGING GO COMMAND @@ -444,7 +459,7 @@ See golang/go#54461 for more details.`) panic(fmt.Sprintf("running %s: %v", listFiles, err)) } } - panic(fmt.Sprintf("detected hanging go command (pid %d): see golang/go#54461 for more details", proc.Pid)) + panic(fmt.Sprintf("detected hanging go command (golang/go#54461); waited %s\n\tcommand:%s\n\tpid:%d", time.Since(start), cmd, cmd.Process.Pid)) } func cmdDebugStr(cmd *exec.Cmd) string { @@ -468,3 +483,73 @@ func cmdDebugStr(cmd *exec.Cmd) string { } return fmt.Sprintf("GOROOT=%v GOPATH=%v GO111MODULE=%v GOPROXY=%v PWD=%v %v", env["GOROOT"], env["GOPATH"], env["GO111MODULE"], env["GOPROXY"], env["PWD"], strings.Join(args, " ")) } + +// WriteOverlays writes each value in the overlay (see the Overlay +// field of go/packages.Config) to a temporary file and returns the name +// of a JSON file describing the mapping that is suitable for the "go +// list -overlay" flag. +// +// On success, the caller must call the cleanup function exactly once +// when the files are no longer needed. +func WriteOverlays(overlay map[string][]byte) (filename string, cleanup func(), err error) { + // Do nothing if there are no overlays in the config. + if len(overlay) == 0 { + return "", func() {}, nil + } + + dir, err := os.MkdirTemp("", "gocommand-*") + if err != nil { + return "", nil, err + } + + // The caller must clean up this directory, + // unless this function returns an error. + // (The cleanup operand of each return + // statement below is ignored.) + defer func() { + cleanup = func() { + os.RemoveAll(dir) + } + if err != nil { + cleanup() + cleanup = nil + } + }() + + // Write each map entry to a temporary file. + overlays := make(map[string]string) + for k, v := range overlay { + // Use a unique basename for each file (001-foo.go), + // to avoid creating nested directories. + base := fmt.Sprintf("%d-%s", 1+len(overlays), filepath.Base(k)) + filename := filepath.Join(dir, base) + err := os.WriteFile(filename, v, 0666) + if err != nil { + return "", nil, err + } + overlays[k] = filename + } + + // Write the JSON overlay file that maps logical file names to temp files. + // + // OverlayJSON is the format overlay files are expected to be in. + // The Replace map maps from overlaid paths to replacement paths: + // the Go command will forward all reads trying to open + // each overlaid path to its replacement path, or consider the overlaid + // path not to exist if the replacement path is empty. + // + // From golang/go#39958. + type OverlayJSON struct { + Replace map[string]string `json:"replace,omitempty"` + } + b, err := json.Marshal(OverlayJSON{Replace: overlays}) + if err != nil { + return "", nil, err + } + filename = filepath.Join(dir, "overlay.json") + if err := os.WriteFile(filename, b, 0666); err != nil { + return "", nil, err + } + + return filename, nil, nil +} diff --git a/vendor/golang.org/x/tools/internal/imports/fix.go b/vendor/golang.org/x/tools/internal/imports/fix.go index 93d49a6efd0cd..dc7d50a7a4012 100644 --- a/vendor/golang.org/x/tools/internal/imports/fix.go +++ b/vendor/golang.org/x/tools/internal/imports/fix.go @@ -27,6 +27,7 @@ import ( "unicode" "unicode/utf8" + "golang.org/x/sync/errgroup" "golang.org/x/tools/go/ast/astutil" "golang.org/x/tools/internal/event" "golang.org/x/tools/internal/gocommand" @@ -104,7 +105,10 @@ type packageInfo struct { // parseOtherFiles parses all the Go files in srcDir except filename, including // test files if filename looks like a test. -func parseOtherFiles(fset *token.FileSet, srcDir, filename string) []*ast.File { +// +// It returns an error only if ctx is cancelled. Files with parse errors are +// ignored. +func parseOtherFiles(ctx context.Context, fset *token.FileSet, srcDir, filename string) ([]*ast.File, error) { // This could use go/packages but it doesn't buy much, and it fails // with https://golang.org/issue/26296 in LoadFiles mode in some cases. considerTests := strings.HasSuffix(filename, "_test.go") @@ -112,11 +116,14 @@ func parseOtherFiles(fset *token.FileSet, srcDir, filename string) []*ast.File { fileBase := filepath.Base(filename) packageFileInfos, err := os.ReadDir(srcDir) if err != nil { - return nil + return nil, ctx.Err() } var files []*ast.File for _, fi := range packageFileInfos { + if ctx.Err() != nil { + return nil, ctx.Err() + } if fi.Name() == fileBase || !strings.HasSuffix(fi.Name(), ".go") { continue } @@ -132,7 +139,7 @@ func parseOtherFiles(fset *token.FileSet, srcDir, filename string) []*ast.File { files = append(files, f) } - return files + return files, ctx.Err() } // addGlobals puts the names of package vars into the provided map. @@ -359,9 +366,7 @@ func (p *pass) load() ([]*ImportFix, bool) { if p.loadRealPackageNames { err := p.loadPackageNames(append(imports, p.candidates...)) if err != nil { - if p.env.Logf != nil { - p.env.Logf("loading package names: %v", err) - } + p.env.logf("loading package names: %v", err) return nil, false } } @@ -560,6 +565,8 @@ func (p *pass) addCandidate(imp *ImportInfo, pkg *packageInfo) { // // This is declared as a variable rather than a function so goimports can // easily be extended by adding a file with an init function. +// +// DO NOT REMOVE: used internally at Google. var fixImports = fixImportsDefault func fixImportsDefault(fset *token.FileSet, f *ast.File, filename string, env *ProcessEnv) error { @@ -579,9 +586,7 @@ func getFixes(ctx context.Context, fset *token.FileSet, f *ast.File, filename st return nil, err } srcDir := filepath.Dir(abs) - if env.Logf != nil { - env.Logf("fixImports(filename=%q), abs=%q, srcDir=%q ...", filename, abs, srcDir) - } + env.logf("fixImports(filename=%q), abs=%q, srcDir=%q ...", filename, abs, srcDir) // First pass: looking only at f, and using the naive algorithm to // derive package names from import paths, see if the file is already @@ -592,7 +597,10 @@ func getFixes(ctx context.Context, fset *token.FileSet, f *ast.File, filename st return fixes, nil } - otherFiles := parseOtherFiles(fset, srcDir, filename) + otherFiles, err := parseOtherFiles(ctx, fset, srcDir, filename) + if err != nil { + return nil, err + } // Second pass: add information from other files in the same package, // like their package vars and imports. @@ -1010,16 +1018,26 @@ func (e *ProcessEnv) GetResolver() (Resolver, error) { // already know the view type. if len(e.Env["GOMOD"]) == 0 && len(e.Env["GOWORK"]) == 0 { e.resolver = newGopathResolver(e) + e.logf("created gopath resolver") } else if r, err := newModuleResolver(e, e.ModCache); err != nil { e.resolverErr = err + e.logf("failed to create module resolver: %v", err) } else { e.resolver = Resolver(r) + e.logf("created module resolver") } } return e.resolver, e.resolverErr } +// logf logs if e.Logf is non-nil. +func (e *ProcessEnv) logf(format string, args ...any) { + if e.Logf != nil { + e.Logf(format, args...) + } +} + // buildContext returns the build.Context to use for matching files. // // TODO(rfindley): support dynamic GOOS, GOARCH here, when doing cross-platform @@ -1123,8 +1141,8 @@ type Resolver interface { // scan works with callback to search for packages. See scanCallback for details. scan(ctx context.Context, callback *scanCallback) error - // loadExports returns the set of exported symbols in the package at dir. - // loadExports may be called concurrently. + // loadExports returns the package name and set of exported symbols in the + // package at dir. loadExports may be called concurrently. loadExports(ctx context.Context, pkg *pkg, includeTest bool) (string, []stdlib.Symbol, error) // scoreImportPath returns the relevance for an import path. @@ -1192,7 +1210,7 @@ func addExternalCandidates(ctx context.Context, pass *pass, refs references, fil if err != nil { return err } - if err = resolver.scan(context.Background(), callback); err != nil { + if err = resolver.scan(ctx, callback); err != nil { return err } @@ -1201,54 +1219,52 @@ func addExternalCandidates(ctx context.Context, pass *pass, refs references, fil imp *ImportInfo pkg *packageInfo } - results := make(chan result, len(refs)) + results := make([]*result, len(refs)) - ctx, cancel := context.WithCancel(context.TODO()) - var wg sync.WaitGroup - defer func() { - cancel() - wg.Wait() - }() - var ( - firstErr error - firstErrOnce sync.Once - ) - for pkgName, symbols := range refs { - wg.Add(1) - go func(pkgName string, symbols map[string]bool) { - defer wg.Done() + g, ctx := errgroup.WithContext(ctx) + + searcher := symbolSearcher{ + logf: pass.env.logf, + srcDir: pass.srcDir, + xtest: strings.HasSuffix(pass.f.Name.Name, "_test"), + loadExports: resolver.loadExports, + } - found, err := findImport(ctx, pass, found[pkgName], pkgName, symbols) + i := 0 + for pkgName, symbols := range refs { + index := i // claim an index in results + i++ + pkgName := pkgName + symbols := symbols + g.Go(func() error { + found, err := searcher.search(ctx, found[pkgName], pkgName, symbols) if err != nil { - firstErrOnce.Do(func() { - firstErr = err - cancel() - }) - return + return err } - if found == nil { - return // No matching package. + return nil // No matching package. } imp := &ImportInfo{ ImportPath: found.importPathShort, } - pkg := &packageInfo{ name: pkgName, exports: symbols, } - results <- result{imp, pkg} - }(pkgName, symbols) + results[index] = &result{imp, pkg} + return nil + }) + } + if err := g.Wait(); err != nil { + return err } - go func() { - wg.Wait() - close(results) - }() - for result := range results { + for _, result := range results { + if result == nil { + continue + } // Don't offer completions that would shadow predeclared // names, such as github.com/coreos/etcd/error. if types.Universe.Lookup(result.pkg.name) != nil { // predeclared @@ -1262,7 +1278,7 @@ func addExternalCandidates(ctx context.Context, pass *pass, refs references, fil } pass.addCandidate(result.imp, result.pkg) } - return firstErr + return nil } // notIdentifier reports whether ch is an invalid identifier character. @@ -1606,9 +1622,7 @@ func loadExportsFromFiles(ctx context.Context, env *ProcessEnv, dir string, incl fullFile := filepath.Join(dir, fi.Name()) f, err := parser.ParseFile(fset, fullFile, nil, 0) if err != nil { - if env.Logf != nil { - env.Logf("error parsing %v: %v", fullFile, err) - } + env.logf("error parsing %v: %v", fullFile, err) continue } if f.Name.Name == "documentation" { @@ -1644,9 +1658,7 @@ func loadExportsFromFiles(ctx context.Context, env *ProcessEnv, dir string, incl } sortSymbols(exports) - if env.Logf != nil { - env.Logf("loaded exports in dir %v (package %v): %v", dir, pkgName, exports) - } + env.logf("loaded exports in dir %v (package %v): %v", dir, pkgName, exports) return pkgName, exports, nil } @@ -1656,25 +1668,39 @@ func sortSymbols(syms []stdlib.Symbol) { }) } -// findImport searches for a package with the given symbols. -// If no package is found, findImport returns ("", false, nil) -func findImport(ctx context.Context, pass *pass, candidates []pkgDistance, pkgName string, symbols map[string]bool) (*pkg, error) { +// A symbolSearcher searches for a package with a set of symbols, among a set +// of candidates. See [symbolSearcher.search]. +// +// The search occurs within the scope of a single file, with context captured +// in srcDir and xtest. +type symbolSearcher struct { + logf func(string, ...any) + srcDir string // directory containing the file + xtest bool // if set, the file containing is an x_test file + loadExports func(ctx context.Context, pkg *pkg, includeTest bool) (string, []stdlib.Symbol, error) +} + +// search searches the provided candidates for a package containing all +// exported symbols. +// +// If successful, returns the resulting package. +func (s *symbolSearcher) search(ctx context.Context, candidates []pkgDistance, pkgName string, symbols map[string]bool) (*pkg, error) { // Sort the candidates by their import package length, // assuming that shorter package names are better than long // ones. Note that this sorts by the de-vendored name, so // there's no "penalty" for vendoring. sort.Sort(byDistanceOrImportPathShortLength(candidates)) - if pass.env.Logf != nil { + if s.logf != nil { for i, c := range candidates { - pass.env.Logf("%s candidate %d/%d: %v in %v", pkgName, i+1, len(candidates), c.pkg.importPathShort, c.pkg.dir) + s.logf("%s candidate %d/%d: %v in %v", pkgName, i+1, len(candidates), c.pkg.importPathShort, c.pkg.dir) } } - resolver, err := pass.env.GetResolver() - if err != nil { - return nil, err - } - // Collect exports for packages with matching names. + // Arrange rescv so that we can we can await results in order of relevance + // and exit as soon as we find the first match. + // + // Search with bounded concurrency, returning as soon as the first result + // among rescv is non-nil. rescv := make([]chan *pkg, len(candidates)) for i := range candidates { rescv[i] = make(chan *pkg, 1) @@ -1682,6 +1708,7 @@ func findImport(ctx context.Context, pass *pass, candidates []pkgDistance, pkgNa const maxConcurrentPackageImport = 4 loadExportsSem := make(chan struct{}, maxConcurrentPackageImport) + // Ensure that all work is completed at exit. ctx, cancel := context.WithCancel(ctx) var wg sync.WaitGroup defer func() { @@ -1689,6 +1716,7 @@ func findImport(ctx context.Context, pass *pass, candidates []pkgDistance, pkgNa wg.Wait() }() + // Start the search. wg.Add(1) go func() { defer wg.Done() @@ -1699,55 +1727,67 @@ func findImport(ctx context.Context, pass *pass, candidates []pkgDistance, pkgNa return } + i := i + c := c wg.Add(1) - go func(c pkgDistance, resc chan<- *pkg) { + go func() { defer func() { <-loadExportsSem wg.Done() }() - - if pass.env.Logf != nil { - pass.env.Logf("loading exports in dir %s (seeking package %s)", c.pkg.dir, pkgName) + if s.logf != nil { + s.logf("loading exports in dir %s (seeking package %s)", c.pkg.dir, pkgName) } - // If we're an x_test, load the package under test's test variant. - includeTest := strings.HasSuffix(pass.f.Name.Name, "_test") && c.pkg.dir == pass.srcDir - _, exports, err := resolver.loadExports(ctx, c.pkg, includeTest) + pkg, err := s.searchOne(ctx, c, symbols) if err != nil { - if pass.env.Logf != nil { - pass.env.Logf("loading exports in dir %s (seeking package %s): %v", c.pkg.dir, pkgName, err) - } - resc <- nil - return - } - - exportsMap := make(map[string]bool, len(exports)) - for _, sym := range exports { - exportsMap[sym.Name] = true - } - - // If it doesn't have the right - // symbols, send nil to mean no match. - for symbol := range symbols { - if !exportsMap[symbol] { - resc <- nil - return + if s.logf != nil && ctx.Err() == nil { + s.logf("loading exports in dir %s (seeking package %s): %v", c.pkg.dir, pkgName, err) } + pkg = nil } - resc <- c.pkg - }(c, rescv[i]) + rescv[i] <- pkg // may be nil + }() } }() + // Await the first (best) result. for _, resc := range rescv { - pkg := <-resc - if pkg == nil { - continue + select { + case r := <-resc: + if r != nil { + return r, nil + } + case <-ctx.Done(): + return nil, ctx.Err() } - return pkg, nil } return nil, nil } +func (s *symbolSearcher) searchOne(ctx context.Context, c pkgDistance, symbols map[string]bool) (*pkg, error) { + if ctx.Err() != nil { + return nil, ctx.Err() + } + // If we're considering the package under test from an x_test, load the + // test variant. + includeTest := s.xtest && c.pkg.dir == s.srcDir + _, exports, err := s.loadExports(ctx, c.pkg, includeTest) + if err != nil { + return nil, err + } + + exportsMap := make(map[string]bool, len(exports)) + for _, sym := range exports { + exportsMap[sym.Name] = true + } + for symbol := range symbols { + if !exportsMap[symbol] { + return nil, nil // no match + } + } + return c.pkg, nil +} + // pkgIsCandidate reports whether pkg is a candidate for satisfying the // finding which package pkgIdent in the file named by filename is trying // to refer to. @@ -1767,58 +1807,24 @@ func pkgIsCandidate(filename string, refs references, pkg *pkg) bool { } // Speed optimization to minimize disk I/O: - // the last two components on disk must contain the - // package name somewhere. // - // This permits mismatch naming like directory - // "go-foo" being package "foo", or "pkg.v3" being "pkg", - // or directory "google.golang.org/api/cloudbilling/v1" - // being package "cloudbilling", but doesn't - // permit a directory "foo" to be package - // "bar", which is strongly discouraged - // anyway. There's no reason goimports needs - // to be slow just to accommodate that. + // Use the matchesPath heuristic to filter to package paths that could + // reasonably match a dangling reference. + // + // This permits mismatch naming like directory "go-foo" being package "foo", + // or "pkg.v3" being "pkg", or directory + // "google.golang.org/api/cloudbilling/v1" being package "cloudbilling", but + // doesn't permit a directory "foo" to be package "bar", which is strongly + // discouraged anyway. There's no reason goimports needs to be slow just to + // accommodate that. for pkgIdent := range refs { - lastTwo := lastTwoComponents(pkg.importPathShort) - if strings.Contains(lastTwo, pkgIdent) { + if matchesPath(pkgIdent, pkg.importPathShort) { return true } - if hasHyphenOrUpperASCII(lastTwo) && !hasHyphenOrUpperASCII(pkgIdent) { - lastTwo = lowerASCIIAndRemoveHyphen(lastTwo) - if strings.Contains(lastTwo, pkgIdent) { - return true - } - } } return false } -func hasHyphenOrUpperASCII(s string) bool { - for i := 0; i < len(s); i++ { - b := s[i] - if b == '-' || ('A' <= b && b <= 'Z') { - return true - } - } - return false -} - -func lowerASCIIAndRemoveHyphen(s string) (ret string) { - buf := make([]byte, 0, len(s)) - for i := 0; i < len(s); i++ { - b := s[i] - switch { - case b == '-': - continue - case 'A' <= b && b <= 'Z': - buf = append(buf, b+('a'-'A')) - default: - buf = append(buf, b) - } - } - return string(buf) -} - // canUse reports whether the package in dir is usable from filename, // respecting the Go "internal" and "vendor" visibility rules. func canUse(filename, dir string) bool { @@ -1859,19 +1865,84 @@ func canUse(filename, dir string) bool { return !strings.Contains(relSlash, "/vendor/") && !strings.Contains(relSlash, "/internal/") && !strings.HasSuffix(relSlash, "/internal") } -// lastTwoComponents returns at most the last two path components -// of v, using either / or \ as the path separator. -func lastTwoComponents(v string) string { +// matchesPath reports whether ident may match a potential package name +// referred to by path, using heuristics to filter out unidiomatic package +// names. +// +// Specifically, it checks whether either of the last two '/'- or '\'-delimited +// path segments matches the identifier. The segment-matching heuristic must +// allow for various conventions around segment naming, including go-foo, +// foo-go, and foo.v3. To handle all of these, matching considers both (1) the +// entire segment, ignoring '-' and '.', as well as (2) the last subsegment +// separated by '-' or '.'. So the segment foo-go matches all of the following +// identifiers: foo, go, and foogo. All matches are case insensitive (for ASCII +// identifiers). +// +// See the docstring for [pkgIsCandidate] for an explanation of how this +// heuristic filters potential candidate packages. +func matchesPath(ident, path string) bool { + // Ignore case, for ASCII. + lowerIfASCII := func(b byte) byte { + if 'A' <= b && b <= 'Z' { + return b + ('a' - 'A') + } + return b + } + + // match reports whether path[start:end] matches ident, ignoring [.-]. + match := func(start, end int) bool { + ii := len(ident) - 1 // current byte in ident + pi := end - 1 // current byte in path + for ; pi >= start && ii >= 0; pi-- { + pb := path[pi] + if pb == '-' || pb == '.' { + continue + } + pb = lowerIfASCII(pb) + ib := lowerIfASCII(ident[ii]) + if pb != ib { + return false + } + ii-- + } + return ii < 0 && pi < start // all bytes matched + } + + // segmentEnd and subsegmentEnd hold the end points of the current segment + // and subsegment intervals. + segmentEnd := len(path) + subsegmentEnd := len(path) + + // Count slashes; we only care about the last two segments. nslash := 0 - for i := len(v) - 1; i >= 0; i-- { - if v[i] == '/' || v[i] == '\\' { + + for i := len(path) - 1; i >= 0; i-- { + switch b := path[i]; b { + // TODO(rfindley): we handle backlashes here only because the previous + // heuristic handled backslashes. This is perhaps overly defensive, but is + // the result of many lessons regarding Chesterton's fence and the + // goimports codebase. + // + // However, this function is only ever called with something called an + // 'importPath'. Is it possible that this is a real import path, and + // therefore we need only consider forward slashes? + case '/', '\\': + if match(i+1, segmentEnd) || match(i+1, subsegmentEnd) { + return true + } nslash++ if nslash == 2 { - return v[i:] + return false // did not match above + } + segmentEnd, subsegmentEnd = i, i // reset + case '-', '.': + if match(i+1, subsegmentEnd) { + return true } + subsegmentEnd = i } } - return v + return match(0, segmentEnd) || match(0, subsegmentEnd) } type visitFn func(node ast.Node) ast.Visitor diff --git a/vendor/golang.org/x/tools/internal/imports/mod.go b/vendor/golang.org/x/tools/internal/imports/mod.go index 82fe644a189b4..91221fda32261 100644 --- a/vendor/golang.org/x/tools/internal/imports/mod.go +++ b/vendor/golang.org/x/tools/internal/imports/mod.go @@ -265,9 +265,7 @@ func (r *ModuleResolver) initAllMods() error { return err } if mod.Dir == "" { - if r.env.Logf != nil { - r.env.Logf("module %v has not been downloaded and will be ignored", mod.Path) - } + r.env.logf("module %v has not been downloaded and will be ignored", mod.Path) // Can't do anything with a module that's not downloaded. continue } @@ -766,9 +764,7 @@ func (r *ModuleResolver) scanDirForPackage(root gopathwalk.Root, dir string) dir } modPath, err := module.UnescapePath(filepath.ToSlash(matches[1])) if err != nil { - if r.env.Logf != nil { - r.env.Logf("decoding module cache path %q: %v", subdir, err) - } + r.env.logf("decoding module cache path %q: %v", subdir, err) return directoryPackageInfo{ status: directoryScanned, err: fmt.Errorf("decoding module cache path %q: %v", subdir, err), diff --git a/vendor/golang.org/x/tools/internal/pkgbits/decoder.go b/vendor/golang.org/x/tools/internal/pkgbits/decoder.go index 2acd85851e362..b92e8e6eb3299 100644 --- a/vendor/golang.org/x/tools/internal/pkgbits/decoder.go +++ b/vendor/golang.org/x/tools/internal/pkgbits/decoder.go @@ -23,9 +23,6 @@ type PkgDecoder struct { // version is the file format version. version uint32 - // aliases determines whether types.Aliases should be created - aliases bool - // sync indicates whether the file uses sync markers. sync bool @@ -76,7 +73,6 @@ func (pr *PkgDecoder) SyncMarkers() bool { return pr.sync } func NewPkgDecoder(pkgPath, input string) PkgDecoder { pr := PkgDecoder{ pkgPath: pkgPath, - //aliases: aliases.Enabled(), } // TODO(mdempsky): Implement direct indexing of input string to diff --git a/vendor/golang.org/x/tools/internal/stdlib/manifest.go b/vendor/golang.org/x/tools/internal/stdlib/manifest.go index fd6892075ee4b..a928acf29facf 100644 --- a/vendor/golang.org/x/tools/internal/stdlib/manifest.go +++ b/vendor/golang.org/x/tools/internal/stdlib/manifest.go @@ -23,6 +23,7 @@ var PackageSymbols = map[string][]Symbol{ {"ErrWriteAfterClose", Var, 0}, {"ErrWriteTooLong", Var, 0}, {"FileInfoHeader", Func, 1}, + {"FileInfoNames", Type, 23}, {"Format", Type, 10}, {"FormatGNU", Const, 10}, {"FormatPAX", Const, 10}, @@ -820,6 +821,7 @@ var PackageSymbols = map[string][]Symbol{ {"(*ConnectionState).ExportKeyingMaterial", Method, 11}, {"(*Dialer).Dial", Method, 15}, {"(*Dialer).DialContext", Method, 15}, + {"(*ECHRejectionError).Error", Method, 23}, {"(*QUICConn).Close", Method, 21}, {"(*QUICConn).ConnectionState", Method, 21}, {"(*QUICConn).HandleData", Method, 21}, @@ -827,6 +829,7 @@ var PackageSymbols = map[string][]Symbol{ {"(*QUICConn).SendSessionTicket", Method, 21}, {"(*QUICConn).SetTransportParameters", Method, 21}, {"(*QUICConn).Start", Method, 21}, + {"(*QUICConn).StoreSession", Method, 23}, {"(*SessionState).Bytes", Method, 21}, {"(AlertError).Error", Method, 21}, {"(ClientAuthType).String", Method, 15}, @@ -877,6 +880,8 @@ var PackageSymbols = map[string][]Symbol{ {"Config.ClientSessionCache", Field, 3}, {"Config.CurvePreferences", Field, 3}, {"Config.DynamicRecordSizingDisabled", Field, 7}, + {"Config.EncryptedClientHelloConfigList", Field, 23}, + {"Config.EncryptedClientHelloRejectionVerify", Field, 23}, {"Config.GetCertificate", Field, 4}, {"Config.GetClientCertificate", Field, 8}, {"Config.GetConfigForClient", Field, 8}, @@ -902,6 +907,7 @@ var PackageSymbols = map[string][]Symbol{ {"ConnectionState", Type, 0}, {"ConnectionState.CipherSuite", Field, 0}, {"ConnectionState.DidResume", Field, 1}, + {"ConnectionState.ECHAccepted", Field, 23}, {"ConnectionState.HandshakeComplete", Field, 0}, {"ConnectionState.NegotiatedProtocol", Field, 0}, {"ConnectionState.NegotiatedProtocolIsMutual", Field, 0}, @@ -925,6 +931,8 @@ var PackageSymbols = map[string][]Symbol{ {"ECDSAWithP384AndSHA384", Const, 8}, {"ECDSAWithP521AndSHA512", Const, 8}, {"ECDSAWithSHA1", Const, 10}, + {"ECHRejectionError", Type, 23}, + {"ECHRejectionError.RetryConfigList", Field, 23}, {"Ed25519", Const, 13}, {"InsecureCipherSuites", Func, 14}, {"Listen", Func, 0}, @@ -943,6 +951,7 @@ var PackageSymbols = map[string][]Symbol{ {"ParseSessionState", Func, 21}, {"QUICClient", Func, 21}, {"QUICConfig", Type, 21}, + {"QUICConfig.EnableStoreSessionEvent", Field, 23}, {"QUICConfig.TLSConfig", Field, 21}, {"QUICConn", Type, 21}, {"QUICEncryptionLevel", Type, 21}, @@ -954,16 +963,20 @@ var PackageSymbols = map[string][]Symbol{ {"QUICEvent.Data", Field, 21}, {"QUICEvent.Kind", Field, 21}, {"QUICEvent.Level", Field, 21}, + {"QUICEvent.SessionState", Field, 23}, {"QUICEvent.Suite", Field, 21}, {"QUICEventKind", Type, 21}, {"QUICHandshakeDone", Const, 21}, {"QUICNoEvent", Const, 21}, {"QUICRejectedEarlyData", Const, 21}, + {"QUICResumeSession", Const, 23}, {"QUICServer", Func, 21}, {"QUICSessionTicketOptions", Type, 21}, {"QUICSessionTicketOptions.EarlyData", Field, 21}, + {"QUICSessionTicketOptions.Extra", Field, 23}, {"QUICSetReadSecret", Const, 21}, {"QUICSetWriteSecret", Const, 21}, + {"QUICStoreSession", Const, 23}, {"QUICTransportParameters", Const, 21}, {"QUICTransportParametersRequired", Const, 21}, {"QUICWriteData", Const, 21}, @@ -1036,6 +1049,8 @@ var PackageSymbols = map[string][]Symbol{ {"(*Certificate).Verify", Method, 0}, {"(*Certificate).VerifyHostname", Method, 0}, {"(*CertificateRequest).CheckSignature", Method, 5}, + {"(*OID).UnmarshalBinary", Method, 23}, + {"(*OID).UnmarshalText", Method, 23}, {"(*RevocationList).CheckSignatureFrom", Method, 19}, {"(CertificateInvalidError).Error", Method, 0}, {"(ConstraintViolationError).Error", Method, 0}, @@ -1043,6 +1058,8 @@ var PackageSymbols = map[string][]Symbol{ {"(InsecureAlgorithmError).Error", Method, 6}, {"(OID).Equal", Method, 22}, {"(OID).EqualASN1OID", Method, 22}, + {"(OID).MarshalBinary", Method, 23}, + {"(OID).MarshalText", Method, 23}, {"(OID).String", Method, 22}, {"(PublicKeyAlgorithm).String", Method, 10}, {"(SignatureAlgorithm).String", Method, 6}, @@ -1196,6 +1213,7 @@ var PackageSymbols = map[string][]Symbol{ {"ParseCertificates", Func, 0}, {"ParseDERCRL", Func, 0}, {"ParseECPrivateKey", Func, 1}, + {"ParseOID", Func, 23}, {"ParsePKCS1PrivateKey", Func, 0}, {"ParsePKCS1PublicKey", Func, 10}, {"ParsePKCS8PrivateKey", Func, 0}, @@ -2541,6 +2559,7 @@ var PackageSymbols = map[string][]Symbol{ {"PT_NOTE", Const, 0}, {"PT_NULL", Const, 0}, {"PT_OPENBSD_BOOTDATA", Const, 16}, + {"PT_OPENBSD_NOBTCFI", Const, 23}, {"PT_OPENBSD_RANDOMIZE", Const, 16}, {"PT_OPENBSD_WXNEEDED", Const, 16}, {"PT_PAX_FLAGS", Const, 16}, @@ -3620,13 +3639,16 @@ var PackageSymbols = map[string][]Symbol{ {"STT_COMMON", Const, 0}, {"STT_FILE", Const, 0}, {"STT_FUNC", Const, 0}, + {"STT_GNU_IFUNC", Const, 23}, {"STT_HIOS", Const, 0}, {"STT_HIPROC", Const, 0}, {"STT_LOOS", Const, 0}, {"STT_LOPROC", Const, 0}, {"STT_NOTYPE", Const, 0}, {"STT_OBJECT", Const, 0}, + {"STT_RELC", Const, 23}, {"STT_SECTION", Const, 0}, + {"STT_SRELC", Const, 23}, {"STT_TLS", Const, 0}, {"STV_DEFAULT", Const, 0}, {"STV_HIDDEN", Const, 0}, @@ -4544,11 +4566,14 @@ var PackageSymbols = map[string][]Symbol{ {"URLEncoding", Var, 0}, }, "encoding/binary": { + {"Append", Func, 23}, {"AppendByteOrder", Type, 19}, {"AppendUvarint", Func, 19}, {"AppendVarint", Func, 19}, {"BigEndian", Var, 0}, {"ByteOrder", Type, 0}, + {"Decode", Func, 23}, + {"Encode", Func, 23}, {"LittleEndian", Var, 0}, {"MaxVarintLen16", Const, 0}, {"MaxVarintLen32", Const, 0}, @@ -5308,6 +5333,7 @@ var PackageSymbols = map[string][]Symbol{ {"ParenExpr.Rparen", Field, 0}, {"ParenExpr.X", Field, 0}, {"Pkg", Const, 0}, + {"Preorder", Func, 23}, {"Print", Func, 0}, {"RECV", Const, 0}, {"RangeStmt", Type, 0}, @@ -5898,7 +5924,12 @@ var PackageSymbols = map[string][]Symbol{ }, "go/types": { {"(*Alias).Obj", Method, 22}, + {"(*Alias).Origin", Method, 23}, + {"(*Alias).Rhs", Method, 23}, + {"(*Alias).SetTypeParams", Method, 23}, {"(*Alias).String", Method, 22}, + {"(*Alias).TypeArgs", Method, 23}, + {"(*Alias).TypeParams", Method, 23}, {"(*Alias).Underlying", Method, 22}, {"(*ArgumentError).Error", Method, 18}, {"(*ArgumentError).Unwrap", Method, 18}, @@ -5943,6 +5974,7 @@ var PackageSymbols = map[string][]Symbol{ {"(*Func).Pkg", Method, 5}, {"(*Func).Pos", Method, 5}, {"(*Func).Scope", Method, 5}, + {"(*Func).Signature", Method, 23}, {"(*Func).String", Method, 5}, {"(*Func).Type", Method, 5}, {"(*Info).ObjectOf", Method, 5}, @@ -6992,6 +7024,12 @@ var PackageSymbols = map[string][]Symbol{ {"TempFile", Func, 0}, {"WriteFile", Func, 0}, }, + "iter": { + {"Pull", Func, 23}, + {"Pull2", Func, 23}, + {"Seq", Type, 23}, + {"Seq2", Type, 23}, + }, "log": { {"(*Logger).Fatal", Method, 0}, {"(*Logger).Fatalf", Method, 0}, @@ -7222,11 +7260,16 @@ var PackageSymbols = map[string][]Symbol{ {"Writer", Type, 0}, }, "maps": { + {"All", Func, 23}, {"Clone", Func, 21}, + {"Collect", Func, 23}, {"Copy", Func, 21}, {"DeleteFunc", Func, 21}, {"Equal", Func, 21}, {"EqualFunc", Func, 21}, + {"Insert", Func, 23}, + {"Keys", Func, 23}, + {"Values", Func, 23}, }, "math": { {"Abs", Func, 0}, @@ -7617,6 +7660,7 @@ var PackageSymbols = map[string][]Symbol{ }, "math/rand/v2": { {"(*ChaCha8).MarshalBinary", Method, 22}, + {"(*ChaCha8).Read", Method, 23}, {"(*ChaCha8).Seed", Method, 22}, {"(*ChaCha8).Uint64", Method, 22}, {"(*ChaCha8).UnmarshalBinary", Method, 22}, @@ -7636,6 +7680,7 @@ var PackageSymbols = map[string][]Symbol{ {"(*Rand).NormFloat64", Method, 22}, {"(*Rand).Perm", Method, 22}, {"(*Rand).Shuffle", Method, 22}, + {"(*Rand).Uint", Method, 23}, {"(*Rand).Uint32", Method, 22}, {"(*Rand).Uint32N", Method, 22}, {"(*Rand).Uint64", Method, 22}, @@ -7663,6 +7708,7 @@ var PackageSymbols = map[string][]Symbol{ {"Rand", Type, 22}, {"Shuffle", Func, 22}, {"Source", Type, 22}, + {"Uint", Func, 23}, {"Uint32", Func, 22}, {"Uint32N", Func, 22}, {"Uint64", Func, 22}, @@ -7743,6 +7789,7 @@ var PackageSymbols = map[string][]Symbol{ {"(*DNSError).Error", Method, 0}, {"(*DNSError).Temporary", Method, 0}, {"(*DNSError).Timeout", Method, 0}, + {"(*DNSError).Unwrap", Method, 23}, {"(*Dialer).Dial", Method, 1}, {"(*Dialer).DialContext", Method, 7}, {"(*Dialer).MultipathTCP", Method, 21}, @@ -7809,6 +7856,7 @@ var PackageSymbols = map[string][]Symbol{ {"(*TCPConn).RemoteAddr", Method, 0}, {"(*TCPConn).SetDeadline", Method, 0}, {"(*TCPConn).SetKeepAlive", Method, 0}, + {"(*TCPConn).SetKeepAliveConfig", Method, 23}, {"(*TCPConn).SetKeepAlivePeriod", Method, 2}, {"(*TCPConn).SetLinger", Method, 0}, {"(*TCPConn).SetNoDelay", Method, 0}, @@ -7922,6 +7970,7 @@ var PackageSymbols = map[string][]Symbol{ {"DNSError.IsTimeout", Field, 0}, {"DNSError.Name", Field, 0}, {"DNSError.Server", Field, 0}, + {"DNSError.UnwrapErr", Field, 23}, {"DefaultResolver", Var, 8}, {"Dial", Func, 0}, {"DialIP", Func, 0}, @@ -7937,6 +7986,7 @@ var PackageSymbols = map[string][]Symbol{ {"Dialer.DualStack", Field, 2}, {"Dialer.FallbackDelay", Field, 5}, {"Dialer.KeepAlive", Field, 3}, + {"Dialer.KeepAliveConfig", Field, 23}, {"Dialer.LocalAddr", Field, 1}, {"Dialer.Resolver", Field, 8}, {"Dialer.Timeout", Field, 1}, @@ -7989,10 +8039,16 @@ var PackageSymbols = map[string][]Symbol{ {"Interfaces", Func, 0}, {"InvalidAddrError", Type, 0}, {"JoinHostPort", Func, 0}, + {"KeepAliveConfig", Type, 23}, + {"KeepAliveConfig.Count", Field, 23}, + {"KeepAliveConfig.Enable", Field, 23}, + {"KeepAliveConfig.Idle", Field, 23}, + {"KeepAliveConfig.Interval", Field, 23}, {"Listen", Func, 0}, {"ListenConfig", Type, 11}, {"ListenConfig.Control", Field, 11}, {"ListenConfig.KeepAlive", Field, 13}, + {"ListenConfig.KeepAliveConfig", Field, 23}, {"ListenIP", Func, 0}, {"ListenMulticastUDP", Func, 0}, {"ListenPacket", Func, 0}, @@ -8081,6 +8137,7 @@ var PackageSymbols = map[string][]Symbol{ {"(*Request).Context", Method, 7}, {"(*Request).Cookie", Method, 0}, {"(*Request).Cookies", Method, 0}, + {"(*Request).CookiesNamed", Method, 23}, {"(*Request).FormFile", Method, 0}, {"(*Request).FormValue", Method, 0}, {"(*Request).MultipartReader", Method, 0}, @@ -8148,7 +8205,9 @@ var PackageSymbols = map[string][]Symbol{ {"Cookie.HttpOnly", Field, 0}, {"Cookie.MaxAge", Field, 0}, {"Cookie.Name", Field, 0}, + {"Cookie.Partitioned", Field, 23}, {"Cookie.Path", Field, 0}, + {"Cookie.Quoted", Field, 23}, {"Cookie.Raw", Field, 0}, {"Cookie.RawExpires", Field, 0}, {"Cookie.SameSite", Field, 11}, @@ -8225,7 +8284,9 @@ var PackageSymbols = map[string][]Symbol{ {"NoBody", Var, 8}, {"NotFound", Func, 0}, {"NotFoundHandler", Func, 0}, + {"ParseCookie", Func, 23}, {"ParseHTTPVersion", Func, 0}, + {"ParseSetCookie", Func, 23}, {"ParseTime", Func, 1}, {"Post", Func, 0}, {"PostForm", Func, 0}, @@ -8252,6 +8313,7 @@ var PackageSymbols = map[string][]Symbol{ {"Request.Host", Field, 0}, {"Request.Method", Field, 0}, {"Request.MultipartForm", Field, 0}, + {"Request.Pattern", Field, 23}, {"Request.PostForm", Field, 1}, {"Request.Proto", Field, 0}, {"Request.ProtoMajor", Field, 0}, @@ -8453,6 +8515,7 @@ var PackageSymbols = map[string][]Symbol{ {"DefaultRemoteAddr", Const, 0}, {"NewRecorder", Func, 0}, {"NewRequest", Func, 7}, + {"NewRequestWithContext", Func, 23}, {"NewServer", Func, 0}, {"NewTLSServer", Func, 0}, {"NewUnstartedServer", Func, 0}, @@ -8917,6 +8980,7 @@ var PackageSymbols = map[string][]Symbol{ {"Chown", Func, 0}, {"Chtimes", Func, 0}, {"Clearenv", Func, 0}, + {"CopyFS", Func, 23}, {"Create", Func, 0}, {"CreateTemp", Func, 16}, {"DevNull", Const, 0}, @@ -9150,6 +9214,7 @@ var PackageSymbols = map[string][]Symbol{ {"IsLocal", Func, 20}, {"Join", Func, 0}, {"ListSeparator", Const, 0}, + {"Localize", Func, 23}, {"Match", Func, 0}, {"Rel", Func, 0}, {"Separator", Const, 0}, @@ -9232,6 +9297,8 @@ var PackageSymbols = map[string][]Symbol{ {"(Value).Pointer", Method, 0}, {"(Value).Recv", Method, 0}, {"(Value).Send", Method, 0}, + {"(Value).Seq", Method, 23}, + {"(Value).Seq2", Method, 23}, {"(Value).Set", Method, 0}, {"(Value).SetBool", Method, 0}, {"(Value).SetBytes", Method, 0}, @@ -9314,6 +9381,7 @@ var PackageSymbols = map[string][]Symbol{ {"SelectSend", Const, 1}, {"SendDir", Const, 0}, {"Slice", Const, 0}, + {"SliceAt", Func, 23}, {"SliceHeader", Type, 0}, {"SliceHeader.Cap", Field, 0}, {"SliceHeader.Data", Field, 0}, @@ -9655,6 +9723,7 @@ var PackageSymbols = map[string][]Symbol{ {"BuildSetting", Type, 18}, {"BuildSetting.Key", Field, 18}, {"BuildSetting.Value", Field, 18}, + {"CrashOptions", Type, 23}, {"FreeOSMemory", Func, 1}, {"GCStats", Type, 1}, {"GCStats.LastGC", Field, 1}, @@ -9672,6 +9741,7 @@ var PackageSymbols = map[string][]Symbol{ {"PrintStack", Func, 0}, {"ReadBuildInfo", Func, 12}, {"ReadGCStats", Func, 1}, + {"SetCrashOutput", Func, 23}, {"SetGCPercent", Func, 1}, {"SetMaxStack", Func, 2}, {"SetMaxThreads", Func, 2}, @@ -9742,10 +9812,15 @@ var PackageSymbols = map[string][]Symbol{ {"WithRegion", Func, 11}, }, "slices": { + {"All", Func, 23}, + {"AppendSeq", Func, 23}, + {"Backward", Func, 23}, {"BinarySearch", Func, 21}, {"BinarySearchFunc", Func, 21}, + {"Chunk", Func, 23}, {"Clip", Func, 21}, {"Clone", Func, 21}, + {"Collect", Func, 23}, {"Compact", Func, 21}, {"CompactFunc", Func, 21}, {"Compare", Func, 21}, @@ -9767,11 +9842,16 @@ var PackageSymbols = map[string][]Symbol{ {"MaxFunc", Func, 21}, {"Min", Func, 21}, {"MinFunc", Func, 21}, + {"Repeat", Func, 23}, {"Replace", Func, 21}, {"Reverse", Func, 21}, {"Sort", Func, 21}, {"SortFunc", Func, 21}, {"SortStableFunc", Func, 21}, + {"Sorted", Func, 23}, + {"SortedFunc", Func, 23}, + {"SortedStableFunc", Func, 23}, + {"Values", Func, 23}, }, "sort": { {"(Float64Slice).Len", Method, 0}, @@ -9936,10 +10016,14 @@ var PackageSymbols = map[string][]Symbol{ {"TrimSpace", Func, 0}, {"TrimSuffix", Func, 1}, }, + "structs": { + {"HostLayout", Type, 23}, + }, "sync": { {"(*Cond).Broadcast", Method, 0}, {"(*Cond).Signal", Method, 0}, {"(*Cond).Wait", Method, 0}, + {"(*Map).Clear", Method, 23}, {"(*Map).CompareAndDelete", Method, 20}, {"(*Map).CompareAndSwap", Method, 20}, {"(*Map).Delete", Method, 9}, @@ -9986,13 +10070,17 @@ var PackageSymbols = map[string][]Symbol{ {"(*Bool).Store", Method, 19}, {"(*Bool).Swap", Method, 19}, {"(*Int32).Add", Method, 19}, + {"(*Int32).And", Method, 23}, {"(*Int32).CompareAndSwap", Method, 19}, {"(*Int32).Load", Method, 19}, + {"(*Int32).Or", Method, 23}, {"(*Int32).Store", Method, 19}, {"(*Int32).Swap", Method, 19}, {"(*Int64).Add", Method, 19}, + {"(*Int64).And", Method, 23}, {"(*Int64).CompareAndSwap", Method, 19}, {"(*Int64).Load", Method, 19}, + {"(*Int64).Or", Method, 23}, {"(*Int64).Store", Method, 19}, {"(*Int64).Swap", Method, 19}, {"(*Pointer).CompareAndSwap", Method, 19}, @@ -10000,18 +10088,24 @@ var PackageSymbols = map[string][]Symbol{ {"(*Pointer).Store", Method, 19}, {"(*Pointer).Swap", Method, 19}, {"(*Uint32).Add", Method, 19}, + {"(*Uint32).And", Method, 23}, {"(*Uint32).CompareAndSwap", Method, 19}, {"(*Uint32).Load", Method, 19}, + {"(*Uint32).Or", Method, 23}, {"(*Uint32).Store", Method, 19}, {"(*Uint32).Swap", Method, 19}, {"(*Uint64).Add", Method, 19}, + {"(*Uint64).And", Method, 23}, {"(*Uint64).CompareAndSwap", Method, 19}, {"(*Uint64).Load", Method, 19}, + {"(*Uint64).Or", Method, 23}, {"(*Uint64).Store", Method, 19}, {"(*Uint64).Swap", Method, 19}, {"(*Uintptr).Add", Method, 19}, + {"(*Uintptr).And", Method, 23}, {"(*Uintptr).CompareAndSwap", Method, 19}, {"(*Uintptr).Load", Method, 19}, + {"(*Uintptr).Or", Method, 23}, {"(*Uintptr).Store", Method, 19}, {"(*Uintptr).Swap", Method, 19}, {"(*Value).CompareAndSwap", Method, 17}, @@ -10023,6 +10117,11 @@ var PackageSymbols = map[string][]Symbol{ {"AddUint32", Func, 0}, {"AddUint64", Func, 0}, {"AddUintptr", Func, 0}, + {"AndInt32", Func, 23}, + {"AndInt64", Func, 23}, + {"AndUint32", Func, 23}, + {"AndUint64", Func, 23}, + {"AndUintptr", Func, 23}, {"Bool", Type, 19}, {"CompareAndSwapInt32", Func, 0}, {"CompareAndSwapInt64", Func, 0}, @@ -10038,6 +10137,11 @@ var PackageSymbols = map[string][]Symbol{ {"LoadUint32", Func, 0}, {"LoadUint64", Func, 0}, {"LoadUintptr", Func, 0}, + {"OrInt32", Func, 23}, + {"OrInt64", Func, 23}, + {"OrUint32", Func, 23}, + {"OrUint64", Func, 23}, + {"OrUintptr", Func, 23}, {"Pointer", Type, 19}, {"StoreInt32", Func, 0}, {"StoreInt64", Func, 0}, @@ -16200,6 +16304,7 @@ var PackageSymbols = map[string][]Symbol{ {"WSAEACCES", Const, 2}, {"WSAECONNABORTED", Const, 9}, {"WSAECONNRESET", Const, 3}, + {"WSAENOPROTOOPT", Const, 23}, {"WSAEnumProtocols", Func, 2}, {"WSAID_CONNECTEX", Var, 1}, {"WSAIoctl", Func, 0}, @@ -17284,6 +17389,7 @@ var PackageSymbols = map[string][]Symbol{ {"Encode", Func, 0}, {"EncodeRune", Func, 0}, {"IsSurrogate", Func, 0}, + {"RuneLen", Func, 23}, }, "unicode/utf8": { {"AppendRune", Func, 18}, @@ -17306,6 +17412,11 @@ var PackageSymbols = map[string][]Symbol{ {"ValidRune", Func, 1}, {"ValidString", Func, 0}, }, + "unique": { + {"(Handle).Value", Method, 23}, + {"Handle", Type, 23}, + {"Make", Func, 23}, + }, "unsafe": { {"Add", Func, 0}, {"Alignof", Func, 0}, diff --git a/vendor/golang.org/x/tools/internal/typesinternal/types.go b/vendor/golang.org/x/tools/internal/typesinternal/types.go index 7c77c2fbc038f..83923286120d4 100644 --- a/vendor/golang.org/x/tools/internal/typesinternal/types.go +++ b/vendor/golang.org/x/tools/internal/typesinternal/types.go @@ -48,3 +48,18 @@ func ReadGo116ErrorData(err types.Error) (code ErrorCode, start, end token.Pos, } return ErrorCode(data[0]), token.Pos(data[1]), token.Pos(data[2]), true } + +// NameRelativeTo returns a types.Qualifier that qualifies members of +// all packages other than pkg, using only the package name. +// (By contrast, [types.RelativeTo] uses the complete package path, +// which is often excessive.) +// +// If pkg is nil, it is equivalent to [*types.Package.Name]. +func NameRelativeTo(pkg *types.Package) types.Qualifier { + return func(other *types.Package) string { + if pkg != nil && pkg == other { + return "" // same package; unqualified + } + return other.Name() + } +} diff --git a/vendor/golang.org/x/tools/internal/versions/constraint.go b/vendor/golang.org/x/tools/internal/versions/constraint.go new file mode 100644 index 0000000000000..179063d484860 --- /dev/null +++ b/vendor/golang.org/x/tools/internal/versions/constraint.go @@ -0,0 +1,13 @@ +// Copyright 2024 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package versions + +import "go/build/constraint" + +// ConstraintGoVersion is constraint.GoVersion (if built with go1.21+). +// Otherwise nil. +// +// Deprecate once x/tools is after go1.21. +var ConstraintGoVersion func(x constraint.Expr) string diff --git a/vendor/golang.org/x/tools/internal/versions/constraint_go121.go b/vendor/golang.org/x/tools/internal/versions/constraint_go121.go new file mode 100644 index 0000000000000..38011407d5f06 --- /dev/null +++ b/vendor/golang.org/x/tools/internal/versions/constraint_go121.go @@ -0,0 +1,14 @@ +// Copyright 2024 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +//go:build go1.21 +// +build go1.21 + +package versions + +import "go/build/constraint" + +func init() { + ConstraintGoVersion = constraint.GoVersion +} diff --git a/vendor/golang.org/x/tools/internal/versions/types_go122.go b/vendor/golang.org/x/tools/internal/versions/types_go122.go index e8180632a5262..aac5db62c983e 100644 --- a/vendor/golang.org/x/tools/internal/versions/types_go122.go +++ b/vendor/golang.org/x/tools/internal/versions/types_go122.go @@ -12,7 +12,7 @@ import ( "go/types" ) -// FileVersions returns a file's Go version. +// FileVersion returns a file's Go version. // The reported version is an unknown Future version if a // version cannot be determined. func FileVersion(info *types.Info, file *ast.File) string { diff --git a/vendor/google.golang.org/genproto/googleapis/api/annotations/client.pb.go b/vendor/google.golang.org/genproto/googleapis/api/annotations/client.pb.go index 636edb460a43a..aa69fb4d509ff 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/annotations/client.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/annotations/client.pb.go @@ -719,6 +719,8 @@ type PythonSettings struct { // Some settings. Common *CommonLanguageSettings `protobuf:"bytes,1,opt,name=common,proto3" json:"common,omitempty"` + // Experimental features to be included during client library generation. + ExperimentalFeatures *PythonSettings_ExperimentalFeatures `protobuf:"bytes,2,opt,name=experimental_features,json=experimentalFeatures,proto3" json:"experimental_features,omitempty"` } func (x *PythonSettings) Reset() { @@ -760,6 +762,13 @@ func (x *PythonSettings) GetCommon() *CommonLanguageSettings { return nil } +func (x *PythonSettings) GetExperimentalFeatures() *PythonSettings_ExperimentalFeatures { + if x != nil { + return x.ExperimentalFeatures + } + return nil +} + // Settings for Node client libraries. type NodeSettings struct { state protoimpl.MessageState @@ -1024,6 +1033,13 @@ type MethodSettings struct { // The fully qualified name of the method, for which the options below apply. // This is used to find the method to apply the options. + // + // Example: + // + // publishing: + // method_settings: + // - selector: google.storage.control.v2.StorageControl.CreateFolder + // # method settings for CreateFolder... Selector string `protobuf:"bytes,1,opt,name=selector,proto3" json:"selector,omitempty"` // Describes settings to use for long-running operations when generating // API methods for RPCs. Complements RPCs that use the annotations in @@ -1033,15 +1049,12 @@ type MethodSettings struct { // // publishing: // method_settings: - // - selector: google.cloud.speech.v2.Speech.BatchRecognize - // long_running: - // initial_poll_delay: - // seconds: 60 # 1 minute - // poll_delay_multiplier: 1.5 - // max_poll_delay: - // seconds: 360 # 6 minutes - // total_poll_timeout: - // seconds: 54000 # 90 minutes + // - selector: google.cloud.speech.v2.Speech.BatchRecognize + // long_running: + // initial_poll_delay: 60s # 1 minute + // poll_delay_multiplier: 1.5 + // max_poll_delay: 360s # 6 minutes + // total_poll_timeout: 54000s # 90 minutes LongRunning *MethodSettings_LongRunning `protobuf:"bytes,2,opt,name=long_running,json=longRunning,proto3" json:"long_running,omitempty"` // List of top-level fields of the request message, that should be // automatically populated by the client libraries based on their @@ -1051,9 +1064,9 @@ type MethodSettings struct { // // publishing: // method_settings: - // - selector: google.example.v1.ExampleService.CreateExample - // auto_populated_fields: - // - request_id + // - selector: google.example.v1.ExampleService.CreateExample + // auto_populated_fields: + // - request_id AutoPopulatedFields []string `protobuf:"bytes,3,rep,name=auto_populated_fields,json=autoPopulatedFields,proto3" json:"auto_populated_fields,omitempty"` } @@ -1110,6 +1123,60 @@ func (x *MethodSettings) GetAutoPopulatedFields() []string { return nil } +// Experimental features to be included during client library generation. +// These fields will be deprecated once the feature graduates and is enabled +// by default. +type PythonSettings_ExperimentalFeatures struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Enables generation of asynchronous REST clients if `rest` transport is + // enabled. By default, asynchronous REST clients will not be generated. + // This feature will be enabled by default 1 month after launching the + // feature in preview packages. + RestAsyncIoEnabled bool `protobuf:"varint,1,opt,name=rest_async_io_enabled,json=restAsyncIoEnabled,proto3" json:"rest_async_io_enabled,omitempty"` +} + +func (x *PythonSettings_ExperimentalFeatures) Reset() { + *x = PythonSettings_ExperimentalFeatures{} + if protoimpl.UnsafeEnabled { + mi := &file_google_api_client_proto_msgTypes[13] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *PythonSettings_ExperimentalFeatures) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*PythonSettings_ExperimentalFeatures) ProtoMessage() {} + +func (x *PythonSettings_ExperimentalFeatures) ProtoReflect() protoreflect.Message { + mi := &file_google_api_client_proto_msgTypes[13] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use PythonSettings_ExperimentalFeatures.ProtoReflect.Descriptor instead. +func (*PythonSettings_ExperimentalFeatures) Descriptor() ([]byte, []int) { + return file_google_api_client_proto_rawDescGZIP(), []int{6, 0} +} + +func (x *PythonSettings_ExperimentalFeatures) GetRestAsyncIoEnabled() bool { + if x != nil { + return x.RestAsyncIoEnabled + } + return false +} + // Describes settings to use when generating API methods that use the // long-running operation pattern. // All default values below are from those used in the client library @@ -1138,7 +1205,7 @@ type MethodSettings_LongRunning struct { func (x *MethodSettings_LongRunning) Reset() { *x = MethodSettings_LongRunning{} if protoimpl.UnsafeEnabled { - mi := &file_google_api_client_proto_msgTypes[15] + mi := &file_google_api_client_proto_msgTypes[16] ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) ms.StoreMessageInfo(mi) } @@ -1151,7 +1218,7 @@ func (x *MethodSettings_LongRunning) String() string { func (*MethodSettings_LongRunning) ProtoMessage() {} func (x *MethodSettings_LongRunning) ProtoReflect() protoreflect.Message { - mi := &file_google_api_client_proto_msgTypes[15] + mi := &file_google_api_client_proto_msgTypes[16] if protoimpl.UnsafeEnabled && x != nil { ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) if ms.LoadMessageInfo() == nil { @@ -1456,132 +1523,143 @@ var file_google_api_client_proto_rawDesc = []byte{ 0x73, 0x12, 0x3a, 0x0a, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x43, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x4c, 0x61, 0x6e, 0x67, 0x75, 0x61, 0x67, 0x65, 0x53, 0x65, 0x74, - 0x74, 0x69, 0x6e, 0x67, 0x73, 0x52, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x22, 0x4c, 0x0a, - 0x0e, 0x50, 0x79, 0x74, 0x68, 0x6f, 0x6e, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x12, - 0x3a, 0x0a, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, - 0x22, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x43, 0x6f, 0x6d, - 0x6d, 0x6f, 0x6e, 0x4c, 0x61, 0x6e, 0x67, 0x75, 0x61, 0x67, 0x65, 0x53, 0x65, 0x74, 0x74, 0x69, - 0x6e, 0x67, 0x73, 0x52, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x22, 0x4a, 0x0a, 0x0c, 0x4e, - 0x6f, 0x64, 0x65, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x12, 0x3a, 0x0a, 0x06, 0x63, - 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x67, 0x6f, - 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x43, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x4c, - 0x61, 0x6e, 0x67, 0x75, 0x61, 0x67, 0x65, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x52, - 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x22, 0xae, 0x04, 0x0a, 0x0e, 0x44, 0x6f, 0x74, 0x6e, - 0x65, 0x74, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x12, 0x3a, 0x0a, 0x06, 0x63, 0x6f, + 0x74, 0x69, 0x6e, 0x67, 0x73, 0x52, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x22, 0xfd, 0x01, + 0x0a, 0x0e, 0x50, 0x79, 0x74, 0x68, 0x6f, 0x6e, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, + 0x12, 0x3a, 0x0a, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, + 0x32, 0x22, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x43, 0x6f, + 0x6d, 0x6d, 0x6f, 0x6e, 0x4c, 0x61, 0x6e, 0x67, 0x75, 0x61, 0x67, 0x65, 0x53, 0x65, 0x74, 0x74, + 0x69, 0x6e, 0x67, 0x73, 0x52, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x12, 0x64, 0x0a, 0x15, + 0x65, 0x78, 0x70, 0x65, 0x72, 0x69, 0x6d, 0x65, 0x6e, 0x74, 0x61, 0x6c, 0x5f, 0x66, 0x65, 0x61, + 0x74, 0x75, 0x72, 0x65, 0x73, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x2f, 0x2e, 0x67, 0x6f, + 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x50, 0x79, 0x74, 0x68, 0x6f, 0x6e, 0x53, + 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x2e, 0x45, 0x78, 0x70, 0x65, 0x72, 0x69, 0x6d, 0x65, + 0x6e, 0x74, 0x61, 0x6c, 0x46, 0x65, 0x61, 0x74, 0x75, 0x72, 0x65, 0x73, 0x52, 0x14, 0x65, 0x78, + 0x70, 0x65, 0x72, 0x69, 0x6d, 0x65, 0x6e, 0x74, 0x61, 0x6c, 0x46, 0x65, 0x61, 0x74, 0x75, 0x72, + 0x65, 0x73, 0x1a, 0x49, 0x0a, 0x14, 0x45, 0x78, 0x70, 0x65, 0x72, 0x69, 0x6d, 0x65, 0x6e, 0x74, + 0x61, 0x6c, 0x46, 0x65, 0x61, 0x74, 0x75, 0x72, 0x65, 0x73, 0x12, 0x31, 0x0a, 0x15, 0x72, 0x65, + 0x73, 0x74, 0x5f, 0x61, 0x73, 0x79, 0x6e, 0x63, 0x5f, 0x69, 0x6f, 0x5f, 0x65, 0x6e, 0x61, 0x62, + 0x6c, 0x65, 0x64, 0x18, 0x01, 0x20, 0x01, 0x28, 0x08, 0x52, 0x12, 0x72, 0x65, 0x73, 0x74, 0x41, + 0x73, 0x79, 0x6e, 0x63, 0x49, 0x6f, 0x45, 0x6e, 0x61, 0x62, 0x6c, 0x65, 0x64, 0x22, 0x4a, 0x0a, + 0x0c, 0x4e, 0x6f, 0x64, 0x65, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x12, 0x3a, 0x0a, + 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, + 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x43, 0x6f, 0x6d, 0x6d, 0x6f, + 0x6e, 0x4c, 0x61, 0x6e, 0x67, 0x75, 0x61, 0x67, 0x65, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, + 0x73, 0x52, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x22, 0xae, 0x04, 0x0a, 0x0e, 0x44, 0x6f, + 0x74, 0x6e, 0x65, 0x74, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x12, 0x3a, 0x0a, 0x06, + 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x67, + 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x43, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, + 0x4c, 0x61, 0x6e, 0x67, 0x75, 0x61, 0x67, 0x65, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, + 0x52, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x12, 0x5a, 0x0a, 0x10, 0x72, 0x65, 0x6e, 0x61, + 0x6d, 0x65, 0x64, 0x5f, 0x73, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x73, 0x18, 0x02, 0x20, 0x03, + 0x28, 0x0b, 0x32, 0x2f, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, + 0x44, 0x6f, 0x74, 0x6e, 0x65, 0x74, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x2e, 0x52, + 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x73, 0x45, 0x6e, + 0x74, 0x72, 0x79, 0x52, 0x0f, 0x72, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x65, 0x72, 0x76, + 0x69, 0x63, 0x65, 0x73, 0x12, 0x5d, 0x0a, 0x11, 0x72, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x5f, + 0x72, 0x65, 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, 0x18, 0x03, 0x20, 0x03, 0x28, 0x0b, 0x32, + 0x30, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x44, 0x6f, 0x74, + 0x6e, 0x65, 0x74, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x2e, 0x52, 0x65, 0x6e, 0x61, + 0x6d, 0x65, 0x64, 0x52, 0x65, 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, 0x45, 0x6e, 0x74, 0x72, + 0x79, 0x52, 0x10, 0x72, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x52, 0x65, 0x73, 0x6f, 0x75, 0x72, + 0x63, 0x65, 0x73, 0x12, 0x2b, 0x0a, 0x11, 0x69, 0x67, 0x6e, 0x6f, 0x72, 0x65, 0x64, 0x5f, 0x72, + 0x65, 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, 0x18, 0x04, 0x20, 0x03, 0x28, 0x09, 0x52, 0x10, + 0x69, 0x67, 0x6e, 0x6f, 0x72, 0x65, 0x64, 0x52, 0x65, 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, + 0x12, 0x38, 0x0a, 0x18, 0x66, 0x6f, 0x72, 0x63, 0x65, 0x64, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x73, + 0x70, 0x61, 0x63, 0x65, 0x5f, 0x61, 0x6c, 0x69, 0x61, 0x73, 0x65, 0x73, 0x18, 0x05, 0x20, 0x03, + 0x28, 0x09, 0x52, 0x16, 0x66, 0x6f, 0x72, 0x63, 0x65, 0x64, 0x4e, 0x61, 0x6d, 0x65, 0x73, 0x70, + 0x61, 0x63, 0x65, 0x41, 0x6c, 0x69, 0x61, 0x73, 0x65, 0x73, 0x12, 0x35, 0x0a, 0x16, 0x68, 0x61, + 0x6e, 0x64, 0x77, 0x72, 0x69, 0x74, 0x74, 0x65, 0x6e, 0x5f, 0x73, 0x69, 0x67, 0x6e, 0x61, 0x74, + 0x75, 0x72, 0x65, 0x73, 0x18, 0x06, 0x20, 0x03, 0x28, 0x09, 0x52, 0x15, 0x68, 0x61, 0x6e, 0x64, + 0x77, 0x72, 0x69, 0x74, 0x74, 0x65, 0x6e, 0x53, 0x69, 0x67, 0x6e, 0x61, 0x74, 0x75, 0x72, 0x65, + 0x73, 0x1a, 0x42, 0x0a, 0x14, 0x52, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x65, 0x72, 0x76, + 0x69, 0x63, 0x65, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, 0x12, 0x10, 0x0a, 0x03, 0x6b, 0x65, 0x79, + 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x03, 0x6b, 0x65, 0x79, 0x12, 0x14, 0x0a, 0x05, 0x76, + 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, + 0x65, 0x3a, 0x02, 0x38, 0x01, 0x1a, 0x43, 0x0a, 0x15, 0x52, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, + 0x52, 0x65, 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, 0x12, 0x10, + 0x0a, 0x03, 0x6b, 0x65, 0x79, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x03, 0x6b, 0x65, 0x79, + 0x12, 0x14, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x3a, 0x02, 0x38, 0x01, 0x22, 0x4a, 0x0a, 0x0c, 0x52, 0x75, + 0x62, 0x79, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x12, 0x3a, 0x0a, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x43, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x4c, 0x61, 0x6e, 0x67, 0x75, 0x61, 0x67, 0x65, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x52, 0x06, - 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x12, 0x5a, 0x0a, 0x10, 0x72, 0x65, 0x6e, 0x61, 0x6d, 0x65, - 0x64, 0x5f, 0x73, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x73, 0x18, 0x02, 0x20, 0x03, 0x28, 0x0b, - 0x32, 0x2f, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x44, 0x6f, - 0x74, 0x6e, 0x65, 0x74, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x2e, 0x52, 0x65, 0x6e, - 0x61, 0x6d, 0x65, 0x64, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x73, 0x45, 0x6e, 0x74, 0x72, - 0x79, 0x52, 0x0f, 0x72, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, - 0x65, 0x73, 0x12, 0x5d, 0x0a, 0x11, 0x72, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x5f, 0x72, 0x65, - 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, 0x18, 0x03, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x30, 0x2e, - 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x44, 0x6f, 0x74, 0x6e, 0x65, - 0x74, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x2e, 0x52, 0x65, 0x6e, 0x61, 0x6d, 0x65, - 0x64, 0x52, 0x65, 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, 0x52, - 0x10, 0x72, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x52, 0x65, 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, - 0x73, 0x12, 0x2b, 0x0a, 0x11, 0x69, 0x67, 0x6e, 0x6f, 0x72, 0x65, 0x64, 0x5f, 0x72, 0x65, 0x73, - 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, 0x18, 0x04, 0x20, 0x03, 0x28, 0x09, 0x52, 0x10, 0x69, 0x67, - 0x6e, 0x6f, 0x72, 0x65, 0x64, 0x52, 0x65, 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, 0x12, 0x38, - 0x0a, 0x18, 0x66, 0x6f, 0x72, 0x63, 0x65, 0x64, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x73, 0x70, 0x61, - 0x63, 0x65, 0x5f, 0x61, 0x6c, 0x69, 0x61, 0x73, 0x65, 0x73, 0x18, 0x05, 0x20, 0x03, 0x28, 0x09, - 0x52, 0x16, 0x66, 0x6f, 0x72, 0x63, 0x65, 0x64, 0x4e, 0x61, 0x6d, 0x65, 0x73, 0x70, 0x61, 0x63, - 0x65, 0x41, 0x6c, 0x69, 0x61, 0x73, 0x65, 0x73, 0x12, 0x35, 0x0a, 0x16, 0x68, 0x61, 0x6e, 0x64, - 0x77, 0x72, 0x69, 0x74, 0x74, 0x65, 0x6e, 0x5f, 0x73, 0x69, 0x67, 0x6e, 0x61, 0x74, 0x75, 0x72, - 0x65, 0x73, 0x18, 0x06, 0x20, 0x03, 0x28, 0x09, 0x52, 0x15, 0x68, 0x61, 0x6e, 0x64, 0x77, 0x72, - 0x69, 0x74, 0x74, 0x65, 0x6e, 0x53, 0x69, 0x67, 0x6e, 0x61, 0x74, 0x75, 0x72, 0x65, 0x73, 0x1a, - 0x42, 0x0a, 0x14, 0x52, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, - 0x65, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, 0x12, 0x10, 0x0a, 0x03, 0x6b, 0x65, 0x79, 0x18, 0x01, - 0x20, 0x01, 0x28, 0x09, 0x52, 0x03, 0x6b, 0x65, 0x79, 0x12, 0x14, 0x0a, 0x05, 0x76, 0x61, 0x6c, - 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x3a, - 0x02, 0x38, 0x01, 0x1a, 0x43, 0x0a, 0x15, 0x52, 0x65, 0x6e, 0x61, 0x6d, 0x65, 0x64, 0x52, 0x65, - 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, 0x12, 0x10, 0x0a, 0x03, - 0x6b, 0x65, 0x79, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x03, 0x6b, 0x65, 0x79, 0x12, 0x14, - 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x05, 0x76, - 0x61, 0x6c, 0x75, 0x65, 0x3a, 0x02, 0x38, 0x01, 0x22, 0x4a, 0x0a, 0x0c, 0x52, 0x75, 0x62, 0x79, - 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x12, 0x3a, 0x0a, 0x06, 0x63, 0x6f, 0x6d, 0x6d, - 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, - 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x43, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x4c, 0x61, 0x6e, 0x67, - 0x75, 0x61, 0x67, 0x65, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x52, 0x06, 0x63, 0x6f, - 0x6d, 0x6d, 0x6f, 0x6e, 0x22, 0x48, 0x0a, 0x0a, 0x47, 0x6f, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, - 0x67, 0x73, 0x12, 0x3a, 0x0a, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, - 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, - 0x43, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x4c, 0x61, 0x6e, 0x67, 0x75, 0x61, 0x67, 0x65, 0x53, 0x65, - 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x52, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x22, 0xc2, - 0x03, 0x0a, 0x0e, 0x4d, 0x65, 0x74, 0x68, 0x6f, 0x64, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, - 0x73, 0x12, 0x1a, 0x0a, 0x08, 0x73, 0x65, 0x6c, 0x65, 0x63, 0x74, 0x6f, 0x72, 0x18, 0x01, 0x20, - 0x01, 0x28, 0x09, 0x52, 0x08, 0x73, 0x65, 0x6c, 0x65, 0x63, 0x74, 0x6f, 0x72, 0x12, 0x49, 0x0a, - 0x0c, 0x6c, 0x6f, 0x6e, 0x67, 0x5f, 0x72, 0x75, 0x6e, 0x6e, 0x69, 0x6e, 0x67, 0x18, 0x02, 0x20, - 0x01, 0x28, 0x0b, 0x32, 0x26, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, - 0x2e, 0x4d, 0x65, 0x74, 0x68, 0x6f, 0x64, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x2e, - 0x4c, 0x6f, 0x6e, 0x67, 0x52, 0x75, 0x6e, 0x6e, 0x69, 0x6e, 0x67, 0x52, 0x0b, 0x6c, 0x6f, 0x6e, - 0x67, 0x52, 0x75, 0x6e, 0x6e, 0x69, 0x6e, 0x67, 0x12, 0x32, 0x0a, 0x15, 0x61, 0x75, 0x74, 0x6f, - 0x5f, 0x70, 0x6f, 0x70, 0x75, 0x6c, 0x61, 0x74, 0x65, 0x64, 0x5f, 0x66, 0x69, 0x65, 0x6c, 0x64, - 0x73, 0x18, 0x03, 0x20, 0x03, 0x28, 0x09, 0x52, 0x13, 0x61, 0x75, 0x74, 0x6f, 0x50, 0x6f, 0x70, - 0x75, 0x6c, 0x61, 0x74, 0x65, 0x64, 0x46, 0x69, 0x65, 0x6c, 0x64, 0x73, 0x1a, 0x94, 0x02, 0x0a, - 0x0b, 0x4c, 0x6f, 0x6e, 0x67, 0x52, 0x75, 0x6e, 0x6e, 0x69, 0x6e, 0x67, 0x12, 0x47, 0x0a, 0x12, - 0x69, 0x6e, 0x69, 0x74, 0x69, 0x61, 0x6c, 0x5f, 0x70, 0x6f, 0x6c, 0x6c, 0x5f, 0x64, 0x65, 0x6c, - 0x61, 0x79, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x19, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, + 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x22, 0x48, 0x0a, 0x0a, 0x47, 0x6f, 0x53, 0x65, 0x74, 0x74, + 0x69, 0x6e, 0x67, 0x73, 0x12, 0x3a, 0x0a, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x18, 0x01, + 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, + 0x69, 0x2e, 0x43, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, 0x4c, 0x61, 0x6e, 0x67, 0x75, 0x61, 0x67, 0x65, + 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, 0x73, 0x52, 0x06, 0x63, 0x6f, 0x6d, 0x6d, 0x6f, 0x6e, + 0x22, 0xc2, 0x03, 0x0a, 0x0e, 0x4d, 0x65, 0x74, 0x68, 0x6f, 0x64, 0x53, 0x65, 0x74, 0x74, 0x69, + 0x6e, 0x67, 0x73, 0x12, 0x1a, 0x0a, 0x08, 0x73, 0x65, 0x6c, 0x65, 0x63, 0x74, 0x6f, 0x72, 0x18, + 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x08, 0x73, 0x65, 0x6c, 0x65, 0x63, 0x74, 0x6f, 0x72, 0x12, + 0x49, 0x0a, 0x0c, 0x6c, 0x6f, 0x6e, 0x67, 0x5f, 0x72, 0x75, 0x6e, 0x6e, 0x69, 0x6e, 0x67, 0x18, + 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x26, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, + 0x70, 0x69, 0x2e, 0x4d, 0x65, 0x74, 0x68, 0x6f, 0x64, 0x53, 0x65, 0x74, 0x74, 0x69, 0x6e, 0x67, + 0x73, 0x2e, 0x4c, 0x6f, 0x6e, 0x67, 0x52, 0x75, 0x6e, 0x6e, 0x69, 0x6e, 0x67, 0x52, 0x0b, 0x6c, + 0x6f, 0x6e, 0x67, 0x52, 0x75, 0x6e, 0x6e, 0x69, 0x6e, 0x67, 0x12, 0x32, 0x0a, 0x15, 0x61, 0x75, + 0x74, 0x6f, 0x5f, 0x70, 0x6f, 0x70, 0x75, 0x6c, 0x61, 0x74, 0x65, 0x64, 0x5f, 0x66, 0x69, 0x65, + 0x6c, 0x64, 0x73, 0x18, 0x03, 0x20, 0x03, 0x28, 0x09, 0x52, 0x13, 0x61, 0x75, 0x74, 0x6f, 0x50, + 0x6f, 0x70, 0x75, 0x6c, 0x61, 0x74, 0x65, 0x64, 0x46, 0x69, 0x65, 0x6c, 0x64, 0x73, 0x1a, 0x94, + 0x02, 0x0a, 0x0b, 0x4c, 0x6f, 0x6e, 0x67, 0x52, 0x75, 0x6e, 0x6e, 0x69, 0x6e, 0x67, 0x12, 0x47, + 0x0a, 0x12, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x61, 0x6c, 0x5f, 0x70, 0x6f, 0x6c, 0x6c, 0x5f, 0x64, + 0x65, 0x6c, 0x61, 0x79, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x19, 0x2e, 0x67, 0x6f, 0x6f, + 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x44, 0x75, 0x72, + 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x10, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x61, 0x6c, 0x50, 0x6f, + 0x6c, 0x6c, 0x44, 0x65, 0x6c, 0x61, 0x79, 0x12, 0x32, 0x0a, 0x15, 0x70, 0x6f, 0x6c, 0x6c, 0x5f, + 0x64, 0x65, 0x6c, 0x61, 0x79, 0x5f, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x69, 0x65, 0x72, + 0x18, 0x02, 0x20, 0x01, 0x28, 0x02, 0x52, 0x13, 0x70, 0x6f, 0x6c, 0x6c, 0x44, 0x65, 0x6c, 0x61, + 0x79, 0x4d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x69, 0x65, 0x72, 0x12, 0x3f, 0x0a, 0x0e, 0x6d, + 0x61, 0x78, 0x5f, 0x70, 0x6f, 0x6c, 0x6c, 0x5f, 0x64, 0x65, 0x6c, 0x61, 0x79, 0x18, 0x03, 0x20, + 0x01, 0x28, 0x0b, 0x32, 0x19, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, + 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x44, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x0c, + 0x6d, 0x61, 0x78, 0x50, 0x6f, 0x6c, 0x6c, 0x44, 0x65, 0x6c, 0x61, 0x79, 0x12, 0x47, 0x0a, 0x12, + 0x74, 0x6f, 0x74, 0x61, 0x6c, 0x5f, 0x70, 0x6f, 0x6c, 0x6c, 0x5f, 0x74, 0x69, 0x6d, 0x65, 0x6f, + 0x75, 0x74, 0x18, 0x04, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x19, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x44, 0x75, 0x72, 0x61, 0x74, - 0x69, 0x6f, 0x6e, 0x52, 0x10, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x61, 0x6c, 0x50, 0x6f, 0x6c, 0x6c, - 0x44, 0x65, 0x6c, 0x61, 0x79, 0x12, 0x32, 0x0a, 0x15, 0x70, 0x6f, 0x6c, 0x6c, 0x5f, 0x64, 0x65, - 0x6c, 0x61, 0x79, 0x5f, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x69, 0x65, 0x72, 0x18, 0x02, - 0x20, 0x01, 0x28, 0x02, 0x52, 0x13, 0x70, 0x6f, 0x6c, 0x6c, 0x44, 0x65, 0x6c, 0x61, 0x79, 0x4d, - 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x69, 0x65, 0x72, 0x12, 0x3f, 0x0a, 0x0e, 0x6d, 0x61, 0x78, - 0x5f, 0x70, 0x6f, 0x6c, 0x6c, 0x5f, 0x64, 0x65, 0x6c, 0x61, 0x79, 0x18, 0x03, 0x20, 0x01, 0x28, - 0x0b, 0x32, 0x19, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, - 0x62, 0x75, 0x66, 0x2e, 0x44, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x0c, 0x6d, 0x61, - 0x78, 0x50, 0x6f, 0x6c, 0x6c, 0x44, 0x65, 0x6c, 0x61, 0x79, 0x12, 0x47, 0x0a, 0x12, 0x74, 0x6f, - 0x74, 0x61, 0x6c, 0x5f, 0x70, 0x6f, 0x6c, 0x6c, 0x5f, 0x74, 0x69, 0x6d, 0x65, 0x6f, 0x75, 0x74, - 0x18, 0x04, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x19, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, - 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x44, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, - 0x6e, 0x52, 0x10, 0x74, 0x6f, 0x74, 0x61, 0x6c, 0x50, 0x6f, 0x6c, 0x6c, 0x54, 0x69, 0x6d, 0x65, - 0x6f, 0x75, 0x74, 0x2a, 0xa3, 0x01, 0x0a, 0x19, 0x43, 0x6c, 0x69, 0x65, 0x6e, 0x74, 0x4c, 0x69, - 0x62, 0x72, 0x61, 0x72, 0x79, 0x4f, 0x72, 0x67, 0x61, 0x6e, 0x69, 0x7a, 0x61, 0x74, 0x69, 0x6f, - 0x6e, 0x12, 0x2b, 0x0a, 0x27, 0x43, 0x4c, 0x49, 0x45, 0x4e, 0x54, 0x5f, 0x4c, 0x49, 0x42, 0x52, - 0x41, 0x52, 0x59, 0x5f, 0x4f, 0x52, 0x47, 0x41, 0x4e, 0x49, 0x5a, 0x41, 0x54, 0x49, 0x4f, 0x4e, - 0x5f, 0x55, 0x4e, 0x53, 0x50, 0x45, 0x43, 0x49, 0x46, 0x49, 0x45, 0x44, 0x10, 0x00, 0x12, 0x09, - 0x0a, 0x05, 0x43, 0x4c, 0x4f, 0x55, 0x44, 0x10, 0x01, 0x12, 0x07, 0x0a, 0x03, 0x41, 0x44, 0x53, - 0x10, 0x02, 0x12, 0x0a, 0x0a, 0x06, 0x50, 0x48, 0x4f, 0x54, 0x4f, 0x53, 0x10, 0x03, 0x12, 0x0f, - 0x0a, 0x0b, 0x53, 0x54, 0x52, 0x45, 0x45, 0x54, 0x5f, 0x56, 0x49, 0x45, 0x57, 0x10, 0x04, 0x12, - 0x0c, 0x0a, 0x08, 0x53, 0x48, 0x4f, 0x50, 0x50, 0x49, 0x4e, 0x47, 0x10, 0x05, 0x12, 0x07, 0x0a, - 0x03, 0x47, 0x45, 0x4f, 0x10, 0x06, 0x12, 0x11, 0x0a, 0x0d, 0x47, 0x45, 0x4e, 0x45, 0x52, 0x41, - 0x54, 0x49, 0x56, 0x45, 0x5f, 0x41, 0x49, 0x10, 0x07, 0x2a, 0x67, 0x0a, 0x18, 0x43, 0x6c, 0x69, - 0x65, 0x6e, 0x74, 0x4c, 0x69, 0x62, 0x72, 0x61, 0x72, 0x79, 0x44, 0x65, 0x73, 0x74, 0x69, 0x6e, - 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x2a, 0x0a, 0x26, 0x43, 0x4c, 0x49, 0x45, 0x4e, 0x54, 0x5f, - 0x4c, 0x49, 0x42, 0x52, 0x41, 0x52, 0x59, 0x5f, 0x44, 0x45, 0x53, 0x54, 0x49, 0x4e, 0x41, 0x54, - 0x49, 0x4f, 0x4e, 0x5f, 0x55, 0x4e, 0x53, 0x50, 0x45, 0x43, 0x49, 0x46, 0x49, 0x45, 0x44, 0x10, - 0x00, 0x12, 0x0a, 0x0a, 0x06, 0x47, 0x49, 0x54, 0x48, 0x55, 0x42, 0x10, 0x0a, 0x12, 0x13, 0x0a, - 0x0f, 0x50, 0x41, 0x43, 0x4b, 0x41, 0x47, 0x45, 0x5f, 0x4d, 0x41, 0x4e, 0x41, 0x47, 0x45, 0x52, - 0x10, 0x14, 0x3a, 0x4a, 0x0a, 0x10, 0x6d, 0x65, 0x74, 0x68, 0x6f, 0x64, 0x5f, 0x73, 0x69, 0x67, - 0x6e, 0x61, 0x74, 0x75, 0x72, 0x65, 0x12, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, - 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x4d, 0x65, 0x74, 0x68, 0x6f, 0x64, 0x4f, - 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x18, 0x9b, 0x08, 0x20, 0x03, 0x28, 0x09, 0x52, 0x0f, 0x6d, - 0x65, 0x74, 0x68, 0x6f, 0x64, 0x53, 0x69, 0x67, 0x6e, 0x61, 0x74, 0x75, 0x72, 0x65, 0x3a, 0x43, - 0x0a, 0x0c, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x5f, 0x68, 0x6f, 0x73, 0x74, 0x12, 0x1f, - 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, - 0x2e, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x4f, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x18, - 0x99, 0x08, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x48, - 0x6f, 0x73, 0x74, 0x3a, 0x43, 0x0a, 0x0c, 0x6f, 0x61, 0x75, 0x74, 0x68, 0x5f, 0x73, 0x63, 0x6f, - 0x70, 0x65, 0x73, 0x12, 0x1f, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, - 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x4f, 0x70, 0x74, - 0x69, 0x6f, 0x6e, 0x73, 0x18, 0x9a, 0x08, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x6f, 0x61, 0x75, - 0x74, 0x68, 0x53, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x3a, 0x44, 0x0a, 0x0b, 0x61, 0x70, 0x69, 0x5f, - 0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x1f, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, - 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, - 0x65, 0x4f, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x18, 0xc1, 0xba, 0xab, 0xfa, 0x01, 0x20, 0x01, - 0x28, 0x09, 0x52, 0x0a, 0x61, 0x70, 0x69, 0x56, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x42, 0x69, - 0x0a, 0x0e, 0x63, 0x6f, 0x6d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, - 0x42, 0x0b, 0x43, 0x6c, 0x69, 0x65, 0x6e, 0x74, 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x50, 0x01, 0x5a, - 0x41, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, 0x2e, 0x6f, - 0x72, 0x67, 0x2f, 0x67, 0x65, 0x6e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x2f, 0x67, 0x6f, 0x6f, 0x67, - 0x6c, 0x65, 0x61, 0x70, 0x69, 0x73, 0x2f, 0x61, 0x70, 0x69, 0x2f, 0x61, 0x6e, 0x6e, 0x6f, 0x74, - 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x3b, 0x61, 0x6e, 0x6e, 0x6f, 0x74, 0x61, 0x74, 0x69, 0x6f, - 0x6e, 0x73, 0xa2, 0x02, 0x04, 0x47, 0x41, 0x50, 0x49, 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, - 0x33, + 0x69, 0x6f, 0x6e, 0x52, 0x10, 0x74, 0x6f, 0x74, 0x61, 0x6c, 0x50, 0x6f, 0x6c, 0x6c, 0x54, 0x69, + 0x6d, 0x65, 0x6f, 0x75, 0x74, 0x2a, 0xa3, 0x01, 0x0a, 0x19, 0x43, 0x6c, 0x69, 0x65, 0x6e, 0x74, + 0x4c, 0x69, 0x62, 0x72, 0x61, 0x72, 0x79, 0x4f, 0x72, 0x67, 0x61, 0x6e, 0x69, 0x7a, 0x61, 0x74, + 0x69, 0x6f, 0x6e, 0x12, 0x2b, 0x0a, 0x27, 0x43, 0x4c, 0x49, 0x45, 0x4e, 0x54, 0x5f, 0x4c, 0x49, + 0x42, 0x52, 0x41, 0x52, 0x59, 0x5f, 0x4f, 0x52, 0x47, 0x41, 0x4e, 0x49, 0x5a, 0x41, 0x54, 0x49, + 0x4f, 0x4e, 0x5f, 0x55, 0x4e, 0x53, 0x50, 0x45, 0x43, 0x49, 0x46, 0x49, 0x45, 0x44, 0x10, 0x00, + 0x12, 0x09, 0x0a, 0x05, 0x43, 0x4c, 0x4f, 0x55, 0x44, 0x10, 0x01, 0x12, 0x07, 0x0a, 0x03, 0x41, + 0x44, 0x53, 0x10, 0x02, 0x12, 0x0a, 0x0a, 0x06, 0x50, 0x48, 0x4f, 0x54, 0x4f, 0x53, 0x10, 0x03, + 0x12, 0x0f, 0x0a, 0x0b, 0x53, 0x54, 0x52, 0x45, 0x45, 0x54, 0x5f, 0x56, 0x49, 0x45, 0x57, 0x10, + 0x04, 0x12, 0x0c, 0x0a, 0x08, 0x53, 0x48, 0x4f, 0x50, 0x50, 0x49, 0x4e, 0x47, 0x10, 0x05, 0x12, + 0x07, 0x0a, 0x03, 0x47, 0x45, 0x4f, 0x10, 0x06, 0x12, 0x11, 0x0a, 0x0d, 0x47, 0x45, 0x4e, 0x45, + 0x52, 0x41, 0x54, 0x49, 0x56, 0x45, 0x5f, 0x41, 0x49, 0x10, 0x07, 0x2a, 0x67, 0x0a, 0x18, 0x43, + 0x6c, 0x69, 0x65, 0x6e, 0x74, 0x4c, 0x69, 0x62, 0x72, 0x61, 0x72, 0x79, 0x44, 0x65, 0x73, 0x74, + 0x69, 0x6e, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x2a, 0x0a, 0x26, 0x43, 0x4c, 0x49, 0x45, 0x4e, + 0x54, 0x5f, 0x4c, 0x49, 0x42, 0x52, 0x41, 0x52, 0x59, 0x5f, 0x44, 0x45, 0x53, 0x54, 0x49, 0x4e, + 0x41, 0x54, 0x49, 0x4f, 0x4e, 0x5f, 0x55, 0x4e, 0x53, 0x50, 0x45, 0x43, 0x49, 0x46, 0x49, 0x45, + 0x44, 0x10, 0x00, 0x12, 0x0a, 0x0a, 0x06, 0x47, 0x49, 0x54, 0x48, 0x55, 0x42, 0x10, 0x0a, 0x12, + 0x13, 0x0a, 0x0f, 0x50, 0x41, 0x43, 0x4b, 0x41, 0x47, 0x45, 0x5f, 0x4d, 0x41, 0x4e, 0x41, 0x47, + 0x45, 0x52, 0x10, 0x14, 0x3a, 0x4a, 0x0a, 0x10, 0x6d, 0x65, 0x74, 0x68, 0x6f, 0x64, 0x5f, 0x73, + 0x69, 0x67, 0x6e, 0x61, 0x74, 0x75, 0x72, 0x65, 0x12, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, + 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x4d, 0x65, 0x74, 0x68, 0x6f, + 0x64, 0x4f, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x18, 0x9b, 0x08, 0x20, 0x03, 0x28, 0x09, 0x52, + 0x0f, 0x6d, 0x65, 0x74, 0x68, 0x6f, 0x64, 0x53, 0x69, 0x67, 0x6e, 0x61, 0x74, 0x75, 0x72, 0x65, + 0x3a, 0x43, 0x0a, 0x0c, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x5f, 0x68, 0x6f, 0x73, 0x74, + 0x12, 0x1f, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, + 0x75, 0x66, 0x2e, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x4f, 0x70, 0x74, 0x69, 0x6f, 0x6e, + 0x73, 0x18, 0x99, 0x08, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, + 0x74, 0x48, 0x6f, 0x73, 0x74, 0x3a, 0x43, 0x0a, 0x0c, 0x6f, 0x61, 0x75, 0x74, 0x68, 0x5f, 0x73, + 0x63, 0x6f, 0x70, 0x65, 0x73, 0x12, 0x1f, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, + 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x4f, + 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x18, 0x9a, 0x08, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x6f, + 0x61, 0x75, 0x74, 0x68, 0x53, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x3a, 0x44, 0x0a, 0x0b, 0x61, 0x70, + 0x69, 0x5f, 0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x1f, 0x2e, 0x67, 0x6f, 0x6f, 0x67, + 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x53, 0x65, 0x72, 0x76, + 0x69, 0x63, 0x65, 0x4f, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x18, 0xc1, 0xba, 0xab, 0xfa, 0x01, + 0x20, 0x01, 0x28, 0x09, 0x52, 0x0a, 0x61, 0x70, 0x69, 0x56, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, + 0x42, 0x69, 0x0a, 0x0e, 0x63, 0x6f, 0x6d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, + 0x70, 0x69, 0x42, 0x0b, 0x43, 0x6c, 0x69, 0x65, 0x6e, 0x74, 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x50, + 0x01, 0x5a, 0x41, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, + 0x2e, 0x6f, 0x72, 0x67, 0x2f, 0x67, 0x65, 0x6e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x2f, 0x67, 0x6f, + 0x6f, 0x67, 0x6c, 0x65, 0x61, 0x70, 0x69, 0x73, 0x2f, 0x61, 0x70, 0x69, 0x2f, 0x61, 0x6e, 0x6e, + 0x6f, 0x74, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x3b, 0x61, 0x6e, 0x6e, 0x6f, 0x74, 0x61, 0x74, + 0x69, 0x6f, 0x6e, 0x73, 0xa2, 0x02, 0x04, 0x47, 0x41, 0x50, 0x49, 0x62, 0x06, 0x70, 0x72, 0x6f, + 0x74, 0x6f, 0x33, } var ( @@ -1597,34 +1675,35 @@ func file_google_api_client_proto_rawDescGZIP() []byte { } var file_google_api_client_proto_enumTypes = make([]protoimpl.EnumInfo, 2) -var file_google_api_client_proto_msgTypes = make([]protoimpl.MessageInfo, 16) +var file_google_api_client_proto_msgTypes = make([]protoimpl.MessageInfo, 17) var file_google_api_client_proto_goTypes = []interface{}{ - (ClientLibraryOrganization)(0), // 0: google.api.ClientLibraryOrganization - (ClientLibraryDestination)(0), // 1: google.api.ClientLibraryDestination - (*CommonLanguageSettings)(nil), // 2: google.api.CommonLanguageSettings - (*ClientLibrarySettings)(nil), // 3: google.api.ClientLibrarySettings - (*Publishing)(nil), // 4: google.api.Publishing - (*JavaSettings)(nil), // 5: google.api.JavaSettings - (*CppSettings)(nil), // 6: google.api.CppSettings - (*PhpSettings)(nil), // 7: google.api.PhpSettings - (*PythonSettings)(nil), // 8: google.api.PythonSettings - (*NodeSettings)(nil), // 9: google.api.NodeSettings - (*DotnetSettings)(nil), // 10: google.api.DotnetSettings - (*RubySettings)(nil), // 11: google.api.RubySettings - (*GoSettings)(nil), // 12: google.api.GoSettings - (*MethodSettings)(nil), // 13: google.api.MethodSettings - nil, // 14: google.api.JavaSettings.ServiceClassNamesEntry - nil, // 15: google.api.DotnetSettings.RenamedServicesEntry - nil, // 16: google.api.DotnetSettings.RenamedResourcesEntry - (*MethodSettings_LongRunning)(nil), // 17: google.api.MethodSettings.LongRunning - (api.LaunchStage)(0), // 18: google.api.LaunchStage - (*durationpb.Duration)(nil), // 19: google.protobuf.Duration - (*descriptorpb.MethodOptions)(nil), // 20: google.protobuf.MethodOptions - (*descriptorpb.ServiceOptions)(nil), // 21: google.protobuf.ServiceOptions + (ClientLibraryOrganization)(0), // 0: google.api.ClientLibraryOrganization + (ClientLibraryDestination)(0), // 1: google.api.ClientLibraryDestination + (*CommonLanguageSettings)(nil), // 2: google.api.CommonLanguageSettings + (*ClientLibrarySettings)(nil), // 3: google.api.ClientLibrarySettings + (*Publishing)(nil), // 4: google.api.Publishing + (*JavaSettings)(nil), // 5: google.api.JavaSettings + (*CppSettings)(nil), // 6: google.api.CppSettings + (*PhpSettings)(nil), // 7: google.api.PhpSettings + (*PythonSettings)(nil), // 8: google.api.PythonSettings + (*NodeSettings)(nil), // 9: google.api.NodeSettings + (*DotnetSettings)(nil), // 10: google.api.DotnetSettings + (*RubySettings)(nil), // 11: google.api.RubySettings + (*GoSettings)(nil), // 12: google.api.GoSettings + (*MethodSettings)(nil), // 13: google.api.MethodSettings + nil, // 14: google.api.JavaSettings.ServiceClassNamesEntry + (*PythonSettings_ExperimentalFeatures)(nil), // 15: google.api.PythonSettings.ExperimentalFeatures + nil, // 16: google.api.DotnetSettings.RenamedServicesEntry + nil, // 17: google.api.DotnetSettings.RenamedResourcesEntry + (*MethodSettings_LongRunning)(nil), // 18: google.api.MethodSettings.LongRunning + (api.LaunchStage)(0), // 19: google.api.LaunchStage + (*durationpb.Duration)(nil), // 20: google.protobuf.Duration + (*descriptorpb.MethodOptions)(nil), // 21: google.protobuf.MethodOptions + (*descriptorpb.ServiceOptions)(nil), // 22: google.protobuf.ServiceOptions } var file_google_api_client_proto_depIdxs = []int32{ 1, // 0: google.api.CommonLanguageSettings.destinations:type_name -> google.api.ClientLibraryDestination - 18, // 1: google.api.ClientLibrarySettings.launch_stage:type_name -> google.api.LaunchStage + 19, // 1: google.api.ClientLibrarySettings.launch_stage:type_name -> google.api.LaunchStage 5, // 2: google.api.ClientLibrarySettings.java_settings:type_name -> google.api.JavaSettings 6, // 3: google.api.ClientLibrarySettings.cpp_settings:type_name -> google.api.CppSettings 7, // 4: google.api.ClientLibrarySettings.php_settings:type_name -> google.api.PhpSettings @@ -1641,25 +1720,26 @@ var file_google_api_client_proto_depIdxs = []int32{ 2, // 15: google.api.CppSettings.common:type_name -> google.api.CommonLanguageSettings 2, // 16: google.api.PhpSettings.common:type_name -> google.api.CommonLanguageSettings 2, // 17: google.api.PythonSettings.common:type_name -> google.api.CommonLanguageSettings - 2, // 18: google.api.NodeSettings.common:type_name -> google.api.CommonLanguageSettings - 2, // 19: google.api.DotnetSettings.common:type_name -> google.api.CommonLanguageSettings - 15, // 20: google.api.DotnetSettings.renamed_services:type_name -> google.api.DotnetSettings.RenamedServicesEntry - 16, // 21: google.api.DotnetSettings.renamed_resources:type_name -> google.api.DotnetSettings.RenamedResourcesEntry - 2, // 22: google.api.RubySettings.common:type_name -> google.api.CommonLanguageSettings - 2, // 23: google.api.GoSettings.common:type_name -> google.api.CommonLanguageSettings - 17, // 24: google.api.MethodSettings.long_running:type_name -> google.api.MethodSettings.LongRunning - 19, // 25: google.api.MethodSettings.LongRunning.initial_poll_delay:type_name -> google.protobuf.Duration - 19, // 26: google.api.MethodSettings.LongRunning.max_poll_delay:type_name -> google.protobuf.Duration - 19, // 27: google.api.MethodSettings.LongRunning.total_poll_timeout:type_name -> google.protobuf.Duration - 20, // 28: google.api.method_signature:extendee -> google.protobuf.MethodOptions - 21, // 29: google.api.default_host:extendee -> google.protobuf.ServiceOptions - 21, // 30: google.api.oauth_scopes:extendee -> google.protobuf.ServiceOptions - 21, // 31: google.api.api_version:extendee -> google.protobuf.ServiceOptions - 32, // [32:32] is the sub-list for method output_type - 32, // [32:32] is the sub-list for method input_type - 32, // [32:32] is the sub-list for extension type_name - 28, // [28:32] is the sub-list for extension extendee - 0, // [0:28] is the sub-list for field type_name + 15, // 18: google.api.PythonSettings.experimental_features:type_name -> google.api.PythonSettings.ExperimentalFeatures + 2, // 19: google.api.NodeSettings.common:type_name -> google.api.CommonLanguageSettings + 2, // 20: google.api.DotnetSettings.common:type_name -> google.api.CommonLanguageSettings + 16, // 21: google.api.DotnetSettings.renamed_services:type_name -> google.api.DotnetSettings.RenamedServicesEntry + 17, // 22: google.api.DotnetSettings.renamed_resources:type_name -> google.api.DotnetSettings.RenamedResourcesEntry + 2, // 23: google.api.RubySettings.common:type_name -> google.api.CommonLanguageSettings + 2, // 24: google.api.GoSettings.common:type_name -> google.api.CommonLanguageSettings + 18, // 25: google.api.MethodSettings.long_running:type_name -> google.api.MethodSettings.LongRunning + 20, // 26: google.api.MethodSettings.LongRunning.initial_poll_delay:type_name -> google.protobuf.Duration + 20, // 27: google.api.MethodSettings.LongRunning.max_poll_delay:type_name -> google.protobuf.Duration + 20, // 28: google.api.MethodSettings.LongRunning.total_poll_timeout:type_name -> google.protobuf.Duration + 21, // 29: google.api.method_signature:extendee -> google.protobuf.MethodOptions + 22, // 30: google.api.default_host:extendee -> google.protobuf.ServiceOptions + 22, // 31: google.api.oauth_scopes:extendee -> google.protobuf.ServiceOptions + 22, // 32: google.api.api_version:extendee -> google.protobuf.ServiceOptions + 33, // [33:33] is the sub-list for method output_type + 33, // [33:33] is the sub-list for method input_type + 33, // [33:33] is the sub-list for extension type_name + 29, // [29:33] is the sub-list for extension extendee + 0, // [0:29] is the sub-list for field type_name } func init() { file_google_api_client_proto_init() } @@ -1812,7 +1892,19 @@ func file_google_api_client_proto_init() { return nil } } - file_google_api_client_proto_msgTypes[15].Exporter = func(v interface{}, i int) interface{} { + file_google_api_client_proto_msgTypes[13].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*PythonSettings_ExperimentalFeatures); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_google_api_client_proto_msgTypes[16].Exporter = func(v interface{}, i int) interface{} { switch v := v.(*MethodSettings_LongRunning); i { case 0: return &v.state @@ -1831,7 +1923,7 @@ func file_google_api_client_proto_init() { GoPackagePath: reflect.TypeOf(x{}).PkgPath(), RawDescriptor: file_google_api_client_proto_rawDesc, NumEnums: 2, - NumMessages: 16, + NumMessages: 17, NumExtensions: 4, NumServices: 0, }, diff --git a/vendor/google.golang.org/genproto/googleapis/api/annotations/field_info.pb.go b/vendor/google.golang.org/genproto/googleapis/api/annotations/field_info.pb.go index d339dfb02ac12..a462e7d013275 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/annotations/field_info.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/annotations/field_info.pb.go @@ -121,6 +121,11 @@ type FieldInfo struct { // any API consumer, just documents the API's format for the field it is // applied to. Format FieldInfo_Format `protobuf:"varint,1,opt,name=format,proto3,enum=google.api.FieldInfo_Format" json:"format,omitempty"` + // The type(s) that the annotated, generic field may represent. + // + // Currently, this must only be used on fields of type `google.protobuf.Any`. + // Supporting other generic types may be considered in the future. + ReferencedTypes []*TypeReference `protobuf:"bytes,2,rep,name=referenced_types,json=referencedTypes,proto3" json:"referenced_types,omitempty"` } func (x *FieldInfo) Reset() { @@ -162,6 +167,70 @@ func (x *FieldInfo) GetFormat() FieldInfo_Format { return FieldInfo_FORMAT_UNSPECIFIED } +func (x *FieldInfo) GetReferencedTypes() []*TypeReference { + if x != nil { + return x.ReferencedTypes + } + return nil +} + +// A reference to a message type, for use in [FieldInfo][google.api.FieldInfo]. +type TypeReference struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // The name of the type that the annotated, generic field may represent. + // If the type is in the same protobuf package, the value can be the simple + // message name e.g., `"MyMessage"`. Otherwise, the value must be the + // fully-qualified message name e.g., `"google.library.v1.Book"`. + // + // If the type(s) are unknown to the service (e.g. the field accepts generic + // user input), use the wildcard `"*"` to denote this behavior. + // + // See [AIP-202](https://google.aip.dev/202#type-references) for more details. + TypeName string `protobuf:"bytes,1,opt,name=type_name,json=typeName,proto3" json:"type_name,omitempty"` +} + +func (x *TypeReference) Reset() { + *x = TypeReference{} + if protoimpl.UnsafeEnabled { + mi := &file_google_api_field_info_proto_msgTypes[1] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *TypeReference) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*TypeReference) ProtoMessage() {} + +func (x *TypeReference) ProtoReflect() protoreflect.Message { + mi := &file_google_api_field_info_proto_msgTypes[1] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use TypeReference.ProtoReflect.Descriptor instead. +func (*TypeReference) Descriptor() ([]byte, []int) { + return file_google_api_field_info_proto_rawDescGZIP(), []int{1} +} + +func (x *TypeReference) GetTypeName() string { + if x != nil { + return x.TypeName + } + return "" +} + var file_google_api_field_info_proto_extTypes = []protoimpl.ExtensionInfo{ { ExtendedType: (*descriptorpb.FieldOptions)(nil), @@ -185,6 +254,13 @@ var ( // string actual_ip_address = 4 [ // (google.api.field_info).format = IPV4_OR_IPV6 // ]; + // google.protobuf.Any generic_field = 5 [ + // (google.api.field_info).referenced_types = {type_name: "ActualType"}, + // (google.api.field_info).referenced_types = {type_name: "OtherType"}, + // ]; + // google.protobuf.Any generic_user_input = 5 [ + // (google.api.field_info).referenced_types = {type_name: "*"}, + // ]; // // optional google.api.FieldInfo field_info = 291403980; E_FieldInfo = &file_google_api_field_info_proto_extTypes[0] @@ -197,30 +273,37 @@ var file_google_api_field_info_proto_rawDesc = []byte{ 0x6c, 0x64, 0x5f, 0x69, 0x6e, 0x66, 0x6f, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x12, 0x0a, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x1a, 0x20, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x64, 0x65, 0x73, 0x63, 0x72, - 0x69, 0x70, 0x74, 0x6f, 0x72, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x22, 0x94, 0x01, 0x0a, 0x09, + 0x69, 0x70, 0x74, 0x6f, 0x72, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x22, 0xda, 0x01, 0x0a, 0x09, 0x46, 0x69, 0x65, 0x6c, 0x64, 0x49, 0x6e, 0x66, 0x6f, 0x12, 0x34, 0x0a, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0e, 0x32, 0x1c, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x46, 0x69, 0x65, 0x6c, 0x64, 0x49, 0x6e, 0x66, 0x6f, - 0x2e, 0x46, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x52, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x22, - 0x51, 0x0a, 0x06, 0x46, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x16, 0x0a, 0x12, 0x46, 0x4f, 0x52, - 0x4d, 0x41, 0x54, 0x5f, 0x55, 0x4e, 0x53, 0x50, 0x45, 0x43, 0x49, 0x46, 0x49, 0x45, 0x44, 0x10, - 0x00, 0x12, 0x09, 0x0a, 0x05, 0x55, 0x55, 0x49, 0x44, 0x34, 0x10, 0x01, 0x12, 0x08, 0x0a, 0x04, - 0x49, 0x50, 0x56, 0x34, 0x10, 0x02, 0x12, 0x08, 0x0a, 0x04, 0x49, 0x50, 0x56, 0x36, 0x10, 0x03, - 0x12, 0x10, 0x0a, 0x0c, 0x49, 0x50, 0x56, 0x34, 0x5f, 0x4f, 0x52, 0x5f, 0x49, 0x50, 0x56, 0x36, - 0x10, 0x04, 0x3a, 0x57, 0x0a, 0x0a, 0x66, 0x69, 0x65, 0x6c, 0x64, 0x5f, 0x69, 0x6e, 0x66, 0x6f, - 0x12, 0x1d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, - 0x75, 0x66, 0x2e, 0x46, 0x69, 0x65, 0x6c, 0x64, 0x4f, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x18, - 0xcc, 0xf1, 0xf9, 0x8a, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x15, 0x2e, 0x67, 0x6f, 0x6f, 0x67, - 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x46, 0x69, 0x65, 0x6c, 0x64, 0x49, 0x6e, 0x66, 0x6f, - 0x52, 0x09, 0x66, 0x69, 0x65, 0x6c, 0x64, 0x49, 0x6e, 0x66, 0x6f, 0x42, 0x6c, 0x0a, 0x0e, 0x63, - 0x6f, 0x6d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x42, 0x0e, 0x46, - 0x69, 0x65, 0x6c, 0x64, 0x49, 0x6e, 0x66, 0x6f, 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x50, 0x01, 0x5a, - 0x41, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, 0x2e, 0x6f, - 0x72, 0x67, 0x2f, 0x67, 0x65, 0x6e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x2f, 0x67, 0x6f, 0x6f, 0x67, - 0x6c, 0x65, 0x61, 0x70, 0x69, 0x73, 0x2f, 0x61, 0x70, 0x69, 0x2f, 0x61, 0x6e, 0x6e, 0x6f, 0x74, - 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x3b, 0x61, 0x6e, 0x6e, 0x6f, 0x74, 0x61, 0x74, 0x69, 0x6f, - 0x6e, 0x73, 0xa2, 0x02, 0x04, 0x47, 0x41, 0x50, 0x49, 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, - 0x33, + 0x2e, 0x46, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x52, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, + 0x44, 0x0a, 0x10, 0x72, 0x65, 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x64, 0x5f, 0x74, 0x79, + 0x70, 0x65, 0x73, 0x18, 0x02, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x19, 0x2e, 0x67, 0x6f, 0x6f, 0x67, + 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x54, 0x79, 0x70, 0x65, 0x52, 0x65, 0x66, 0x65, 0x72, + 0x65, 0x6e, 0x63, 0x65, 0x52, 0x0f, 0x72, 0x65, 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x64, + 0x54, 0x79, 0x70, 0x65, 0x73, 0x22, 0x51, 0x0a, 0x06, 0x46, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, + 0x16, 0x0a, 0x12, 0x46, 0x4f, 0x52, 0x4d, 0x41, 0x54, 0x5f, 0x55, 0x4e, 0x53, 0x50, 0x45, 0x43, + 0x49, 0x46, 0x49, 0x45, 0x44, 0x10, 0x00, 0x12, 0x09, 0x0a, 0x05, 0x55, 0x55, 0x49, 0x44, 0x34, + 0x10, 0x01, 0x12, 0x08, 0x0a, 0x04, 0x49, 0x50, 0x56, 0x34, 0x10, 0x02, 0x12, 0x08, 0x0a, 0x04, + 0x49, 0x50, 0x56, 0x36, 0x10, 0x03, 0x12, 0x10, 0x0a, 0x0c, 0x49, 0x50, 0x56, 0x34, 0x5f, 0x4f, + 0x52, 0x5f, 0x49, 0x50, 0x56, 0x36, 0x10, 0x04, 0x22, 0x2c, 0x0a, 0x0d, 0x54, 0x79, 0x70, 0x65, + 0x52, 0x65, 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x12, 0x1b, 0x0a, 0x09, 0x74, 0x79, 0x70, + 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x08, 0x74, 0x79, + 0x70, 0x65, 0x4e, 0x61, 0x6d, 0x65, 0x3a, 0x57, 0x0a, 0x0a, 0x66, 0x69, 0x65, 0x6c, 0x64, 0x5f, + 0x69, 0x6e, 0x66, 0x6f, 0x12, 0x1d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, + 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x46, 0x69, 0x65, 0x6c, 0x64, 0x4f, 0x70, 0x74, 0x69, + 0x6f, 0x6e, 0x73, 0x18, 0xcc, 0xf1, 0xf9, 0x8a, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x15, 0x2e, + 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x46, 0x69, 0x65, 0x6c, 0x64, + 0x49, 0x6e, 0x66, 0x6f, 0x52, 0x09, 0x66, 0x69, 0x65, 0x6c, 0x64, 0x49, 0x6e, 0x66, 0x6f, 0x42, + 0x6c, 0x0a, 0x0e, 0x63, 0x6f, 0x6d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, + 0x69, 0x42, 0x0e, 0x46, 0x69, 0x65, 0x6c, 0x64, 0x49, 0x6e, 0x66, 0x6f, 0x50, 0x72, 0x6f, 0x74, + 0x6f, 0x50, 0x01, 0x5a, 0x41, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x67, 0x6f, 0x6c, 0x61, + 0x6e, 0x67, 0x2e, 0x6f, 0x72, 0x67, 0x2f, 0x67, 0x65, 0x6e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x2f, + 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x61, 0x70, 0x69, 0x73, 0x2f, 0x61, 0x70, 0x69, 0x2f, 0x61, + 0x6e, 0x6e, 0x6f, 0x74, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x3b, 0x61, 0x6e, 0x6e, 0x6f, 0x74, + 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0xa2, 0x02, 0x04, 0x47, 0x41, 0x50, 0x49, 0x62, 0x06, 0x70, + 0x72, 0x6f, 0x74, 0x6f, 0x33, } var ( @@ -236,21 +319,23 @@ func file_google_api_field_info_proto_rawDescGZIP() []byte { } var file_google_api_field_info_proto_enumTypes = make([]protoimpl.EnumInfo, 1) -var file_google_api_field_info_proto_msgTypes = make([]protoimpl.MessageInfo, 1) +var file_google_api_field_info_proto_msgTypes = make([]protoimpl.MessageInfo, 2) var file_google_api_field_info_proto_goTypes = []interface{}{ (FieldInfo_Format)(0), // 0: google.api.FieldInfo.Format (*FieldInfo)(nil), // 1: google.api.FieldInfo - (*descriptorpb.FieldOptions)(nil), // 2: google.protobuf.FieldOptions + (*TypeReference)(nil), // 2: google.api.TypeReference + (*descriptorpb.FieldOptions)(nil), // 3: google.protobuf.FieldOptions } var file_google_api_field_info_proto_depIdxs = []int32{ 0, // 0: google.api.FieldInfo.format:type_name -> google.api.FieldInfo.Format - 2, // 1: google.api.field_info:extendee -> google.protobuf.FieldOptions - 1, // 2: google.api.field_info:type_name -> google.api.FieldInfo - 3, // [3:3] is the sub-list for method output_type - 3, // [3:3] is the sub-list for method input_type - 2, // [2:3] is the sub-list for extension type_name - 1, // [1:2] is the sub-list for extension extendee - 0, // [0:1] is the sub-list for field type_name + 2, // 1: google.api.FieldInfo.referenced_types:type_name -> google.api.TypeReference + 3, // 2: google.api.field_info:extendee -> google.protobuf.FieldOptions + 1, // 3: google.api.field_info:type_name -> google.api.FieldInfo + 4, // [4:4] is the sub-list for method output_type + 4, // [4:4] is the sub-list for method input_type + 3, // [3:4] is the sub-list for extension type_name + 2, // [2:3] is the sub-list for extension extendee + 0, // [0:2] is the sub-list for field type_name } func init() { file_google_api_field_info_proto_init() } @@ -271,6 +356,18 @@ func file_google_api_field_info_proto_init() { return nil } } + file_google_api_field_info_proto_msgTypes[1].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*TypeReference); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } } type x struct{} out := protoimpl.TypeBuilder{ @@ -278,7 +375,7 @@ func file_google_api_field_info_proto_init() { GoPackagePath: reflect.TypeOf(x{}).PkgPath(), RawDescriptor: file_google_api_field_info_proto_rawDesc, NumEnums: 1, - NumMessages: 1, + NumMessages: 2, NumExtensions: 1, NumServices: 0, }, diff --git a/vendor/google.golang.org/genproto/googleapis/api/annotations/http.pb.go b/vendor/google.golang.org/genproto/googleapis/api/annotations/http.pb.go index 76ea76df330fa..ffb5838cb18be 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/annotations/http.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/annotations/http.pb.go @@ -102,7 +102,7 @@ func (x *Http) GetFullyDecodeReservedExpansion() bool { return false } -// # gRPC Transcoding +// gRPC Transcoding // // gRPC Transcoding is a feature for mapping between a gRPC method and one or // more HTTP REST endpoints. It allows developers to build a single API service @@ -143,9 +143,8 @@ func (x *Http) GetFullyDecodeReservedExpansion() bool { // // This enables an HTTP REST to gRPC mapping as below: // -// HTTP | gRPC -// -----|----- -// `GET /v1/messages/123456` | `GetMessage(name: "messages/123456")` +// - HTTP: `GET /v1/messages/123456` +// - gRPC: `GetMessage(name: "messages/123456")` // // Any fields in the request message which are not bound by the path template // automatically become HTTP query parameters if there is no HTTP request body. @@ -169,11 +168,9 @@ func (x *Http) GetFullyDecodeReservedExpansion() bool { // // This enables a HTTP JSON to RPC mapping as below: // -// HTTP | gRPC -// -----|----- -// `GET /v1/messages/123456?revision=2&sub.subfield=foo` | -// `GetMessage(message_id: "123456" revision: 2 sub: SubMessage(subfield: -// "foo"))` +// - HTTP: `GET /v1/messages/123456?revision=2&sub.subfield=foo` +// - gRPC: `GetMessage(message_id: "123456" revision: 2 sub: +// SubMessage(subfield: "foo"))` // // Note that fields which are mapped to URL query parameters must have a // primitive type or a repeated primitive type or a non-repeated message type. @@ -203,10 +200,8 @@ func (x *Http) GetFullyDecodeReservedExpansion() bool { // representation of the JSON in the request body is determined by // protos JSON encoding: // -// HTTP | gRPC -// -----|----- -// `PATCH /v1/messages/123456 { "text": "Hi!" }` | `UpdateMessage(message_id: -// "123456" message { text: "Hi!" })` +// - HTTP: `PATCH /v1/messages/123456 { "text": "Hi!" }` +// - gRPC: `UpdateMessage(message_id: "123456" message { text: "Hi!" })` // // The special name `*` can be used in the body mapping to define that // every field not bound by the path template should be mapped to the @@ -228,10 +223,8 @@ func (x *Http) GetFullyDecodeReservedExpansion() bool { // // The following HTTP JSON to RPC mapping is enabled: // -// HTTP | gRPC -// -----|----- -// `PATCH /v1/messages/123456 { "text": "Hi!" }` | `UpdateMessage(message_id: -// "123456" text: "Hi!")` +// - HTTP: `PATCH /v1/messages/123456 { "text": "Hi!" }` +// - gRPC: `UpdateMessage(message_id: "123456" text: "Hi!")` // // Note that when using `*` in the body mapping, it is not possible to // have HTTP parameters, as all fields not bound by the path end in @@ -259,13 +252,13 @@ func (x *Http) GetFullyDecodeReservedExpansion() bool { // // This enables the following two alternative HTTP JSON to RPC mappings: // -// HTTP | gRPC -// -----|----- -// `GET /v1/messages/123456` | `GetMessage(message_id: "123456")` -// `GET /v1/users/me/messages/123456` | `GetMessage(user_id: "me" message_id: -// "123456")` +// - HTTP: `GET /v1/messages/123456` +// - gRPC: `GetMessage(message_id: "123456")` // -// ## Rules for HTTP mapping +// - HTTP: `GET /v1/users/me/messages/123456` +// - gRPC: `GetMessage(user_id: "me" message_id: "123456")` +// +// # Rules for HTTP mapping // // 1. Leaf request fields (recursive expansion nested messages in the request // message) are classified into three categories: @@ -284,7 +277,7 @@ func (x *Http) GetFullyDecodeReservedExpansion() bool { // request body, all // fields are passed via URL path and URL query parameters. // -// ### Path template syntax +// Path template syntax // // Template = "/" Segments [ Verb ] ; // Segments = Segment { "/" Segment } ; @@ -323,7 +316,7 @@ func (x *Http) GetFullyDecodeReservedExpansion() bool { // Document](https://developers.google.com/discovery/v1/reference/apis) as // `{+var}`. // -// ## Using gRPC API Service Configuration +// # Using gRPC API Service Configuration // // gRPC API Service Configuration (service config) is a configuration language // for configuring a gRPC service to become a user-facing product. The @@ -338,15 +331,14 @@ func (x *Http) GetFullyDecodeReservedExpansion() bool { // specified in the service config will override any matching transcoding // configuration in the proto. // -// Example: +// The following example selects a gRPC method and applies an `HttpRule` to it: // // http: // rules: -// # Selects a gRPC method and applies HttpRule to it. // - selector: example.v1.Messaging.GetMessage // get: /v1/messages/{message_id}/{sub.subfield} // -// ## Special notes +// # Special notes // // When gRPC Transcoding is used to map a gRPC to JSON REST endpoints, the // proto to JSON conversion must follow the [proto3 diff --git a/vendor/google.golang.org/genproto/googleapis/api/annotations/resource.pb.go b/vendor/google.golang.org/genproto/googleapis/api/annotations/resource.pb.go index 7a3fd93fcd96d..b5db279aebf6c 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/annotations/resource.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/annotations/resource.pb.go @@ -253,8 +253,13 @@ type ResourceDescriptor struct { History ResourceDescriptor_History `protobuf:"varint,4,opt,name=history,proto3,enum=google.api.ResourceDescriptor_History" json:"history,omitempty"` // The plural name used in the resource name and permission names, such as // 'projects' for the resource name of 'projects/{project}' and the permission - // name of 'cloudresourcemanager.googleapis.com/projects.get'. It is the same - // concept of the `plural` field in k8s CRD spec + // name of 'cloudresourcemanager.googleapis.com/projects.get'. One exception + // to this is for Nested Collections that have stuttering names, as defined + // in [AIP-122](https://google.aip.dev/122#nested-collections), where the + // collection ID in the resource name pattern does not necessarily directly + // match the `plural` value. + // + // It is the same concept of the `plural` field in k8s CRD spec // https://kubernetes.io/docs/tasks/access-kubernetes-api/custom-resources/custom-resource-definitions/ // // Note: The plural form is required even for singleton resources. See diff --git a/vendor/google.golang.org/genproto/googleapis/api/expr/v1alpha1/syntax.pb.go b/vendor/google.golang.org/genproto/googleapis/api/expr/v1alpha1/syntax.pb.go index 6b867a46ede9f..c90c6015d26c3 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/expr/v1alpha1/syntax.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/expr/v1alpha1/syntax.pb.go @@ -1105,25 +1105,66 @@ func (x *Expr_CreateStruct) GetEntries() []*Expr_CreateStruct_Entry { // messages `has(m.x)` is defined as 'defined, but not set`. For proto3, the // macro tests whether the property is set to its default. For map and struct // types, the macro tests whether the property `x` is defined on `m`. +// +// Comprehensions for the standard environment macros evaluation can be best +// visualized as the following pseudocode: +// +// ``` +// let `accu_var` = `accu_init` +// +// for (let `iter_var` in `iter_range`) { +// if (!`loop_condition`) { +// break +// } +// `accu_var` = `loop_step` +// } +// +// return `result` +// ``` +// +// Comprehensions for the optional V2 macros which support map-to-map +// translation differ slightly from the standard environment macros in that +// they expose both the key or index in addition to the value for each list +// or map entry: +// +// ``` +// let `accu_var` = `accu_init` +// +// for (let `iter_var`, `iter_var2` in `iter_range`) { +// if (!`loop_condition`) { +// break +// } +// `accu_var` = `loop_step` +// } +// +// return `result` +// ``` type Expr_Comprehension struct { state protoimpl.MessageState sizeCache protoimpl.SizeCache unknownFields protoimpl.UnknownFields - // The name of the iteration variable. + // The name of the first iteration variable. + // When the iter_range is a list, this variable is the list element. + // When the iter_range is a map, this variable is the map entry key. IterVar string `protobuf:"bytes,1,opt,name=iter_var,json=iterVar,proto3" json:"iter_var,omitempty"` - // The range over which var iterates. + // The name of the second iteration variable, empty if not set. + // When the iter_range is a list, this variable is the integer index. + // When the iter_range is a map, this variable is the map entry value. + // This field is only set for comprehension v2 macros. + IterVar2 string `protobuf:"bytes,8,opt,name=iter_var2,json=iterVar2,proto3" json:"iter_var2,omitempty"` + // The range over which the comprehension iterates. IterRange *Expr `protobuf:"bytes,2,opt,name=iter_range,json=iterRange,proto3" json:"iter_range,omitempty"` // The name of the variable used for accumulation of the result. AccuVar string `protobuf:"bytes,3,opt,name=accu_var,json=accuVar,proto3" json:"accu_var,omitempty"` // The initial value of the accumulator. AccuInit *Expr `protobuf:"bytes,4,opt,name=accu_init,json=accuInit,proto3" json:"accu_init,omitempty"` - // An expression which can contain iter_var and accu_var. + // An expression which can contain iter_var, iter_var2, and accu_var. // // Returns false when the result has been computed and may be used as // a hint to short-circuit the remainder of the comprehension. LoopCondition *Expr `protobuf:"bytes,5,opt,name=loop_condition,json=loopCondition,proto3" json:"loop_condition,omitempty"` - // An expression which can contain iter_var and accu_var. + // An expression which can contain iter_var, iter_var2, and accu_var. // // Computes the next value of accu_var. LoopStep *Expr `protobuf:"bytes,6,opt,name=loop_step,json=loopStep,proto3" json:"loop_step,omitempty"` @@ -1172,6 +1213,13 @@ func (x *Expr_Comprehension) GetIterVar() string { return "" } +func (x *Expr_Comprehension) GetIterVar2() string { + if x != nil { + return x.IterVar2 + } + return "" +} + func (x *Expr_Comprehension) GetIterRange() *Expr { if x != nil { return x.IterRange @@ -1485,7 +1533,7 @@ var file_google_api_expr_v1alpha1_syntax_proto_rawDesc = []byte{ 0x66, 0x6f, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x24, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x52, 0x0a, - 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x22, 0xae, 0x0d, 0x0a, 0x04, 0x45, + 0x73, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x22, 0xcb, 0x0d, 0x0a, 0x04, 0x45, 0x78, 0x70, 0x72, 0x12, 0x0e, 0x0a, 0x02, 0x69, 0x64, 0x18, 0x02, 0x20, 0x01, 0x28, 0x03, 0x52, 0x02, 0x69, 0x64, 0x12, 0x43, 0x0a, 0x0a, 0x63, 0x6f, 0x6e, 0x73, 0x74, 0x5f, 0x65, 0x78, 0x70, 0x72, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, @@ -1567,132 +1615,134 @@ var file_google_api_expr_v1alpha1_syntax_proto_rawDesc = []byte{ 0x45, 0x78, 0x70, 0x72, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x25, 0x0a, 0x0e, 0x6f, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, 0x65, 0x6e, 0x74, 0x72, 0x79, 0x18, 0x05, 0x20, 0x01, 0x28, 0x08, 0x52, 0x0d, 0x6f, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x45, 0x6e, 0x74, - 0x72, 0x79, 0x42, 0x0a, 0x0a, 0x08, 0x6b, 0x65, 0x79, 0x5f, 0x6b, 0x69, 0x6e, 0x64, 0x1a, 0xfd, - 0x02, 0x0a, 0x0d, 0x43, 0x6f, 0x6d, 0x70, 0x72, 0x65, 0x68, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, + 0x72, 0x79, 0x42, 0x0a, 0x0a, 0x08, 0x6b, 0x65, 0x79, 0x5f, 0x6b, 0x69, 0x6e, 0x64, 0x1a, 0x9a, + 0x03, 0x0a, 0x0d, 0x43, 0x6f, 0x6d, 0x70, 0x72, 0x65, 0x68, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x19, 0x0a, 0x08, 0x69, 0x74, 0x65, 0x72, 0x5f, 0x76, 0x61, 0x72, 0x18, 0x01, 0x20, 0x01, - 0x28, 0x09, 0x52, 0x07, 0x69, 0x74, 0x65, 0x72, 0x56, 0x61, 0x72, 0x12, 0x3d, 0x0a, 0x0a, 0x69, - 0x74, 0x65, 0x72, 0x5f, 0x72, 0x61, 0x6e, 0x67, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, - 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, - 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x45, 0x78, 0x70, 0x72, 0x52, - 0x09, 0x69, 0x74, 0x65, 0x72, 0x52, 0x61, 0x6e, 0x67, 0x65, 0x12, 0x19, 0x0a, 0x08, 0x61, 0x63, - 0x63, 0x75, 0x5f, 0x76, 0x61, 0x72, 0x18, 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x61, 0x63, - 0x63, 0x75, 0x56, 0x61, 0x72, 0x12, 0x3b, 0x0a, 0x09, 0x61, 0x63, 0x63, 0x75, 0x5f, 0x69, 0x6e, - 0x69, 0x74, 0x18, 0x04, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, - 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, - 0x68, 0x61, 0x31, 0x2e, 0x45, 0x78, 0x70, 0x72, 0x52, 0x08, 0x61, 0x63, 0x63, 0x75, 0x49, 0x6e, - 0x69, 0x74, 0x12, 0x45, 0x0a, 0x0e, 0x6c, 0x6f, 0x6f, 0x70, 0x5f, 0x63, 0x6f, 0x6e, 0x64, 0x69, - 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x05, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, - 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, - 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x45, 0x78, 0x70, 0x72, 0x52, 0x0d, 0x6c, 0x6f, 0x6f, 0x70, - 0x43, 0x6f, 0x6e, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x3b, 0x0a, 0x09, 0x6c, 0x6f, 0x6f, - 0x70, 0x5f, 0x73, 0x74, 0x65, 0x70, 0x18, 0x06, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x67, + 0x28, 0x09, 0x52, 0x07, 0x69, 0x74, 0x65, 0x72, 0x56, 0x61, 0x72, 0x12, 0x1b, 0x0a, 0x09, 0x69, + 0x74, 0x65, 0x72, 0x5f, 0x76, 0x61, 0x72, 0x32, 0x18, 0x08, 0x20, 0x01, 0x28, 0x09, 0x52, 0x08, + 0x69, 0x74, 0x65, 0x72, 0x56, 0x61, 0x72, 0x32, 0x12, 0x3d, 0x0a, 0x0a, 0x69, 0x74, 0x65, 0x72, + 0x5f, 0x72, 0x61, 0x6e, 0x67, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, - 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x45, 0x78, 0x70, 0x72, 0x52, 0x08, 0x6c, 0x6f, - 0x6f, 0x70, 0x53, 0x74, 0x65, 0x70, 0x12, 0x36, 0x0a, 0x06, 0x72, 0x65, 0x73, 0x75, 0x6c, 0x74, - 0x18, 0x07, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, - 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, - 0x31, 0x2e, 0x45, 0x78, 0x70, 0x72, 0x52, 0x06, 0x72, 0x65, 0x73, 0x75, 0x6c, 0x74, 0x42, 0x0b, - 0x0a, 0x09, 0x65, 0x78, 0x70, 0x72, 0x5f, 0x6b, 0x69, 0x6e, 0x64, 0x22, 0xc1, 0x03, 0x0a, 0x08, - 0x43, 0x6f, 0x6e, 0x73, 0x74, 0x61, 0x6e, 0x74, 0x12, 0x3b, 0x0a, 0x0a, 0x6e, 0x75, 0x6c, 0x6c, - 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0e, 0x32, 0x1a, 0x2e, 0x67, - 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x4e, - 0x75, 0x6c, 0x6c, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x48, 0x00, 0x52, 0x09, 0x6e, 0x75, 0x6c, 0x6c, - 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x1f, 0x0a, 0x0a, 0x62, 0x6f, 0x6f, 0x6c, 0x5f, 0x76, 0x61, - 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x08, 0x48, 0x00, 0x52, 0x09, 0x62, 0x6f, 0x6f, - 0x6c, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x21, 0x0a, 0x0b, 0x69, 0x6e, 0x74, 0x36, 0x34, 0x5f, - 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x03, 0x20, 0x01, 0x28, 0x03, 0x48, 0x00, 0x52, 0x0a, 0x69, - 0x6e, 0x74, 0x36, 0x34, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x23, 0x0a, 0x0c, 0x75, 0x69, 0x6e, - 0x74, 0x36, 0x34, 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x04, 0x20, 0x01, 0x28, 0x04, 0x48, - 0x00, 0x52, 0x0b, 0x75, 0x69, 0x6e, 0x74, 0x36, 0x34, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x23, - 0x0a, 0x0c, 0x64, 0x6f, 0x75, 0x62, 0x6c, 0x65, 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x05, - 0x20, 0x01, 0x28, 0x01, 0x48, 0x00, 0x52, 0x0b, 0x64, 0x6f, 0x75, 0x62, 0x6c, 0x65, 0x56, 0x61, - 0x6c, 0x75, 0x65, 0x12, 0x23, 0x0a, 0x0c, 0x73, 0x74, 0x72, 0x69, 0x6e, 0x67, 0x5f, 0x76, 0x61, - 0x6c, 0x75, 0x65, 0x18, 0x06, 0x20, 0x01, 0x28, 0x09, 0x48, 0x00, 0x52, 0x0b, 0x73, 0x74, 0x72, - 0x69, 0x6e, 0x67, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x21, 0x0a, 0x0b, 0x62, 0x79, 0x74, 0x65, - 0x73, 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x07, 0x20, 0x01, 0x28, 0x0c, 0x48, 0x00, 0x52, - 0x0a, 0x62, 0x79, 0x74, 0x65, 0x73, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x46, 0x0a, 0x0e, 0x64, - 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x08, 0x20, - 0x01, 0x28, 0x0b, 0x32, 0x19, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, - 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x44, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x42, 0x02, - 0x18, 0x01, 0x48, 0x00, 0x52, 0x0d, 0x64, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x56, 0x61, - 0x6c, 0x75, 0x65, 0x12, 0x49, 0x0a, 0x0f, 0x74, 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, - 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x09, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1a, 0x2e, 0x67, - 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x54, - 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x42, 0x02, 0x18, 0x01, 0x48, 0x00, 0x52, 0x0e, - 0x74, 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x42, 0x0f, - 0x0a, 0x0d, 0x63, 0x6f, 0x6e, 0x73, 0x74, 0x61, 0x6e, 0x74, 0x5f, 0x6b, 0x69, 0x6e, 0x64, 0x22, - 0x8c, 0x07, 0x0a, 0x0a, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x12, 0x25, - 0x0a, 0x0e, 0x73, 0x79, 0x6e, 0x74, 0x61, 0x78, 0x5f, 0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, - 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0d, 0x73, 0x79, 0x6e, 0x74, 0x61, 0x78, 0x56, 0x65, - 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x1a, 0x0a, 0x08, 0x6c, 0x6f, 0x63, 0x61, 0x74, 0x69, 0x6f, - 0x6e, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x08, 0x6c, 0x6f, 0x63, 0x61, 0x74, 0x69, 0x6f, - 0x6e, 0x12, 0x21, 0x0a, 0x0c, 0x6c, 0x69, 0x6e, 0x65, 0x5f, 0x6f, 0x66, 0x66, 0x73, 0x65, 0x74, - 0x73, 0x18, 0x03, 0x20, 0x03, 0x28, 0x05, 0x52, 0x0b, 0x6c, 0x69, 0x6e, 0x65, 0x4f, 0x66, 0x66, - 0x73, 0x65, 0x74, 0x73, 0x12, 0x51, 0x0a, 0x09, 0x70, 0x6f, 0x73, 0x69, 0x74, 0x69, 0x6f, 0x6e, - 0x73, 0x18, 0x04, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x33, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, + 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x45, 0x78, 0x70, 0x72, 0x52, 0x09, 0x69, 0x74, + 0x65, 0x72, 0x52, 0x61, 0x6e, 0x67, 0x65, 0x12, 0x19, 0x0a, 0x08, 0x61, 0x63, 0x63, 0x75, 0x5f, + 0x76, 0x61, 0x72, 0x18, 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x61, 0x63, 0x63, 0x75, 0x56, + 0x61, 0x72, 0x12, 0x3b, 0x0a, 0x09, 0x61, 0x63, 0x63, 0x75, 0x5f, 0x69, 0x6e, 0x69, 0x74, 0x18, + 0x04, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, + 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, + 0x2e, 0x45, 0x78, 0x70, 0x72, 0x52, 0x08, 0x61, 0x63, 0x63, 0x75, 0x49, 0x6e, 0x69, 0x74, 0x12, + 0x45, 0x0a, 0x0e, 0x6c, 0x6f, 0x6f, 0x70, 0x5f, 0x63, 0x6f, 0x6e, 0x64, 0x69, 0x74, 0x69, 0x6f, + 0x6e, 0x18, 0x05, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, - 0x61, 0x31, 0x2e, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x2e, 0x50, 0x6f, - 0x73, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, 0x52, 0x09, 0x70, 0x6f, - 0x73, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x12, 0x55, 0x0a, 0x0b, 0x6d, 0x61, 0x63, 0x72, 0x6f, - 0x5f, 0x63, 0x61, 0x6c, 0x6c, 0x73, 0x18, 0x05, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x34, 0x2e, 0x67, - 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, - 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, - 0x66, 0x6f, 0x2e, 0x4d, 0x61, 0x63, 0x72, 0x6f, 0x43, 0x61, 0x6c, 0x6c, 0x73, 0x45, 0x6e, 0x74, - 0x72, 0x79, 0x52, 0x0a, 0x6d, 0x61, 0x63, 0x72, 0x6f, 0x43, 0x61, 0x6c, 0x6c, 0x73, 0x12, 0x4e, - 0x0a, 0x0a, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x73, 0x18, 0x06, 0x20, 0x03, - 0x28, 0x0b, 0x32, 0x2e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, - 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x53, 0x6f, - 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x2e, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, - 0x6f, 0x6e, 0x52, 0x0a, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x73, 0x1a, 0x80, - 0x03, 0x0a, 0x09, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x0e, 0x0a, 0x02, - 0x69, 0x64, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x02, 0x69, 0x64, 0x12, 0x69, 0x0a, 0x13, - 0x61, 0x66, 0x66, 0x65, 0x63, 0x74, 0x65, 0x64, 0x5f, 0x63, 0x6f, 0x6d, 0x70, 0x6f, 0x6e, 0x65, - 0x6e, 0x74, 0x73, 0x18, 0x02, 0x20, 0x03, 0x28, 0x0e, 0x32, 0x38, 0x2e, 0x67, 0x6f, 0x6f, 0x67, + 0x61, 0x31, 0x2e, 0x45, 0x78, 0x70, 0x72, 0x52, 0x0d, 0x6c, 0x6f, 0x6f, 0x70, 0x43, 0x6f, 0x6e, + 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x3b, 0x0a, 0x09, 0x6c, 0x6f, 0x6f, 0x70, 0x5f, 0x73, + 0x74, 0x65, 0x70, 0x18, 0x06, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, - 0x70, 0x68, 0x61, 0x31, 0x2e, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x2e, - 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x2e, 0x43, 0x6f, 0x6d, 0x70, 0x6f, 0x6e, - 0x65, 0x6e, 0x74, 0x52, 0x12, 0x61, 0x66, 0x66, 0x65, 0x63, 0x74, 0x65, 0x64, 0x43, 0x6f, 0x6d, - 0x70, 0x6f, 0x6e, 0x65, 0x6e, 0x74, 0x73, 0x12, 0x50, 0x0a, 0x07, 0x76, 0x65, 0x72, 0x73, 0x69, - 0x6f, 0x6e, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x36, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, - 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, - 0x68, 0x61, 0x31, 0x2e, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x2e, 0x45, - 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x2e, 0x56, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, - 0x52, 0x07, 0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x1a, 0x35, 0x0a, 0x07, 0x56, 0x65, 0x72, - 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x14, 0x0a, 0x05, 0x6d, 0x61, 0x6a, 0x6f, 0x72, 0x18, 0x01, 0x20, - 0x01, 0x28, 0x03, 0x52, 0x05, 0x6d, 0x61, 0x6a, 0x6f, 0x72, 0x12, 0x14, 0x0a, 0x05, 0x6d, 0x69, - 0x6e, 0x6f, 0x72, 0x18, 0x02, 0x20, 0x01, 0x28, 0x03, 0x52, 0x05, 0x6d, 0x69, 0x6e, 0x6f, 0x72, - 0x22, 0x6f, 0x0a, 0x09, 0x43, 0x6f, 0x6d, 0x70, 0x6f, 0x6e, 0x65, 0x6e, 0x74, 0x12, 0x19, 0x0a, - 0x15, 0x43, 0x4f, 0x4d, 0x50, 0x4f, 0x4e, 0x45, 0x4e, 0x54, 0x5f, 0x55, 0x4e, 0x53, 0x50, 0x45, - 0x43, 0x49, 0x46, 0x49, 0x45, 0x44, 0x10, 0x00, 0x12, 0x14, 0x0a, 0x10, 0x43, 0x4f, 0x4d, 0x50, - 0x4f, 0x4e, 0x45, 0x4e, 0x54, 0x5f, 0x50, 0x41, 0x52, 0x53, 0x45, 0x52, 0x10, 0x01, 0x12, 0x1a, - 0x0a, 0x16, 0x43, 0x4f, 0x4d, 0x50, 0x4f, 0x4e, 0x45, 0x4e, 0x54, 0x5f, 0x54, 0x59, 0x50, 0x45, - 0x5f, 0x43, 0x48, 0x45, 0x43, 0x4b, 0x45, 0x52, 0x10, 0x02, 0x12, 0x15, 0x0a, 0x11, 0x43, 0x4f, - 0x4d, 0x50, 0x4f, 0x4e, 0x45, 0x4e, 0x54, 0x5f, 0x52, 0x55, 0x4e, 0x54, 0x49, 0x4d, 0x45, 0x10, - 0x03, 0x1a, 0x3c, 0x0a, 0x0e, 0x50, 0x6f, 0x73, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x45, 0x6e, - 0x74, 0x72, 0x79, 0x12, 0x10, 0x0a, 0x03, 0x6b, 0x65, 0x79, 0x18, 0x01, 0x20, 0x01, 0x28, 0x03, - 0x52, 0x03, 0x6b, 0x65, 0x79, 0x12, 0x14, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, - 0x20, 0x01, 0x28, 0x05, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x3a, 0x02, 0x38, 0x01, 0x1a, - 0x5d, 0x0a, 0x0f, 0x4d, 0x61, 0x63, 0x72, 0x6f, 0x43, 0x61, 0x6c, 0x6c, 0x73, 0x45, 0x6e, 0x74, - 0x72, 0x79, 0x12, 0x10, 0x0a, 0x03, 0x6b, 0x65, 0x79, 0x18, 0x01, 0x20, 0x01, 0x28, 0x03, 0x52, - 0x03, 0x6b, 0x65, 0x79, 0x12, 0x34, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, + 0x70, 0x68, 0x61, 0x31, 0x2e, 0x45, 0x78, 0x70, 0x72, 0x52, 0x08, 0x6c, 0x6f, 0x6f, 0x70, 0x53, + 0x74, 0x65, 0x70, 0x12, 0x36, 0x0a, 0x06, 0x72, 0x65, 0x73, 0x75, 0x6c, 0x74, 0x18, 0x07, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x45, - 0x78, 0x70, 0x72, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x3a, 0x02, 0x38, 0x01, 0x22, 0x70, - 0x0a, 0x0e, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x50, 0x6f, 0x73, 0x69, 0x74, 0x69, 0x6f, 0x6e, - 0x12, 0x1a, 0x0a, 0x08, 0x6c, 0x6f, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, - 0x28, 0x09, 0x52, 0x08, 0x6c, 0x6f, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x16, 0x0a, 0x06, - 0x6f, 0x66, 0x66, 0x73, 0x65, 0x74, 0x18, 0x02, 0x20, 0x01, 0x28, 0x05, 0x52, 0x06, 0x6f, 0x66, - 0x66, 0x73, 0x65, 0x74, 0x12, 0x12, 0x0a, 0x04, 0x6c, 0x69, 0x6e, 0x65, 0x18, 0x03, 0x20, 0x01, - 0x28, 0x05, 0x52, 0x04, 0x6c, 0x69, 0x6e, 0x65, 0x12, 0x16, 0x0a, 0x06, 0x63, 0x6f, 0x6c, 0x75, - 0x6d, 0x6e, 0x18, 0x04, 0x20, 0x01, 0x28, 0x05, 0x52, 0x06, 0x63, 0x6f, 0x6c, 0x75, 0x6d, 0x6e, - 0x42, 0x6e, 0x0a, 0x1c, 0x63, 0x6f, 0x6d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, + 0x78, 0x70, 0x72, 0x52, 0x06, 0x72, 0x65, 0x73, 0x75, 0x6c, 0x74, 0x42, 0x0b, 0x0a, 0x09, 0x65, + 0x78, 0x70, 0x72, 0x5f, 0x6b, 0x69, 0x6e, 0x64, 0x22, 0xc1, 0x03, 0x0a, 0x08, 0x43, 0x6f, 0x6e, + 0x73, 0x74, 0x61, 0x6e, 0x74, 0x12, 0x3b, 0x0a, 0x0a, 0x6e, 0x75, 0x6c, 0x6c, 0x5f, 0x76, 0x61, + 0x6c, 0x75, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0e, 0x32, 0x1a, 0x2e, 0x67, 0x6f, 0x6f, 0x67, + 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x4e, 0x75, 0x6c, 0x6c, + 0x56, 0x61, 0x6c, 0x75, 0x65, 0x48, 0x00, 0x52, 0x09, 0x6e, 0x75, 0x6c, 0x6c, 0x56, 0x61, 0x6c, + 0x75, 0x65, 0x12, 0x1f, 0x0a, 0x0a, 0x62, 0x6f, 0x6f, 0x6c, 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, + 0x18, 0x02, 0x20, 0x01, 0x28, 0x08, 0x48, 0x00, 0x52, 0x09, 0x62, 0x6f, 0x6f, 0x6c, 0x56, 0x61, + 0x6c, 0x75, 0x65, 0x12, 0x21, 0x0a, 0x0b, 0x69, 0x6e, 0x74, 0x36, 0x34, 0x5f, 0x76, 0x61, 0x6c, + 0x75, 0x65, 0x18, 0x03, 0x20, 0x01, 0x28, 0x03, 0x48, 0x00, 0x52, 0x0a, 0x69, 0x6e, 0x74, 0x36, + 0x34, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x23, 0x0a, 0x0c, 0x75, 0x69, 0x6e, 0x74, 0x36, 0x34, + 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x04, 0x20, 0x01, 0x28, 0x04, 0x48, 0x00, 0x52, 0x0b, + 0x75, 0x69, 0x6e, 0x74, 0x36, 0x34, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x23, 0x0a, 0x0c, 0x64, + 0x6f, 0x75, 0x62, 0x6c, 0x65, 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x05, 0x20, 0x01, 0x28, + 0x01, 0x48, 0x00, 0x52, 0x0b, 0x64, 0x6f, 0x75, 0x62, 0x6c, 0x65, 0x56, 0x61, 0x6c, 0x75, 0x65, + 0x12, 0x23, 0x0a, 0x0c, 0x73, 0x74, 0x72, 0x69, 0x6e, 0x67, 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, + 0x18, 0x06, 0x20, 0x01, 0x28, 0x09, 0x48, 0x00, 0x52, 0x0b, 0x73, 0x74, 0x72, 0x69, 0x6e, 0x67, + 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x21, 0x0a, 0x0b, 0x62, 0x79, 0x74, 0x65, 0x73, 0x5f, 0x76, + 0x61, 0x6c, 0x75, 0x65, 0x18, 0x07, 0x20, 0x01, 0x28, 0x0c, 0x48, 0x00, 0x52, 0x0a, 0x62, 0x79, + 0x74, 0x65, 0x73, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x46, 0x0a, 0x0e, 0x64, 0x75, 0x72, 0x61, + 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x08, 0x20, 0x01, 0x28, 0x0b, + 0x32, 0x19, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, + 0x75, 0x66, 0x2e, 0x44, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x42, 0x02, 0x18, 0x01, 0x48, + 0x00, 0x52, 0x0d, 0x64, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x56, 0x61, 0x6c, 0x75, 0x65, + 0x12, 0x49, 0x0a, 0x0f, 0x74, 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x5f, 0x76, 0x61, + 0x6c, 0x75, 0x65, 0x18, 0x09, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1a, 0x2e, 0x67, 0x6f, 0x6f, 0x67, + 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x54, 0x69, 0x6d, 0x65, + 0x73, 0x74, 0x61, 0x6d, 0x70, 0x42, 0x02, 0x18, 0x01, 0x48, 0x00, 0x52, 0x0e, 0x74, 0x69, 0x6d, + 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x42, 0x0f, 0x0a, 0x0d, 0x63, + 0x6f, 0x6e, 0x73, 0x74, 0x61, 0x6e, 0x74, 0x5f, 0x6b, 0x69, 0x6e, 0x64, 0x22, 0x8c, 0x07, 0x0a, + 0x0a, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x12, 0x25, 0x0a, 0x0e, 0x73, + 0x79, 0x6e, 0x74, 0x61, 0x78, 0x5f, 0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x01, 0x20, + 0x01, 0x28, 0x09, 0x52, 0x0d, 0x73, 0x79, 0x6e, 0x74, 0x61, 0x78, 0x56, 0x65, 0x72, 0x73, 0x69, + 0x6f, 0x6e, 0x12, 0x1a, 0x0a, 0x08, 0x6c, 0x6f, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x02, + 0x20, 0x01, 0x28, 0x09, 0x52, 0x08, 0x6c, 0x6f, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x21, + 0x0a, 0x0c, 0x6c, 0x69, 0x6e, 0x65, 0x5f, 0x6f, 0x66, 0x66, 0x73, 0x65, 0x74, 0x73, 0x18, 0x03, + 0x20, 0x03, 0x28, 0x05, 0x52, 0x0b, 0x6c, 0x69, 0x6e, 0x65, 0x4f, 0x66, 0x66, 0x73, 0x65, 0x74, + 0x73, 0x12, 0x51, 0x0a, 0x09, 0x70, 0x6f, 0x73, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x18, 0x04, + 0x20, 0x03, 0x28, 0x0b, 0x32, 0x33, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, + 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, + 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x2e, 0x50, 0x6f, 0x73, 0x69, 0x74, + 0x69, 0x6f, 0x6e, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, 0x52, 0x09, 0x70, 0x6f, 0x73, 0x69, 0x74, + 0x69, 0x6f, 0x6e, 0x73, 0x12, 0x55, 0x0a, 0x0b, 0x6d, 0x61, 0x63, 0x72, 0x6f, 0x5f, 0x63, 0x61, + 0x6c, 0x6c, 0x73, 0x18, 0x05, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x34, 0x2e, 0x67, 0x6f, 0x6f, 0x67, + 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, + 0x70, 0x68, 0x61, 0x31, 0x2e, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x2e, + 0x4d, 0x61, 0x63, 0x72, 0x6f, 0x43, 0x61, 0x6c, 0x6c, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, 0x52, + 0x0a, 0x6d, 0x61, 0x63, 0x72, 0x6f, 0x43, 0x61, 0x6c, 0x6c, 0x73, 0x12, 0x4e, 0x0a, 0x0a, 0x65, + 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x73, 0x18, 0x06, 0x20, 0x03, 0x28, 0x0b, 0x32, + 0x2e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, + 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x53, 0x6f, 0x75, 0x72, 0x63, + 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x2e, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x52, + 0x0a, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x73, 0x1a, 0x80, 0x03, 0x0a, 0x09, + 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x0e, 0x0a, 0x02, 0x69, 0x64, 0x18, + 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x02, 0x69, 0x64, 0x12, 0x69, 0x0a, 0x13, 0x61, 0x66, 0x66, + 0x65, 0x63, 0x74, 0x65, 0x64, 0x5f, 0x63, 0x6f, 0x6d, 0x70, 0x6f, 0x6e, 0x65, 0x6e, 0x74, 0x73, + 0x18, 0x02, 0x20, 0x03, 0x28, 0x0e, 0x32, 0x38, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, + 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, + 0x31, 0x2e, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x2e, 0x45, 0x78, 0x74, + 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x2e, 0x43, 0x6f, 0x6d, 0x70, 0x6f, 0x6e, 0x65, 0x6e, 0x74, + 0x52, 0x12, 0x61, 0x66, 0x66, 0x65, 0x63, 0x74, 0x65, 0x64, 0x43, 0x6f, 0x6d, 0x70, 0x6f, 0x6e, + 0x65, 0x6e, 0x74, 0x73, 0x12, 0x50, 0x0a, 0x07, 0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x18, + 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x36, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, - 0x42, 0x0b, 0x53, 0x79, 0x6e, 0x74, 0x61, 0x78, 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x50, 0x01, 0x5a, - 0x3c, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, 0x2e, 0x6f, - 0x72, 0x67, 0x2f, 0x67, 0x65, 0x6e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x2f, 0x67, 0x6f, 0x6f, 0x67, - 0x6c, 0x65, 0x61, 0x70, 0x69, 0x73, 0x2f, 0x61, 0x70, 0x69, 0x2f, 0x65, 0x78, 0x70, 0x72, 0x2f, - 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x3b, 0x65, 0x78, 0x70, 0x72, 0xf8, 0x01, 0x01, - 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x33, + 0x2e, 0x53, 0x6f, 0x75, 0x72, 0x63, 0x65, 0x49, 0x6e, 0x66, 0x6f, 0x2e, 0x45, 0x78, 0x74, 0x65, + 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x2e, 0x56, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x52, 0x07, 0x76, + 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x1a, 0x35, 0x0a, 0x07, 0x56, 0x65, 0x72, 0x73, 0x69, 0x6f, + 0x6e, 0x12, 0x14, 0x0a, 0x05, 0x6d, 0x61, 0x6a, 0x6f, 0x72, 0x18, 0x01, 0x20, 0x01, 0x28, 0x03, + 0x52, 0x05, 0x6d, 0x61, 0x6a, 0x6f, 0x72, 0x12, 0x14, 0x0a, 0x05, 0x6d, 0x69, 0x6e, 0x6f, 0x72, + 0x18, 0x02, 0x20, 0x01, 0x28, 0x03, 0x52, 0x05, 0x6d, 0x69, 0x6e, 0x6f, 0x72, 0x22, 0x6f, 0x0a, + 0x09, 0x43, 0x6f, 0x6d, 0x70, 0x6f, 0x6e, 0x65, 0x6e, 0x74, 0x12, 0x19, 0x0a, 0x15, 0x43, 0x4f, + 0x4d, 0x50, 0x4f, 0x4e, 0x45, 0x4e, 0x54, 0x5f, 0x55, 0x4e, 0x53, 0x50, 0x45, 0x43, 0x49, 0x46, + 0x49, 0x45, 0x44, 0x10, 0x00, 0x12, 0x14, 0x0a, 0x10, 0x43, 0x4f, 0x4d, 0x50, 0x4f, 0x4e, 0x45, + 0x4e, 0x54, 0x5f, 0x50, 0x41, 0x52, 0x53, 0x45, 0x52, 0x10, 0x01, 0x12, 0x1a, 0x0a, 0x16, 0x43, + 0x4f, 0x4d, 0x50, 0x4f, 0x4e, 0x45, 0x4e, 0x54, 0x5f, 0x54, 0x59, 0x50, 0x45, 0x5f, 0x43, 0x48, + 0x45, 0x43, 0x4b, 0x45, 0x52, 0x10, 0x02, 0x12, 0x15, 0x0a, 0x11, 0x43, 0x4f, 0x4d, 0x50, 0x4f, + 0x4e, 0x45, 0x4e, 0x54, 0x5f, 0x52, 0x55, 0x4e, 0x54, 0x49, 0x4d, 0x45, 0x10, 0x03, 0x1a, 0x3c, + 0x0a, 0x0e, 0x50, 0x6f, 0x73, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, + 0x12, 0x10, 0x0a, 0x03, 0x6b, 0x65, 0x79, 0x18, 0x01, 0x20, 0x01, 0x28, 0x03, 0x52, 0x03, 0x6b, + 0x65, 0x79, 0x12, 0x14, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, + 0x05, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x3a, 0x02, 0x38, 0x01, 0x1a, 0x5d, 0x0a, 0x0f, + 0x4d, 0x61, 0x63, 0x72, 0x6f, 0x43, 0x61, 0x6c, 0x6c, 0x73, 0x45, 0x6e, 0x74, 0x72, 0x79, 0x12, + 0x10, 0x0a, 0x03, 0x6b, 0x65, 0x79, 0x18, 0x01, 0x20, 0x01, 0x28, 0x03, 0x52, 0x03, 0x6b, 0x65, + 0x79, 0x12, 0x34, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, + 0x32, 0x1e, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, 0x65, 0x78, + 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x2e, 0x45, 0x78, 0x70, 0x72, + 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x3a, 0x02, 0x38, 0x01, 0x22, 0x70, 0x0a, 0x0e, 0x53, + 0x6f, 0x75, 0x72, 0x63, 0x65, 0x50, 0x6f, 0x73, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x1a, 0x0a, + 0x08, 0x6c, 0x6f, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x08, 0x6c, 0x6f, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x16, 0x0a, 0x06, 0x6f, 0x66, 0x66, + 0x73, 0x65, 0x74, 0x18, 0x02, 0x20, 0x01, 0x28, 0x05, 0x52, 0x06, 0x6f, 0x66, 0x66, 0x73, 0x65, + 0x74, 0x12, 0x12, 0x0a, 0x04, 0x6c, 0x69, 0x6e, 0x65, 0x18, 0x03, 0x20, 0x01, 0x28, 0x05, 0x52, + 0x04, 0x6c, 0x69, 0x6e, 0x65, 0x12, 0x16, 0x0a, 0x06, 0x63, 0x6f, 0x6c, 0x75, 0x6d, 0x6e, 0x18, + 0x04, 0x20, 0x01, 0x28, 0x05, 0x52, 0x06, 0x63, 0x6f, 0x6c, 0x75, 0x6d, 0x6e, 0x42, 0x6e, 0x0a, + 0x1c, 0x63, 0x6f, 0x6d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x61, 0x70, 0x69, 0x2e, + 0x65, 0x78, 0x70, 0x72, 0x2e, 0x76, 0x31, 0x61, 0x6c, 0x70, 0x68, 0x61, 0x31, 0x42, 0x0b, 0x53, + 0x79, 0x6e, 0x74, 0x61, 0x78, 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x50, 0x01, 0x5a, 0x3c, 0x67, 0x6f, + 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, 0x2e, 0x6f, 0x72, 0x67, 0x2f, + 0x67, 0x65, 0x6e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x2f, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x61, + 0x70, 0x69, 0x73, 0x2f, 0x61, 0x70, 0x69, 0x2f, 0x65, 0x78, 0x70, 0x72, 0x2f, 0x76, 0x31, 0x61, + 0x6c, 0x70, 0x68, 0x61, 0x31, 0x3b, 0x65, 0x78, 0x70, 0x72, 0xf8, 0x01, 0x01, 0x62, 0x06, 0x70, + 0x72, 0x6f, 0x74, 0x6f, 0x33, } var ( diff --git a/vendor/google.golang.org/grpc/MAINTAINERS.md b/vendor/google.golang.org/grpc/MAINTAINERS.md index 6a8a07781ae34..5d4096d46a048 100644 --- a/vendor/google.golang.org/grpc/MAINTAINERS.md +++ b/vendor/google.golang.org/grpc/MAINTAINERS.md @@ -9,21 +9,28 @@ for general contribution guidelines. ## Maintainers (in alphabetical order) +- [aranjans](https://github.com/aranjans), Google LLC +- [arjan-bal](https://github.com/arjan-bal), Google LLC +- [arvindbr8](https://github.com/arvindbr8), Google LLC - [atollena](https://github.com/atollena), Datadog, Inc. -- [cesarghali](https://github.com/cesarghali), Google LLC - [dfawley](https://github.com/dfawley), Google LLC - [easwars](https://github.com/easwars), Google LLC -- [menghanl](https://github.com/menghanl), Google LLC -- [srini100](https://github.com/srini100), Google LLC +- [erm-g](https://github.com/erm-g), Google LLC +- [gtcooke94](https://github.com/gtcooke94), Google LLC +- [purnesh42h](https://github.com/purnesh42h), Google LLC +- [zasweq](https://github.com/zasweq), Google LLC ## Emeritus Maintainers (in alphabetical order) -- [adelez](https://github.com/adelez), Google LLC -- [canguler](https://github.com/canguler), Google LLC -- [iamqizhao](https://github.com/iamqizhao), Google LLC -- [jadekler](https://github.com/jadekler), Google LLC -- [jtattermusch](https://github.com/jtattermusch), Google LLC -- [lyuxuan](https://github.com/lyuxuan), Google LLC -- [makmukhi](https://github.com/makmukhi), Google LLC -- [matt-kwong](https://github.com/matt-kwong), Google LLC -- [nicolasnoble](https://github.com/nicolasnoble), Google LLC -- [yongni](https://github.com/yongni), Google LLC +- [adelez](https://github.com/adelez) +- [canguler](https://github.com/canguler) +- [cesarghali](https://github.com/cesarghali) +- [iamqizhao](https://github.com/iamqizhao) +- [jeanbza](https://github.com/jeanbza) +- [jtattermusch](https://github.com/jtattermusch) +- [lyuxuan](https://github.com/lyuxuan) +- [makmukhi](https://github.com/makmukhi) +- [matt-kwong](https://github.com/matt-kwong) +- [menghanl](https://github.com/menghanl) +- [nicolasnoble](https://github.com/nicolasnoble) +- [srini100](https://github.com/srini100) +- [yongni](https://github.com/yongni) diff --git a/vendor/google.golang.org/grpc/SECURITY.md b/vendor/google.golang.org/grpc/SECURITY.md index be6e108705c48..abab279379ba8 100644 --- a/vendor/google.golang.org/grpc/SECURITY.md +++ b/vendor/google.golang.org/grpc/SECURITY.md @@ -1,3 +1,3 @@ # Security Policy -For information on gRPC Security Policy and reporting potentional security issues, please see [gRPC CVE Process](https://github.com/grpc/proposal/blob/master/P4-grpc-cve-process.md). +For information on gRPC Security Policy and reporting potential security issues, please see [gRPC CVE Process](https://github.com/grpc/proposal/blob/master/P4-grpc-cve-process.md). diff --git a/vendor/google.golang.org/grpc/backoff/backoff.go b/vendor/google.golang.org/grpc/backoff/backoff.go index 0787d0b50ce94..d7b40b7cb66f1 100644 --- a/vendor/google.golang.org/grpc/backoff/backoff.go +++ b/vendor/google.golang.org/grpc/backoff/backoff.go @@ -39,7 +39,7 @@ type Config struct { MaxDelay time.Duration } -// DefaultConfig is a backoff configuration with the default values specfied +// DefaultConfig is a backoff configuration with the default values specified // at https://github.com/grpc/grpc/blob/master/doc/connection-backoff.md. // // This should be useful for callers who want to configure backoff with diff --git a/vendor/google.golang.org/grpc/balancer/balancer.go b/vendor/google.golang.org/grpc/balancer/balancer.go index f391744f7299b..b181f386a1ba6 100644 --- a/vendor/google.golang.org/grpc/balancer/balancer.go +++ b/vendor/google.golang.org/grpc/balancer/balancer.go @@ -30,6 +30,7 @@ import ( "google.golang.org/grpc/channelz" "google.golang.org/grpc/connectivity" "google.golang.org/grpc/credentials" + estats "google.golang.org/grpc/experimental/stats" "google.golang.org/grpc/grpclog" "google.golang.org/grpc/internal" "google.golang.org/grpc/metadata" @@ -72,8 +73,21 @@ func unregisterForTesting(name string) { delete(m, name) } +// connectedAddress returns the connected address for a SubConnState. The +// address is only valid if the state is READY. +func connectedAddress(scs SubConnState) resolver.Address { + return scs.connectedAddress +} + +// setConnectedAddress sets the connected address for a SubConnState. +func setConnectedAddress(scs *SubConnState, addr resolver.Address) { + scs.connectedAddress = addr +} + func init() { internal.BalancerUnregister = unregisterForTesting + internal.ConnectedAddress = connectedAddress + internal.SetConnectedAddress = setConnectedAddress } // Get returns the resolver builder registered with the given name. @@ -243,6 +257,10 @@ type BuildOptions struct { // same resolver.Target as passed to the resolver. See the documentation for // the resolver.Target type for details about what it contains. Target resolver.Target + // MetricsRecorder is the metrics recorder that balancers can use to record + // metrics. Balancer implementations which do not register metrics on + // metrics registry and record on them can ignore this field. + MetricsRecorder estats.MetricsRecorder } // Builder creates a balancer. @@ -410,6 +428,9 @@ type SubConnState struct { // ConnectionError is set if the ConnectivityState is TransientFailure, // describing the reason the SubConn failed. Otherwise, it is nil. ConnectionError error + // connectedAddr contains the connected address when ConnectivityState is + // Ready. Otherwise, it is indeterminate. + connectedAddress resolver.Address } // ClientConnState describes the state of a ClientConn relevant to the diff --git a/vendor/google.golang.org/grpc/balancer/base/balancer.go b/vendor/google.golang.org/grpc/balancer/base/balancer.go index a7f1eeec8e6ae..2b87bd79c7573 100644 --- a/vendor/google.golang.org/grpc/balancer/base/balancer.go +++ b/vendor/google.golang.org/grpc/balancer/base/balancer.go @@ -36,7 +36,7 @@ type baseBuilder struct { config Config } -func (bb *baseBuilder) Build(cc balancer.ClientConn, opt balancer.BuildOptions) balancer.Balancer { +func (bb *baseBuilder) Build(cc balancer.ClientConn, _ balancer.BuildOptions) balancer.Balancer { bal := &baseBalancer{ cc: cc, pickerBuilder: bb.pickerBuilder, @@ -259,6 +259,6 @@ type errPicker struct { err error // Pick() always returns this err. } -func (p *errPicker) Pick(info balancer.PickInfo) (balancer.PickResult, error) { +func (p *errPicker) Pick(balancer.PickInfo) (balancer.PickResult, error) { return balancer.PickResult{}, p.err } diff --git a/vendor/google.golang.org/grpc/balancer/pickfirst/pickfirst.go b/vendor/google.golang.org/grpc/balancer/pickfirst/pickfirst.go index 07527603f1d4e..4d69b4052f8e3 100644 --- a/vendor/google.golang.org/grpc/balancer/pickfirst/pickfirst.go +++ b/vendor/google.golang.org/grpc/balancer/pickfirst/pickfirst.go @@ -50,7 +50,7 @@ const ( type pickfirstBuilder struct{} -func (pickfirstBuilder) Build(cc balancer.ClientConn, opt balancer.BuildOptions) balancer.Balancer { +func (pickfirstBuilder) Build(cc balancer.ClientConn, _ balancer.BuildOptions) balancer.Balancer { b := &pickfirstBalancer{cc: cc} b.logger = internalgrpclog.NewPrefixLogger(logger, fmt.Sprintf(logPrefix, b)) return b @@ -155,7 +155,7 @@ func (b *pickfirstBalancer) UpdateClientConnState(state balancer.ClientConnState // Endpoints not set, process addresses until we migrate resolver // emissions fully to Endpoints. The top channel does wrap emitted // addresses with endpoints, however some balancers such as weighted - // target do not forwarrd the corresponding correct endpoints down/split + // target do not forward the corresponding correct endpoints down/split // endpoints properly. Once all balancers correctly forward endpoints // down, can delete this else conditional. addrs = state.ResolverState.Addresses diff --git a/vendor/google.golang.org/grpc/balancer_wrapper.go b/vendor/google.golang.org/grpc/balancer_wrapper.go index 4161fdf47a8b1..8ad6ce2f0950a 100644 --- a/vendor/google.golang.org/grpc/balancer_wrapper.go +++ b/vendor/google.golang.org/grpc/balancer_wrapper.go @@ -25,12 +25,15 @@ import ( "google.golang.org/grpc/balancer" "google.golang.org/grpc/connectivity" + "google.golang.org/grpc/internal" "google.golang.org/grpc/internal/balancer/gracefulswitch" "google.golang.org/grpc/internal/channelz" "google.golang.org/grpc/internal/grpcsync" "google.golang.org/grpc/resolver" ) +var setConnectedAddress = internal.SetConnectedAddress.(func(*balancer.SubConnState, resolver.Address)) + // ccBalancerWrapper sits between the ClientConn and the Balancer. // // ccBalancerWrapper implements methods corresponding to the ones on the @@ -79,6 +82,7 @@ func newCCBalancerWrapper(cc *ClientConn) *ccBalancerWrapper { CustomUserAgent: cc.dopts.copts.UserAgent, ChannelzParent: cc.channelz, Target: cc.parsedTarget, + MetricsRecorder: cc.metricsRecorderList, }, serializer: grpcsync.NewCallbackSerializer(ctx), serializerCancel: cancel, @@ -92,7 +96,7 @@ func newCCBalancerWrapper(cc *ClientConn) *ccBalancerWrapper { // it is safe to call into the balancer here. func (ccb *ccBalancerWrapper) updateClientConnState(ccs *balancer.ClientConnState) error { errCh := make(chan error) - ok := ccb.serializer.Schedule(func(ctx context.Context) { + uccs := func(ctx context.Context) { defer close(errCh) if ctx.Err() != nil || ccb.balancer == nil { return @@ -107,17 +111,23 @@ func (ccb *ccBalancerWrapper) updateClientConnState(ccs *balancer.ClientConnStat logger.Infof("error from balancer.UpdateClientConnState: %v", err) } errCh <- err - }) - if !ok { - return nil } + onFailure := func() { close(errCh) } + + // UpdateClientConnState can race with Close, and when the latter wins, the + // serializer is closed, and the attempt to schedule the callback will fail. + // It is acceptable to ignore this failure. But since we want to handle the + // state update in a blocking fashion (when we successfully schedule the + // callback), we have to use the ScheduleOr method and not the MaybeSchedule + // method on the serializer. + ccb.serializer.ScheduleOr(uccs, onFailure) return <-errCh } // resolverError is invoked by grpc to push a resolver error to the underlying // balancer. The call to the balancer is executed from the serializer. func (ccb *ccBalancerWrapper) resolverError(err error) { - ccb.serializer.Schedule(func(ctx context.Context) { + ccb.serializer.TrySchedule(func(ctx context.Context) { if ctx.Err() != nil || ccb.balancer == nil { return } @@ -133,7 +143,7 @@ func (ccb *ccBalancerWrapper) close() { ccb.closed = true ccb.mu.Unlock() channelz.Info(logger, ccb.cc.channelz, "ccBalancerWrapper: closing") - ccb.serializer.Schedule(func(context.Context) { + ccb.serializer.TrySchedule(func(context.Context) { if ccb.balancer == nil { return } @@ -145,7 +155,7 @@ func (ccb *ccBalancerWrapper) close() { // exitIdle invokes the balancer's exitIdle method in the serializer. func (ccb *ccBalancerWrapper) exitIdle() { - ccb.serializer.Schedule(func(ctx context.Context) { + ccb.serializer.TrySchedule(func(ctx context.Context) { if ctx.Err() != nil || ccb.balancer == nil { return } @@ -182,7 +192,7 @@ func (ccb *ccBalancerWrapper) NewSubConn(addrs []resolver.Address, opts balancer return acbw, nil } -func (ccb *ccBalancerWrapper) RemoveSubConn(sc balancer.SubConn) { +func (ccb *ccBalancerWrapper) RemoveSubConn(balancer.SubConn) { // The graceful switch balancer will never call this. logger.Errorf("ccb RemoveSubConn(%v) called unexpectedly, sc") } @@ -252,15 +262,29 @@ type acBalancerWrapper struct { // updateState is invoked by grpc to push a subConn state update to the // underlying balancer. -func (acbw *acBalancerWrapper) updateState(s connectivity.State, err error) { - acbw.ccb.serializer.Schedule(func(ctx context.Context) { +func (acbw *acBalancerWrapper) updateState(s connectivity.State, curAddr resolver.Address, err error) { + acbw.ccb.serializer.TrySchedule(func(ctx context.Context) { if ctx.Err() != nil || acbw.ccb.balancer == nil { return } // Even though it is optional for balancers, gracefulswitch ensures // opts.StateListener is set, so this cannot ever be nil. // TODO: delete this comment when UpdateSubConnState is removed. - acbw.stateListener(balancer.SubConnState{ConnectivityState: s, ConnectionError: err}) + scs := balancer.SubConnState{ConnectivityState: s, ConnectionError: err} + if s == connectivity.Ready { + setConnectedAddress(&scs, curAddr) + } + acbw.stateListener(scs) + acbw.ac.mu.Lock() + defer acbw.ac.mu.Unlock() + if s == connectivity.Ready { + // When changing states to READY, reset stateReadyChan. Wait until + // after we notify the LB policy's listener(s) in order to prevent + // ac.getTransport() from unblocking before the LB policy starts + // tracking the subchannel as READY. + close(acbw.ac.stateReadyChan) + acbw.ac.stateReadyChan = make(chan struct{}) + } }) } @@ -318,8 +342,8 @@ func (acbw *acBalancerWrapper) GetOrBuildProducer(pb balancer.ProducerBuilder) ( pData := acbw.producers[pb] if pData == nil { // Not found; create a new one and add it to the producers map. - p, close := pb.Build(acbw) - pData = &refCountedProducer{producer: p, close: close} + p, closeFn := pb.Build(acbw) + pData = &refCountedProducer{producer: p, close: closeFn} acbw.producers[pb] = pData } // Account for this new reference. diff --git a/vendor/google.golang.org/grpc/binarylog/grpc_binarylog_v1/binarylog.pb.go b/vendor/google.golang.org/grpc/binarylog/grpc_binarylog_v1/binarylog.pb.go index 63c639e4fe933..55bffaa77ef0f 100644 --- a/vendor/google.golang.org/grpc/binarylog/grpc_binarylog_v1/binarylog.pb.go +++ b/vendor/google.golang.org/grpc/binarylog/grpc_binarylog_v1/binarylog.pb.go @@ -18,8 +18,8 @@ // Code generated by protoc-gen-go. DO NOT EDIT. // versions: -// protoc-gen-go v1.34.1 -// protoc v4.25.2 +// protoc-gen-go v1.34.2 +// protoc v5.27.1 // source: grpc/binlog/v1/binarylog.proto package grpc_binarylog_v1 @@ -1015,7 +1015,7 @@ func file_grpc_binlog_v1_binarylog_proto_rawDescGZIP() []byte { var file_grpc_binlog_v1_binarylog_proto_enumTypes = make([]protoimpl.EnumInfo, 3) var file_grpc_binlog_v1_binarylog_proto_msgTypes = make([]protoimpl.MessageInfo, 8) -var file_grpc_binlog_v1_binarylog_proto_goTypes = []interface{}{ +var file_grpc_binlog_v1_binarylog_proto_goTypes = []any{ (GrpcLogEntry_EventType)(0), // 0: grpc.binarylog.v1.GrpcLogEntry.EventType (GrpcLogEntry_Logger)(0), // 1: grpc.binarylog.v1.GrpcLogEntry.Logger (Address_Type)(0), // 2: grpc.binarylog.v1.Address.Type @@ -1058,7 +1058,7 @@ func file_grpc_binlog_v1_binarylog_proto_init() { return } if !protoimpl.UnsafeEnabled { - file_grpc_binlog_v1_binarylog_proto_msgTypes[0].Exporter = func(v interface{}, i int) interface{} { + file_grpc_binlog_v1_binarylog_proto_msgTypes[0].Exporter = func(v any, i int) any { switch v := v.(*GrpcLogEntry); i { case 0: return &v.state @@ -1070,7 +1070,7 @@ func file_grpc_binlog_v1_binarylog_proto_init() { return nil } } - file_grpc_binlog_v1_binarylog_proto_msgTypes[1].Exporter = func(v interface{}, i int) interface{} { + file_grpc_binlog_v1_binarylog_proto_msgTypes[1].Exporter = func(v any, i int) any { switch v := v.(*ClientHeader); i { case 0: return &v.state @@ -1082,7 +1082,7 @@ func file_grpc_binlog_v1_binarylog_proto_init() { return nil } } - file_grpc_binlog_v1_binarylog_proto_msgTypes[2].Exporter = func(v interface{}, i int) interface{} { + file_grpc_binlog_v1_binarylog_proto_msgTypes[2].Exporter = func(v any, i int) any { switch v := v.(*ServerHeader); i { case 0: return &v.state @@ -1094,7 +1094,7 @@ func file_grpc_binlog_v1_binarylog_proto_init() { return nil } } - file_grpc_binlog_v1_binarylog_proto_msgTypes[3].Exporter = func(v interface{}, i int) interface{} { + file_grpc_binlog_v1_binarylog_proto_msgTypes[3].Exporter = func(v any, i int) any { switch v := v.(*Trailer); i { case 0: return &v.state @@ -1106,7 +1106,7 @@ func file_grpc_binlog_v1_binarylog_proto_init() { return nil } } - file_grpc_binlog_v1_binarylog_proto_msgTypes[4].Exporter = func(v interface{}, i int) interface{} { + file_grpc_binlog_v1_binarylog_proto_msgTypes[4].Exporter = func(v any, i int) any { switch v := v.(*Message); i { case 0: return &v.state @@ -1118,7 +1118,7 @@ func file_grpc_binlog_v1_binarylog_proto_init() { return nil } } - file_grpc_binlog_v1_binarylog_proto_msgTypes[5].Exporter = func(v interface{}, i int) interface{} { + file_grpc_binlog_v1_binarylog_proto_msgTypes[5].Exporter = func(v any, i int) any { switch v := v.(*Metadata); i { case 0: return &v.state @@ -1130,7 +1130,7 @@ func file_grpc_binlog_v1_binarylog_proto_init() { return nil } } - file_grpc_binlog_v1_binarylog_proto_msgTypes[6].Exporter = func(v interface{}, i int) interface{} { + file_grpc_binlog_v1_binarylog_proto_msgTypes[6].Exporter = func(v any, i int) any { switch v := v.(*MetadataEntry); i { case 0: return &v.state @@ -1142,7 +1142,7 @@ func file_grpc_binlog_v1_binarylog_proto_init() { return nil } } - file_grpc_binlog_v1_binarylog_proto_msgTypes[7].Exporter = func(v interface{}, i int) interface{} { + file_grpc_binlog_v1_binarylog_proto_msgTypes[7].Exporter = func(v any, i int) any { switch v := v.(*Address); i { case 0: return &v.state @@ -1155,7 +1155,7 @@ func file_grpc_binlog_v1_binarylog_proto_init() { } } } - file_grpc_binlog_v1_binarylog_proto_msgTypes[0].OneofWrappers = []interface{}{ + file_grpc_binlog_v1_binarylog_proto_msgTypes[0].OneofWrappers = []any{ (*GrpcLogEntry_ClientHeader)(nil), (*GrpcLogEntry_ServerHeader)(nil), (*GrpcLogEntry_Message)(nil), diff --git a/vendor/google.golang.org/grpc/clientconn.go b/vendor/google.golang.org/grpc/clientconn.go index 423be7b43b00c..9c8850e3fdd5b 100644 --- a/vendor/google.golang.org/grpc/clientconn.go +++ b/vendor/google.golang.org/grpc/clientconn.go @@ -24,6 +24,7 @@ import ( "fmt" "math" "net/url" + "slices" "strings" "sync" "sync/atomic" @@ -39,6 +40,7 @@ import ( "google.golang.org/grpc/internal/grpcsync" "google.golang.org/grpc/internal/idle" iresolver "google.golang.org/grpc/internal/resolver" + "google.golang.org/grpc/internal/stats" "google.golang.org/grpc/internal/transport" "google.golang.org/grpc/keepalive" "google.golang.org/grpc/resolver" @@ -194,8 +196,11 @@ func NewClient(target string, opts ...DialOption) (conn *ClientConn, err error) cc.csMgr = newConnectivityStateManager(cc.ctx, cc.channelz) cc.pickerWrapper = newPickerWrapper(cc.dopts.copts.StatsHandlers) + cc.metricsRecorderList = stats.NewMetricsRecorderList(cc.dopts.copts.StatsHandlers) + cc.initIdleStateLocked() // Safe to call without the lock, since nothing else has a reference to cc. cc.idlenessMgr = idle.NewManager((*idler)(cc), cc.dopts.idleTimeout) + return cc, nil } @@ -590,13 +595,14 @@ type ClientConn struct { cancel context.CancelFunc // Cancelled on close. // The following are initialized at dial time, and are read-only after that. - target string // User's dial target. - parsedTarget resolver.Target // See initParsedTargetAndResolverBuilder(). - authority string // See initAuthority(). - dopts dialOptions // Default and user specified dial options. - channelz *channelz.Channel // Channelz object. - resolverBuilder resolver.Builder // See initParsedTargetAndResolverBuilder(). - idlenessMgr *idle.Manager + target string // User's dial target. + parsedTarget resolver.Target // See initParsedTargetAndResolverBuilder(). + authority string // See initAuthority(). + dopts dialOptions // Default and user specified dial options. + channelz *channelz.Channel // Channelz object. + resolverBuilder resolver.Builder // See initParsedTargetAndResolverBuilder(). + idlenessMgr *idle.Manager + metricsRecorderList *stats.MetricsRecorderList // The following provide their own synchronization, and therefore don't // require cc.mu to be held to access them. @@ -626,11 +632,6 @@ type ClientConn struct { // WaitForStateChange waits until the connectivity.State of ClientConn changes from sourceState or // ctx expires. A true value is returned in former case and false in latter. -// -// # Experimental -// -// Notice: This API is EXPERIMENTAL and may be changed or removed in a -// later release. func (cc *ClientConn) WaitForStateChange(ctx context.Context, sourceState connectivity.State) bool { ch := cc.csMgr.getNotifyChan() if cc.csMgr.getState() != sourceState { @@ -645,11 +646,6 @@ func (cc *ClientConn) WaitForStateChange(ctx context.Context, sourceState connec } // GetState returns the connectivity.State of ClientConn. -// -// # Experimental -// -// Notice: This API is EXPERIMENTAL and may be changed or removed in a later -// release. func (cc *ClientConn) GetState() connectivity.State { return cc.csMgr.getState() } @@ -812,17 +808,11 @@ func (cc *ClientConn) applyFailingLBLocked(sc *serviceconfig.ParseResult) { cc.csMgr.updateState(connectivity.TransientFailure) } -// Makes a copy of the input addresses slice and clears out the balancer -// attributes field. Addresses are passed during subconn creation and address -// update operations. In both cases, we will clear the balancer attributes by -// calling this function, and therefore we will be able to use the Equal method -// provided by the resolver.Address type for comparison. -func copyAddressesWithoutBalancerAttributes(in []resolver.Address) []resolver.Address { +// Makes a copy of the input addresses slice. Addresses are passed during +// subconn creation and address update operations. +func copyAddresses(in []resolver.Address) []resolver.Address { out := make([]resolver.Address, len(in)) - for i := range in { - out[i] = in[i] - out[i].BalancerAttributes = nil - } + copy(out, in) return out } @@ -835,14 +825,14 @@ func (cc *ClientConn) newAddrConnLocked(addrs []resolver.Address, opts balancer. } ac := &addrConn{ - state: connectivity.Idle, - cc: cc, - addrs: copyAddressesWithoutBalancerAttributes(addrs), - scopts: opts, - dopts: cc.dopts, - channelz: channelz.RegisterSubChannel(cc.channelz, ""), - resetBackoff: make(chan struct{}), - stateChan: make(chan struct{}), + state: connectivity.Idle, + cc: cc, + addrs: copyAddresses(addrs), + scopts: opts, + dopts: cc.dopts, + channelz: channelz.RegisterSubChannel(cc.channelz, ""), + resetBackoff: make(chan struct{}), + stateReadyChan: make(chan struct{}), } ac.ctx, ac.cancel = context.WithCancel(cc.ctx) // Start with our address set to the first address; this may be updated if @@ -918,28 +908,29 @@ func (ac *addrConn) connect() error { ac.mu.Unlock() return nil } - ac.mu.Unlock() - ac.resetTransport() + ac.resetTransportAndUnlock() return nil } -func equalAddresses(a, b []resolver.Address) bool { - if len(a) != len(b) { - return false - } - for i, v := range a { - if !v.Equal(b[i]) { - return false - } - } - return true +// equalAddressIgnoringBalAttributes returns true is a and b are considered equal. +// This is different from the Equal method on the resolver.Address type which +// considers all fields to determine equality. Here, we only consider fields +// that are meaningful to the subConn. +func equalAddressIgnoringBalAttributes(a, b *resolver.Address) bool { + return a.Addr == b.Addr && a.ServerName == b.ServerName && + a.Attributes.Equal(b.Attributes) && + a.Metadata == b.Metadata +} + +func equalAddressesIgnoringBalAttributes(a, b []resolver.Address) bool { + return slices.EqualFunc(a, b, func(a, b resolver.Address) bool { return equalAddressIgnoringBalAttributes(&a, &b) }) } // updateAddrs updates ac.addrs with the new addresses list and handles active // connections or connection attempts. func (ac *addrConn) updateAddrs(addrs []resolver.Address) { - addrs = copyAddressesWithoutBalancerAttributes(addrs) + addrs = copyAddresses(addrs) limit := len(addrs) if limit > 5 { limit = 5 @@ -947,7 +938,7 @@ func (ac *addrConn) updateAddrs(addrs []resolver.Address) { channelz.Infof(logger, ac.channelz, "addrConn: updateAddrs addrs (%d of %d): %v", limit, len(addrs), addrs[:limit]) ac.mu.Lock() - if equalAddresses(ac.addrs, addrs) { + if equalAddressesIgnoringBalAttributes(ac.addrs, addrs) { ac.mu.Unlock() return } @@ -966,7 +957,7 @@ func (ac *addrConn) updateAddrs(addrs []resolver.Address) { // Try to find the connected address. for _, a := range addrs { a.ServerName = ac.cc.getServerName(a) - if a.Equal(ac.curAddr) { + if equalAddressIgnoringBalAttributes(&a, &ac.curAddr) { // We are connected to a valid address, so do nothing but // update the addresses. ac.mu.Unlock() @@ -992,11 +983,9 @@ func (ac *addrConn) updateAddrs(addrs []resolver.Address) { ac.updateConnectivityState(connectivity.Idle, nil) } - ac.mu.Unlock() - // Since we were connecting/connected, we should start a new connection // attempt. - go ac.resetTransport() + go ac.resetTransportAndUnlock() } // getServerName determines the serverName to be used in the connection @@ -1190,8 +1179,8 @@ type addrConn struct { addrs []resolver.Address // All addresses that the resolver resolved to. // Use updateConnectivityState for updating addrConn's connectivity state. - state connectivity.State - stateChan chan struct{} // closed and recreated on every state change. + state connectivity.State + stateReadyChan chan struct{} // closed and recreated on every READY state change. backoffIdx int // Needs to be stateful for resetConnectBackoff. resetBackoff chan struct{} @@ -1204,9 +1193,6 @@ func (ac *addrConn) updateConnectivityState(s connectivity.State, lastErr error) if ac.state == s { return } - // When changing states, reset the state change channel. - close(ac.stateChan) - ac.stateChan = make(chan struct{}) ac.state = s ac.channelz.ChannelMetrics.State.Store(&s) if lastErr == nil { @@ -1214,7 +1200,7 @@ func (ac *addrConn) updateConnectivityState(s connectivity.State, lastErr error) } else { channelz.Infof(logger, ac.channelz, "Subchannel Connectivity change to %v, last error: %s", s, lastErr) } - ac.acbw.updateState(s, lastErr) + ac.acbw.updateState(s, ac.curAddr, lastErr) } // adjustParams updates parameters used to create transports upon @@ -1231,8 +1217,10 @@ func (ac *addrConn) adjustParams(r transport.GoAwayReason) { } } -func (ac *addrConn) resetTransport() { - ac.mu.Lock() +// resetTransportAndUnlock unconditionally connects the addrConn. +// +// ac.mu must be held by the caller, and this function will guarantee it is released. +func (ac *addrConn) resetTransportAndUnlock() { acCtx := ac.ctx if acCtx.Err() != nil { ac.mu.Unlock() @@ -1522,7 +1510,7 @@ func (ac *addrConn) getReadyTransport() transport.ClientTransport { func (ac *addrConn) getTransport(ctx context.Context) (transport.ClientTransport, error) { for ctx.Err() == nil { ac.mu.Lock() - t, state, sc := ac.transport, ac.state, ac.stateChan + t, state, sc := ac.transport, ac.state, ac.stateReadyChan ac.mu.Unlock() if state == connectivity.Ready { return t, nil @@ -1585,7 +1573,7 @@ func (ac *addrConn) tearDown(err error) { } else { // Hard close the transport when the channel is entering idle or is // being shutdown. In the case where the channel is being shutdown, - // closing of transports is also taken care of by cancelation of cc.ctx. + // closing of transports is also taken care of by cancellation of cc.ctx. // But in the case where the channel is entering idle, we need to // explicitly close the transports here. Instead of distinguishing // between these two cases, it is simpler to close the transport diff --git a/vendor/google.golang.org/grpc/codec.go b/vendor/google.golang.org/grpc/codec.go index 411e3dfd47ccd..e840858b77b18 100644 --- a/vendor/google.golang.org/grpc/codec.go +++ b/vendor/google.golang.org/grpc/codec.go @@ -21,18 +21,73 @@ package grpc import ( "google.golang.org/grpc/encoding" _ "google.golang.org/grpc/encoding/proto" // to register the Codec for "proto" + "google.golang.org/grpc/mem" ) -// baseCodec contains the functionality of both Codec and encoding.Codec, but -// omits the name/string, which vary between the two and are not needed for -// anything besides the registry in the encoding package. +// baseCodec captures the new encoding.CodecV2 interface without the Name +// function, allowing it to be implemented by older Codec and encoding.Codec +// implementations. The omitted Name function is only needed for the register in +// the encoding package and is not part of the core functionality. type baseCodec interface { - Marshal(v any) ([]byte, error) - Unmarshal(data []byte, v any) error + Marshal(v any) (mem.BufferSlice, error) + Unmarshal(data mem.BufferSlice, v any) error +} + +// getCodec returns an encoding.CodecV2 for the codec of the given name (if +// registered). Initially checks the V2 registry with encoding.GetCodecV2 and +// returns the V2 codec if it is registered. Otherwise, it checks the V1 registry +// with encoding.GetCodec and if it is registered wraps it with newCodecV1Bridge +// to turn it into an encoding.CodecV2. Returns nil otherwise. +func getCodec(name string) encoding.CodecV2 { + if codecV1 := encoding.GetCodec(name); codecV1 != nil { + return newCodecV1Bridge(codecV1) + } + + return encoding.GetCodecV2(name) +} + +func newCodecV0Bridge(c Codec) baseCodec { + return codecV0Bridge{codec: c} +} + +func newCodecV1Bridge(c encoding.Codec) encoding.CodecV2 { + return codecV1Bridge{ + codecV0Bridge: codecV0Bridge{codec: c}, + name: c.Name(), + } +} + +var _ baseCodec = codecV0Bridge{} + +type codecV0Bridge struct { + codec interface { + Marshal(v any) ([]byte, error) + Unmarshal(data []byte, v any) error + } +} + +func (c codecV0Bridge) Marshal(v any) (mem.BufferSlice, error) { + data, err := c.codec.Marshal(v) + if err != nil { + return nil, err + } + return mem.BufferSlice{mem.NewBuffer(&data, nil)}, nil +} + +func (c codecV0Bridge) Unmarshal(data mem.BufferSlice, v any) (err error) { + return c.codec.Unmarshal(data.Materialize(), v) } -var _ baseCodec = Codec(nil) -var _ baseCodec = encoding.Codec(nil) +var _ encoding.CodecV2 = codecV1Bridge{} + +type codecV1Bridge struct { + codecV0Bridge + name string +} + +func (c codecV1Bridge) Name() string { + return c.name +} // Codec defines the interface gRPC uses to encode and decode messages. // Note that implementations of this interface must be thread safe; diff --git a/vendor/google.golang.org/grpc/credentials/insecure/insecure.go b/vendor/google.golang.org/grpc/credentials/insecure/insecure.go index 82bee1443bfee..4c805c64462c9 100644 --- a/vendor/google.golang.org/grpc/credentials/insecure/insecure.go +++ b/vendor/google.golang.org/grpc/credentials/insecure/insecure.go @@ -40,7 +40,7 @@ func NewCredentials() credentials.TransportCredentials { // NoSecurity. type insecureTC struct{} -func (insecureTC) ClientHandshake(ctx context.Context, _ string, conn net.Conn) (net.Conn, credentials.AuthInfo, error) { +func (insecureTC) ClientHandshake(_ context.Context, _ string, conn net.Conn) (net.Conn, credentials.AuthInfo, error) { return conn, info{credentials.CommonAuthInfo{SecurityLevel: credentials.NoSecurity}}, nil } diff --git a/vendor/google.golang.org/grpc/dialoptions.go b/vendor/google.golang.org/grpc/dialoptions.go index f5453d48a53f3..2b285beee376b 100644 --- a/vendor/google.golang.org/grpc/dialoptions.go +++ b/vendor/google.golang.org/grpc/dialoptions.go @@ -33,6 +33,7 @@ import ( "google.golang.org/grpc/internal/binarylog" "google.golang.org/grpc/internal/transport" "google.golang.org/grpc/keepalive" + "google.golang.org/grpc/mem" "google.golang.org/grpc/resolver" "google.golang.org/grpc/stats" ) @@ -60,7 +61,7 @@ func init() { internal.WithBinaryLogger = withBinaryLogger internal.JoinDialOptions = newJoinDialOption internal.DisableGlobalDialOptions = newDisableGlobalDialOptions - internal.WithRecvBufferPool = withRecvBufferPool + internal.WithBufferPool = withBufferPool } // dialOptions configure a Dial call. dialOptions are set by the DialOption @@ -92,7 +93,6 @@ type dialOptions struct { defaultServiceConfigRawJSON *string resolvers []resolver.Builder idleTimeout time.Duration - recvBufferPool SharedBufferPool defaultScheme string maxCallAttempts int } @@ -518,6 +518,8 @@ func WithUserAgent(s string) DialOption { // WithKeepaliveParams returns a DialOption that specifies keepalive parameters // for the client transport. +// +// Keepalive is disabled by default. func WithKeepaliveParams(kp keepalive.ClientParameters) DialOption { if kp.Time < internal.KeepaliveMinPingTime { logger.Warningf("Adjusting keepalive ping interval to minimum period of %v", internal.KeepaliveMinPingTime) @@ -677,11 +679,11 @@ func defaultDialOptions() dialOptions { WriteBufferSize: defaultWriteBufSize, UseProxy: true, UserAgent: grpcUA, + BufferPool: mem.DefaultBufferPool(), }, bs: internalbackoff.DefaultExponential, healthCheckFunc: internal.HealthCheckFunc, idleTimeout: 30 * time.Minute, - recvBufferPool: nopBufferPool{}, defaultScheme: "dns", maxCallAttempts: defaultMaxCallAttempts, } @@ -758,25 +760,8 @@ func WithMaxCallAttempts(n int) DialOption { }) } -// WithRecvBufferPool returns a DialOption that configures the ClientConn -// to use the provided shared buffer pool for parsing incoming messages. Depending -// on the application's workload, this could result in reduced memory allocation. -// -// If you are unsure about how to implement a memory pool but want to utilize one, -// begin with grpc.NewSharedBufferPool. -// -// Note: The shared buffer pool feature will not be active if any of the following -// options are used: WithStatsHandler, EnableTracing, or binary logging. In such -// cases, the shared buffer pool will be ignored. -// -// Deprecated: use experimental.WithRecvBufferPool instead. Will be deleted in -// v1.60.0 or later. -func WithRecvBufferPool(bufferPool SharedBufferPool) DialOption { - return withRecvBufferPool(bufferPool) -} - -func withRecvBufferPool(bufferPool SharedBufferPool) DialOption { +func withBufferPool(bufferPool mem.BufferPool) DialOption { return newFuncDialOption(func(o *dialOptions) { - o.recvBufferPool = bufferPool + o.copts.BufferPool = bufferPool }) } diff --git a/vendor/google.golang.org/grpc/doc.go b/vendor/google.golang.org/grpc/doc.go index 0022859ad7465..e7b532b6f806f 100644 --- a/vendor/google.golang.org/grpc/doc.go +++ b/vendor/google.golang.org/grpc/doc.go @@ -16,7 +16,7 @@ * */ -//go:generate ./regenerate.sh +//go:generate ./scripts/regenerate.sh /* Package grpc implements an RPC system called gRPC. diff --git a/vendor/google.golang.org/grpc/encoding/encoding.go b/vendor/google.golang.org/grpc/encoding/encoding.go index 5ebf88d7147f2..11d0ae142c429 100644 --- a/vendor/google.golang.org/grpc/encoding/encoding.go +++ b/vendor/google.golang.org/grpc/encoding/encoding.go @@ -94,7 +94,7 @@ type Codec interface { Name() string } -var registeredCodecs = make(map[string]Codec) +var registeredCodecs = make(map[string]any) // RegisterCodec registers the provided Codec for use with all gRPC clients and // servers. @@ -126,5 +126,6 @@ func RegisterCodec(codec Codec) { // // The content-subtype is expected to be lowercase. func GetCodec(contentSubtype string) Codec { - return registeredCodecs[contentSubtype] + c, _ := registeredCodecs[contentSubtype].(Codec) + return c } diff --git a/vendor/google.golang.org/grpc/encoding/encoding_v2.go b/vendor/google.golang.org/grpc/encoding/encoding_v2.go new file mode 100644 index 0000000000000..074c5e234a7b3 --- /dev/null +++ b/vendor/google.golang.org/grpc/encoding/encoding_v2.go @@ -0,0 +1,81 @@ +/* + * + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package encoding + +import ( + "strings" + + "google.golang.org/grpc/mem" +) + +// CodecV2 defines the interface gRPC uses to encode and decode messages. Note +// that implementations of this interface must be thread safe; a CodecV2's +// methods can be called from concurrent goroutines. +type CodecV2 interface { + // Marshal returns the wire format of v. The buffers in the returned + // [mem.BufferSlice] must have at least one reference each, which will be freed + // by gRPC when they are no longer needed. + Marshal(v any) (out mem.BufferSlice, err error) + // Unmarshal parses the wire format into v. Note that data will be freed as soon + // as this function returns. If the codec wishes to guarantee access to the data + // after this function, it must take its own reference that it frees when it is + // no longer needed. + Unmarshal(data mem.BufferSlice, v any) error + // Name returns the name of the Codec implementation. The returned string + // will be used as part of content type in transmission. The result must be + // static; the result cannot change between calls. + Name() string +} + +// RegisterCodecV2 registers the provided CodecV2 for use with all gRPC clients and +// servers. +// +// The CodecV2 will be stored and looked up by result of its Name() method, which +// should match the content-subtype of the encoding handled by the CodecV2. This +// is case-insensitive, and is stored and looked up as lowercase. If the +// result of calling Name() is an empty string, RegisterCodecV2 will panic. See +// Content-Type on +// https://github.com/grpc/grpc/blob/master/doc/PROTOCOL-HTTP2.md#requests for +// more details. +// +// If both a Codec and CodecV2 are registered with the same name, the CodecV2 +// will be used. +// +// NOTE: this function must only be called during initialization time (i.e. in +// an init() function), and is not thread-safe. If multiple Codecs are +// registered with the same name, the one registered last will take effect. +func RegisterCodecV2(codec CodecV2) { + if codec == nil { + panic("cannot register a nil CodecV2") + } + if codec.Name() == "" { + panic("cannot register CodecV2 with empty string result for Name()") + } + contentSubtype := strings.ToLower(codec.Name()) + registeredCodecs[contentSubtype] = codec +} + +// GetCodecV2 gets a registered CodecV2 by content-subtype, or nil if no CodecV2 is +// registered for the content-subtype. +// +// The content-subtype is expected to be lowercase. +func GetCodecV2(contentSubtype string) CodecV2 { + c, _ := registeredCodecs[contentSubtype].(CodecV2) + return c +} diff --git a/vendor/google.golang.org/grpc/encoding/proto/proto.go b/vendor/google.golang.org/grpc/encoding/proto/proto.go index 66d5cdf03ec58..ceec319dd2fb4 100644 --- a/vendor/google.golang.org/grpc/encoding/proto/proto.go +++ b/vendor/google.golang.org/grpc/encoding/proto/proto.go @@ -1,6 +1,6 @@ /* * - * Copyright 2018 gRPC authors. + * Copyright 2024 gRPC authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. @@ -24,6 +24,7 @@ import ( "fmt" "google.golang.org/grpc/encoding" + "google.golang.org/grpc/mem" "google.golang.org/protobuf/proto" "google.golang.org/protobuf/protoadapt" ) @@ -32,28 +33,51 @@ import ( const Name = "proto" func init() { - encoding.RegisterCodec(codec{}) + encoding.RegisterCodecV2(&codecV2{}) } -// codec is a Codec implementation with protobuf. It is the default codec for gRPC. -type codec struct{} +// codec is a CodecV2 implementation with protobuf. It is the default codec for +// gRPC. +type codecV2 struct{} -func (codec) Marshal(v any) ([]byte, error) { +func (c *codecV2) Marshal(v any) (data mem.BufferSlice, err error) { vv := messageV2Of(v) if vv == nil { - return nil, fmt.Errorf("failed to marshal, message is %T, want proto.Message", v) + return nil, fmt.Errorf("proto: failed to marshal, message is %T, want proto.Message", v) } - return proto.Marshal(vv) + size := proto.Size(vv) + if mem.IsBelowBufferPoolingThreshold(size) { + buf, err := proto.Marshal(vv) + if err != nil { + return nil, err + } + data = append(data, mem.SliceBuffer(buf)) + } else { + pool := mem.DefaultBufferPool() + buf := pool.Get(size) + if _, err := (proto.MarshalOptions{}).MarshalAppend((*buf)[:0], vv); err != nil { + pool.Put(buf) + return nil, err + } + data = append(data, mem.NewBuffer(buf, pool)) + } + + return data, nil } -func (codec) Unmarshal(data []byte, v any) error { +func (c *codecV2) Unmarshal(data mem.BufferSlice, v any) (err error) { vv := messageV2Of(v) if vv == nil { return fmt.Errorf("failed to unmarshal, message is %T, want proto.Message", v) } - return proto.Unmarshal(data, vv) + buf := data.MaterializeToBuffer(mem.DefaultBufferPool()) + defer buf.Free() + // TODO: Upgrade proto.Unmarshal to support mem.BufferSlice. Right now, it's not + // really possible without a major overhaul of the proto package, but the + // vtprotobuf library may be able to support this. + return proto.Unmarshal(buf.ReadOnlyData(), vv) } func messageV2Of(v any) proto.Message { @@ -67,6 +91,6 @@ func messageV2Of(v any) proto.Message { return nil } -func (codec) Name() string { +func (c *codecV2) Name() string { return Name } diff --git a/vendor/google.golang.org/grpc/experimental/stats/metricregistry.go b/vendor/google.golang.org/grpc/experimental/stats/metricregistry.go new file mode 100644 index 0000000000000..1d827dd5d9d41 --- /dev/null +++ b/vendor/google.golang.org/grpc/experimental/stats/metricregistry.go @@ -0,0 +1,269 @@ +/* + * + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package stats + +import ( + "maps" + + "google.golang.org/grpc/grpclog" + "google.golang.org/grpc/internal" +) + +func init() { + internal.SnapshotMetricRegistryForTesting = snapshotMetricsRegistryForTesting +} + +var logger = grpclog.Component("metrics-registry") + +// DefaultMetrics are the default metrics registered through global metrics +// registry. This is written to at initialization time only, and is read only +// after initialization. +var DefaultMetrics = NewMetrics() + +// MetricDescriptor is the data for a registered metric. +type MetricDescriptor struct { + // The name of this metric. This name must be unique across the whole binary + // (including any per call metrics). See + // https://github.com/grpc/proposal/blob/master/A79-non-per-call-metrics-architecture.md#metric-instrument-naming-conventions + // for metric naming conventions. + Name Metric + // The description of this metric. + Description string + // The unit (e.g. entries, seconds) of this metric. + Unit string + // The required label keys for this metric. These are intended to + // metrics emitted from a stats handler. + Labels []string + // The optional label keys for this metric. These are intended to attached + // to metrics emitted from a stats handler if configured. + OptionalLabels []string + // Whether this metric is on by default. + Default bool + // The type of metric. This is set by the metric registry, and not intended + // to be set by a component registering a metric. + Type MetricType + // Bounds are the bounds of this metric. This only applies to histogram + // metrics. If unset or set with length 0, stats handlers will fall back to + // default bounds. + Bounds []float64 +} + +// MetricType is the type of metric. +type MetricType int + +// Type of metric supported by this instrument registry. +const ( + MetricTypeIntCount MetricType = iota + MetricTypeFloatCount + MetricTypeIntHisto + MetricTypeFloatHisto + MetricTypeIntGauge +) + +// Int64CountHandle is a typed handle for a int count metric. This handle +// is passed at the recording point in order to know which metric to record +// on. +type Int64CountHandle MetricDescriptor + +// Descriptor returns the int64 count handle typecast to a pointer to a +// MetricDescriptor. +func (h *Int64CountHandle) Descriptor() *MetricDescriptor { + return (*MetricDescriptor)(h) +} + +// Record records the int64 count value on the metrics recorder provided. +func (h *Int64CountHandle) Record(recorder MetricsRecorder, incr int64, labels ...string) { + recorder.RecordInt64Count(h, incr, labels...) +} + +// Float64CountHandle is a typed handle for a float count metric. This handle is +// passed at the recording point in order to know which metric to record on. +type Float64CountHandle MetricDescriptor + +// Descriptor returns the float64 count handle typecast to a pointer to a +// MetricDescriptor. +func (h *Float64CountHandle) Descriptor() *MetricDescriptor { + return (*MetricDescriptor)(h) +} + +// Record records the float64 count value on the metrics recorder provided. +func (h *Float64CountHandle) Record(recorder MetricsRecorder, incr float64, labels ...string) { + recorder.RecordFloat64Count(h, incr, labels...) +} + +// Int64HistoHandle is a typed handle for an int histogram metric. This handle +// is passed at the recording point in order to know which metric to record on. +type Int64HistoHandle MetricDescriptor + +// Descriptor returns the int64 histo handle typecast to a pointer to a +// MetricDescriptor. +func (h *Int64HistoHandle) Descriptor() *MetricDescriptor { + return (*MetricDescriptor)(h) +} + +// Record records the int64 histo value on the metrics recorder provided. +func (h *Int64HistoHandle) Record(recorder MetricsRecorder, incr int64, labels ...string) { + recorder.RecordInt64Histo(h, incr, labels...) +} + +// Float64HistoHandle is a typed handle for a float histogram metric. This +// handle is passed at the recording point in order to know which metric to +// record on. +type Float64HistoHandle MetricDescriptor + +// Descriptor returns the float64 histo handle typecast to a pointer to a +// MetricDescriptor. +func (h *Float64HistoHandle) Descriptor() *MetricDescriptor { + return (*MetricDescriptor)(h) +} + +// Record records the float64 histo value on the metrics recorder provided. +func (h *Float64HistoHandle) Record(recorder MetricsRecorder, incr float64, labels ...string) { + recorder.RecordFloat64Histo(h, incr, labels...) +} + +// Int64GaugeHandle is a typed handle for an int gauge metric. This handle is +// passed at the recording point in order to know which metric to record on. +type Int64GaugeHandle MetricDescriptor + +// Descriptor returns the int64 gauge handle typecast to a pointer to a +// MetricDescriptor. +func (h *Int64GaugeHandle) Descriptor() *MetricDescriptor { + return (*MetricDescriptor)(h) +} + +// Record records the int64 histo value on the metrics recorder provided. +func (h *Int64GaugeHandle) Record(recorder MetricsRecorder, incr int64, labels ...string) { + recorder.RecordInt64Gauge(h, incr, labels...) +} + +// registeredMetrics are the registered metric descriptor names. +var registeredMetrics = make(map[Metric]bool) + +// metricsRegistry contains all of the registered metrics. +// +// This is written to only at init time, and read only after that. +var metricsRegistry = make(map[Metric]*MetricDescriptor) + +// DescriptorForMetric returns the MetricDescriptor from the global registry. +// +// Returns nil if MetricDescriptor not present. +func DescriptorForMetric(metric Metric) *MetricDescriptor { + return metricsRegistry[metric] +} + +func registerMetric(name Metric, def bool) { + if registeredMetrics[name] { + logger.Fatalf("metric %v already registered", name) + } + registeredMetrics[name] = true + if def { + DefaultMetrics = DefaultMetrics.Add(name) + } +} + +// RegisterInt64Count registers the metric description onto the global registry. +// It returns a typed handle to use to recording data. +// +// NOTE: this function must only be called during initialization time (i.e. in +// an init() function), and is not thread-safe. If multiple metrics are +// registered with the same name, this function will panic. +func RegisterInt64Count(descriptor MetricDescriptor) *Int64CountHandle { + registerMetric(descriptor.Name, descriptor.Default) + descriptor.Type = MetricTypeIntCount + descPtr := &descriptor + metricsRegistry[descriptor.Name] = descPtr + return (*Int64CountHandle)(descPtr) +} + +// RegisterFloat64Count registers the metric description onto the global +// registry. It returns a typed handle to use to recording data. +// +// NOTE: this function must only be called during initialization time (i.e. in +// an init() function), and is not thread-safe. If multiple metrics are +// registered with the same name, this function will panic. +func RegisterFloat64Count(descriptor MetricDescriptor) *Float64CountHandle { + registerMetric(descriptor.Name, descriptor.Default) + descriptor.Type = MetricTypeFloatCount + descPtr := &descriptor + metricsRegistry[descriptor.Name] = descPtr + return (*Float64CountHandle)(descPtr) +} + +// RegisterInt64Histo registers the metric description onto the global registry. +// It returns a typed handle to use to recording data. +// +// NOTE: this function must only be called during initialization time (i.e. in +// an init() function), and is not thread-safe. If multiple metrics are +// registered with the same name, this function will panic. +func RegisterInt64Histo(descriptor MetricDescriptor) *Int64HistoHandle { + registerMetric(descriptor.Name, descriptor.Default) + descriptor.Type = MetricTypeIntHisto + descPtr := &descriptor + metricsRegistry[descriptor.Name] = descPtr + return (*Int64HistoHandle)(descPtr) +} + +// RegisterFloat64Histo registers the metric description onto the global +// registry. It returns a typed handle to use to recording data. +// +// NOTE: this function must only be called during initialization time (i.e. in +// an init() function), and is not thread-safe. If multiple metrics are +// registered with the same name, this function will panic. +func RegisterFloat64Histo(descriptor MetricDescriptor) *Float64HistoHandle { + registerMetric(descriptor.Name, descriptor.Default) + descriptor.Type = MetricTypeFloatHisto + descPtr := &descriptor + metricsRegistry[descriptor.Name] = descPtr + return (*Float64HistoHandle)(descPtr) +} + +// RegisterInt64Gauge registers the metric description onto the global registry. +// It returns a typed handle to use to recording data. +// +// NOTE: this function must only be called during initialization time (i.e. in +// an init() function), and is not thread-safe. If multiple metrics are +// registered with the same name, this function will panic. +func RegisterInt64Gauge(descriptor MetricDescriptor) *Int64GaugeHandle { + registerMetric(descriptor.Name, descriptor.Default) + descriptor.Type = MetricTypeIntGauge + descPtr := &descriptor + metricsRegistry[descriptor.Name] = descPtr + return (*Int64GaugeHandle)(descPtr) +} + +// snapshotMetricsRegistryForTesting snapshots the global data of the metrics +// registry. Returns a cleanup function that sets the metrics registry to its +// original state. +func snapshotMetricsRegistryForTesting() func() { + oldDefaultMetrics := DefaultMetrics + oldRegisteredMetrics := registeredMetrics + oldMetricsRegistry := metricsRegistry + + registeredMetrics = make(map[Metric]bool) + metricsRegistry = make(map[Metric]*MetricDescriptor) + maps.Copy(registeredMetrics, registeredMetrics) + maps.Copy(metricsRegistry, metricsRegistry) + + return func() { + DefaultMetrics = oldDefaultMetrics + registeredMetrics = oldRegisteredMetrics + metricsRegistry = oldMetricsRegistry + } +} diff --git a/vendor/google.golang.org/grpc/experimental/stats/metrics.go b/vendor/google.golang.org/grpc/experimental/stats/metrics.go new file mode 100644 index 0000000000000..3221f7a633a37 --- /dev/null +++ b/vendor/google.golang.org/grpc/experimental/stats/metrics.go @@ -0,0 +1,114 @@ +/* + * + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +// Package stats contains experimental metrics/stats API's. +package stats + +import "maps" + +// MetricsRecorder records on metrics derived from metric registry. +type MetricsRecorder interface { + // RecordInt64Count records the measurement alongside labels on the int + // count associated with the provided handle. + RecordInt64Count(handle *Int64CountHandle, incr int64, labels ...string) + // RecordFloat64Count records the measurement alongside labels on the float + // count associated with the provided handle. + RecordFloat64Count(handle *Float64CountHandle, incr float64, labels ...string) + // RecordInt64Histo records the measurement alongside labels on the int + // histo associated with the provided handle. + RecordInt64Histo(handle *Int64HistoHandle, incr int64, labels ...string) + // RecordFloat64Histo records the measurement alongside labels on the float + // histo associated with the provided handle. + RecordFloat64Histo(handle *Float64HistoHandle, incr float64, labels ...string) + // RecordInt64Gauge records the measurement alongside labels on the int + // gauge associated with the provided handle. + RecordInt64Gauge(handle *Int64GaugeHandle, incr int64, labels ...string) +} + +// Metric is an identifier for a metric. +type Metric string + +// Metrics is a set of metrics to record. Once created, Metrics is immutable, +// however Add and Remove can make copies with specific metrics added or +// removed, respectively. +// +// Do not construct directly; use NewMetrics instead. +type Metrics struct { + // metrics are the set of metrics to initialize. + metrics map[Metric]bool +} + +// NewMetrics returns a Metrics containing Metrics. +func NewMetrics(metrics ...Metric) *Metrics { + newMetrics := make(map[Metric]bool) + for _, metric := range metrics { + newMetrics[metric] = true + } + return &Metrics{ + metrics: newMetrics, + } +} + +// Metrics returns the metrics set. The returned map is read-only and must not +// be modified. +func (m *Metrics) Metrics() map[Metric]bool { + return m.metrics +} + +// Add adds the metrics to the metrics set and returns a new copy with the +// additional metrics. +func (m *Metrics) Add(metrics ...Metric) *Metrics { + newMetrics := make(map[Metric]bool) + for metric := range m.metrics { + newMetrics[metric] = true + } + + for _, metric := range metrics { + newMetrics[metric] = true + } + return &Metrics{ + metrics: newMetrics, + } +} + +// Join joins the metrics passed in with the metrics set, and returns a new copy +// with the merged metrics. +func (m *Metrics) Join(metrics *Metrics) *Metrics { + newMetrics := make(map[Metric]bool) + maps.Copy(newMetrics, m.metrics) + maps.Copy(newMetrics, metrics.metrics) + return &Metrics{ + metrics: newMetrics, + } +} + +// Remove removes the metrics from the metrics set and returns a new copy with +// the metrics removed. +func (m *Metrics) Remove(metrics ...Metric) *Metrics { + newMetrics := make(map[Metric]bool) + for metric := range m.metrics { + newMetrics[metric] = true + } + + for _, metric := range metrics { + delete(newMetrics, metric) + } + return &Metrics{ + metrics: newMetrics, + } +} diff --git a/vendor/google.golang.org/grpc/grpclog/component.go b/vendor/google.golang.org/grpc/grpclog/component.go index ac73c9ced2553..f1ae080dcb816 100644 --- a/vendor/google.golang.org/grpc/grpclog/component.go +++ b/vendor/google.golang.org/grpc/grpclog/component.go @@ -20,8 +20,6 @@ package grpclog import ( "fmt" - - "google.golang.org/grpc/internal/grpclog" ) // componentData records the settings for a component. @@ -33,22 +31,22 @@ var cache = map[string]*componentData{} func (c *componentData) InfoDepth(depth int, args ...any) { args = append([]any{"[" + string(c.name) + "]"}, args...) - grpclog.InfoDepth(depth+1, args...) + InfoDepth(depth+1, args...) } func (c *componentData) WarningDepth(depth int, args ...any) { args = append([]any{"[" + string(c.name) + "]"}, args...) - grpclog.WarningDepth(depth+1, args...) + WarningDepth(depth+1, args...) } func (c *componentData) ErrorDepth(depth int, args ...any) { args = append([]any{"[" + string(c.name) + "]"}, args...) - grpclog.ErrorDepth(depth+1, args...) + ErrorDepth(depth+1, args...) } func (c *componentData) FatalDepth(depth int, args ...any) { args = append([]any{"[" + string(c.name) + "]"}, args...) - grpclog.FatalDepth(depth+1, args...) + FatalDepth(depth+1, args...) } func (c *componentData) Info(args ...any) { diff --git a/vendor/google.golang.org/grpc/grpclog/grpclog.go b/vendor/google.golang.org/grpc/grpclog/grpclog.go index 16928c9cb993c..db320105e64e2 100644 --- a/vendor/google.golang.org/grpc/grpclog/grpclog.go +++ b/vendor/google.golang.org/grpc/grpclog/grpclog.go @@ -18,18 +18,15 @@ // Package grpclog defines logging for grpc. // -// All logs in transport and grpclb packages only go to verbose level 2. -// All logs in other packages in grpc are logged in spite of the verbosity level. -// -// In the default logger, -// severity level can be set by environment variable GRPC_GO_LOG_SEVERITY_LEVEL, -// verbosity level can be set by GRPC_GO_LOG_VERBOSITY_LEVEL. -package grpclog // import "google.golang.org/grpc/grpclog" +// In the default logger, severity level can be set by environment variable +// GRPC_GO_LOG_SEVERITY_LEVEL, verbosity level can be set by +// GRPC_GO_LOG_VERBOSITY_LEVEL. +package grpclog import ( "os" - "google.golang.org/grpc/internal/grpclog" + "google.golang.org/grpc/grpclog/internal" ) func init() { @@ -38,58 +35,58 @@ func init() { // V reports whether verbosity level l is at least the requested verbose level. func V(l int) bool { - return grpclog.Logger.V(l) + return internal.LoggerV2Impl.V(l) } // Info logs to the INFO log. func Info(args ...any) { - grpclog.Logger.Info(args...) + internal.LoggerV2Impl.Info(args...) } // Infof logs to the INFO log. Arguments are handled in the manner of fmt.Printf. func Infof(format string, args ...any) { - grpclog.Logger.Infof(format, args...) + internal.LoggerV2Impl.Infof(format, args...) } // Infoln logs to the INFO log. Arguments are handled in the manner of fmt.Println. func Infoln(args ...any) { - grpclog.Logger.Infoln(args...) + internal.LoggerV2Impl.Infoln(args...) } // Warning logs to the WARNING log. func Warning(args ...any) { - grpclog.Logger.Warning(args...) + internal.LoggerV2Impl.Warning(args...) } // Warningf logs to the WARNING log. Arguments are handled in the manner of fmt.Printf. func Warningf(format string, args ...any) { - grpclog.Logger.Warningf(format, args...) + internal.LoggerV2Impl.Warningf(format, args...) } // Warningln logs to the WARNING log. Arguments are handled in the manner of fmt.Println. func Warningln(args ...any) { - grpclog.Logger.Warningln(args...) + internal.LoggerV2Impl.Warningln(args...) } // Error logs to the ERROR log. func Error(args ...any) { - grpclog.Logger.Error(args...) + internal.LoggerV2Impl.Error(args...) } // Errorf logs to the ERROR log. Arguments are handled in the manner of fmt.Printf. func Errorf(format string, args ...any) { - grpclog.Logger.Errorf(format, args...) + internal.LoggerV2Impl.Errorf(format, args...) } // Errorln logs to the ERROR log. Arguments are handled in the manner of fmt.Println. func Errorln(args ...any) { - grpclog.Logger.Errorln(args...) + internal.LoggerV2Impl.Errorln(args...) } // Fatal logs to the FATAL log. Arguments are handled in the manner of fmt.Print. // It calls os.Exit() with exit code 1. func Fatal(args ...any) { - grpclog.Logger.Fatal(args...) + internal.LoggerV2Impl.Fatal(args...) // Make sure fatal logs will exit. os.Exit(1) } @@ -97,15 +94,15 @@ func Fatal(args ...any) { // Fatalf logs to the FATAL log. Arguments are handled in the manner of fmt.Printf. // It calls os.Exit() with exit code 1. func Fatalf(format string, args ...any) { - grpclog.Logger.Fatalf(format, args...) + internal.LoggerV2Impl.Fatalf(format, args...) // Make sure fatal logs will exit. os.Exit(1) } // Fatalln logs to the FATAL log. Arguments are handled in the manner of fmt.Println. -// It calle os.Exit()) with exit code 1. +// It calls os.Exit() with exit code 1. func Fatalln(args ...any) { - grpclog.Logger.Fatalln(args...) + internal.LoggerV2Impl.Fatalln(args...) // Make sure fatal logs will exit. os.Exit(1) } @@ -114,19 +111,76 @@ func Fatalln(args ...any) { // // Deprecated: use Info. func Print(args ...any) { - grpclog.Logger.Info(args...) + internal.LoggerV2Impl.Info(args...) } // Printf prints to the logger. Arguments are handled in the manner of fmt.Printf. // // Deprecated: use Infof. func Printf(format string, args ...any) { - grpclog.Logger.Infof(format, args...) + internal.LoggerV2Impl.Infof(format, args...) } // Println prints to the logger. Arguments are handled in the manner of fmt.Println. // // Deprecated: use Infoln. func Println(args ...any) { - grpclog.Logger.Infoln(args...) + internal.LoggerV2Impl.Infoln(args...) +} + +// InfoDepth logs to the INFO log at the specified depth. +// +// # Experimental +// +// Notice: This API is EXPERIMENTAL and may be changed or removed in a +// later release. +func InfoDepth(depth int, args ...any) { + if internal.DepthLoggerV2Impl != nil { + internal.DepthLoggerV2Impl.InfoDepth(depth, args...) + } else { + internal.LoggerV2Impl.Infoln(args...) + } +} + +// WarningDepth logs to the WARNING log at the specified depth. +// +// # Experimental +// +// Notice: This API is EXPERIMENTAL and may be changed or removed in a +// later release. +func WarningDepth(depth int, args ...any) { + if internal.DepthLoggerV2Impl != nil { + internal.DepthLoggerV2Impl.WarningDepth(depth, args...) + } else { + internal.LoggerV2Impl.Warningln(args...) + } +} + +// ErrorDepth logs to the ERROR log at the specified depth. +// +// # Experimental +// +// Notice: This API is EXPERIMENTAL and may be changed or removed in a +// later release. +func ErrorDepth(depth int, args ...any) { + if internal.DepthLoggerV2Impl != nil { + internal.DepthLoggerV2Impl.ErrorDepth(depth, args...) + } else { + internal.LoggerV2Impl.Errorln(args...) + } +} + +// FatalDepth logs to the FATAL log at the specified depth. +// +// # Experimental +// +// Notice: This API is EXPERIMENTAL and may be changed or removed in a +// later release. +func FatalDepth(depth int, args ...any) { + if internal.DepthLoggerV2Impl != nil { + internal.DepthLoggerV2Impl.FatalDepth(depth, args...) + } else { + internal.LoggerV2Impl.Fatalln(args...) + } + os.Exit(1) } diff --git a/vendor/google.golang.org/grpc/grpclog/internal/grpclog.go b/vendor/google.golang.org/grpc/grpclog/internal/grpclog.go new file mode 100644 index 0000000000000..59c03bc14c2a9 --- /dev/null +++ b/vendor/google.golang.org/grpc/grpclog/internal/grpclog.go @@ -0,0 +1,26 @@ +/* + * + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +// Package internal contains functionality internal to the grpclog package. +package internal + +// LoggerV2Impl is the logger used for the non-depth log functions. +var LoggerV2Impl LoggerV2 + +// DepthLoggerV2Impl is the logger used for the depth log functions. +var DepthLoggerV2Impl DepthLoggerV2 diff --git a/vendor/google.golang.org/grpc/grpclog/internal/logger.go b/vendor/google.golang.org/grpc/grpclog/internal/logger.go new file mode 100644 index 0000000000000..e524fdd40b236 --- /dev/null +++ b/vendor/google.golang.org/grpc/grpclog/internal/logger.go @@ -0,0 +1,87 @@ +/* + * + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package internal + +// Logger mimics golang's standard Logger as an interface. +// +// Deprecated: use LoggerV2. +type Logger interface { + Fatal(args ...any) + Fatalf(format string, args ...any) + Fatalln(args ...any) + Print(args ...any) + Printf(format string, args ...any) + Println(args ...any) +} + +// LoggerWrapper wraps Logger into a LoggerV2. +type LoggerWrapper struct { + Logger +} + +// Info logs to INFO log. Arguments are handled in the manner of fmt.Print. +func (l *LoggerWrapper) Info(args ...any) { + l.Logger.Print(args...) +} + +// Infoln logs to INFO log. Arguments are handled in the manner of fmt.Println. +func (l *LoggerWrapper) Infoln(args ...any) { + l.Logger.Println(args...) +} + +// Infof logs to INFO log. Arguments are handled in the manner of fmt.Printf. +func (l *LoggerWrapper) Infof(format string, args ...any) { + l.Logger.Printf(format, args...) +} + +// Warning logs to WARNING log. Arguments are handled in the manner of fmt.Print. +func (l *LoggerWrapper) Warning(args ...any) { + l.Logger.Print(args...) +} + +// Warningln logs to WARNING log. Arguments are handled in the manner of fmt.Println. +func (l *LoggerWrapper) Warningln(args ...any) { + l.Logger.Println(args...) +} + +// Warningf logs to WARNING log. Arguments are handled in the manner of fmt.Printf. +func (l *LoggerWrapper) Warningf(format string, args ...any) { + l.Logger.Printf(format, args...) +} + +// Error logs to ERROR log. Arguments are handled in the manner of fmt.Print. +func (l *LoggerWrapper) Error(args ...any) { + l.Logger.Print(args...) +} + +// Errorln logs to ERROR log. Arguments are handled in the manner of fmt.Println. +func (l *LoggerWrapper) Errorln(args ...any) { + l.Logger.Println(args...) +} + +// Errorf logs to ERROR log. Arguments are handled in the manner of fmt.Printf. +func (l *LoggerWrapper) Errorf(format string, args ...any) { + l.Logger.Printf(format, args...) +} + +// V reports whether verbosity level l is at least the requested verbose level. +func (*LoggerWrapper) V(int) bool { + // Returns true for all verbose level. + return true +} diff --git a/vendor/google.golang.org/grpc/internal/grpclog/grpclog.go b/vendor/google.golang.org/grpc/grpclog/internal/loggerv2.go similarity index 52% rename from vendor/google.golang.org/grpc/internal/grpclog/grpclog.go rename to vendor/google.golang.org/grpc/grpclog/internal/loggerv2.go index bfc45102ab245..07df71e98a87a 100644 --- a/vendor/google.golang.org/grpc/internal/grpclog/grpclog.go +++ b/vendor/google.golang.org/grpc/grpclog/internal/loggerv2.go @@ -1,6 +1,6 @@ /* * - * Copyright 2020 gRPC authors. + * Copyright 2024 gRPC authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. @@ -16,59 +16,17 @@ * */ -// Package grpclog (internal) defines depth logging for grpc. -package grpclog +package internal import ( + "encoding/json" + "fmt" + "io" + "log" "os" ) -// Logger is the logger used for the non-depth log functions. -var Logger LoggerV2 - -// DepthLogger is the logger used for the depth log functions. -var DepthLogger DepthLoggerV2 - -// InfoDepth logs to the INFO log at the specified depth. -func InfoDepth(depth int, args ...any) { - if DepthLogger != nil { - DepthLogger.InfoDepth(depth, args...) - } else { - Logger.Infoln(args...) - } -} - -// WarningDepth logs to the WARNING log at the specified depth. -func WarningDepth(depth int, args ...any) { - if DepthLogger != nil { - DepthLogger.WarningDepth(depth, args...) - } else { - Logger.Warningln(args...) - } -} - -// ErrorDepth logs to the ERROR log at the specified depth. -func ErrorDepth(depth int, args ...any) { - if DepthLogger != nil { - DepthLogger.ErrorDepth(depth, args...) - } else { - Logger.Errorln(args...) - } -} - -// FatalDepth logs to the FATAL log at the specified depth. -func FatalDepth(depth int, args ...any) { - if DepthLogger != nil { - DepthLogger.FatalDepth(depth, args...) - } else { - Logger.Fatalln(args...) - } - os.Exit(1) -} - // LoggerV2 does underlying logging work for grpclog. -// This is a copy of the LoggerV2 defined in the external grpclog package. It -// is defined here to avoid a circular dependency. type LoggerV2 interface { // Info logs to INFO log. Arguments are handled in the manner of fmt.Print. Info(args ...any) @@ -107,14 +65,13 @@ type LoggerV2 interface { // DepthLoggerV2 logs at a specified call frame. If a LoggerV2 also implements // DepthLoggerV2, the below functions will be called with the appropriate stack // depth set for trivial functions the logger may ignore. -// This is a copy of the DepthLoggerV2 defined in the external grpclog package. -// It is defined here to avoid a circular dependency. // // # Experimental // // Notice: This type is EXPERIMENTAL and may be changed or removed in a // later release. type DepthLoggerV2 interface { + LoggerV2 // InfoDepth logs to INFO log at the specified depth. Arguments are handled in the manner of fmt.Println. InfoDepth(depth int, args ...any) // WarningDepth logs to WARNING log at the specified depth. Arguments are handled in the manner of fmt.Println. @@ -124,3 +81,124 @@ type DepthLoggerV2 interface { // FatalDepth logs to FATAL log at the specified depth. Arguments are handled in the manner of fmt.Println. FatalDepth(depth int, args ...any) } + +const ( + // infoLog indicates Info severity. + infoLog int = iota + // warningLog indicates Warning severity. + warningLog + // errorLog indicates Error severity. + errorLog + // fatalLog indicates Fatal severity. + fatalLog +) + +// severityName contains the string representation of each severity. +var severityName = []string{ + infoLog: "INFO", + warningLog: "WARNING", + errorLog: "ERROR", + fatalLog: "FATAL", +} + +// loggerT is the default logger used by grpclog. +type loggerT struct { + m []*log.Logger + v int + jsonFormat bool +} + +func (g *loggerT) output(severity int, s string) { + sevStr := severityName[severity] + if !g.jsonFormat { + g.m[severity].Output(2, fmt.Sprintf("%v: %v", sevStr, s)) + return + } + // TODO: we can also include the logging component, but that needs more + // (API) changes. + b, _ := json.Marshal(map[string]string{ + "severity": sevStr, + "message": s, + }) + g.m[severity].Output(2, string(b)) +} + +func (g *loggerT) Info(args ...any) { + g.output(infoLog, fmt.Sprint(args...)) +} + +func (g *loggerT) Infoln(args ...any) { + g.output(infoLog, fmt.Sprintln(args...)) +} + +func (g *loggerT) Infof(format string, args ...any) { + g.output(infoLog, fmt.Sprintf(format, args...)) +} + +func (g *loggerT) Warning(args ...any) { + g.output(warningLog, fmt.Sprint(args...)) +} + +func (g *loggerT) Warningln(args ...any) { + g.output(warningLog, fmt.Sprintln(args...)) +} + +func (g *loggerT) Warningf(format string, args ...any) { + g.output(warningLog, fmt.Sprintf(format, args...)) +} + +func (g *loggerT) Error(args ...any) { + g.output(errorLog, fmt.Sprint(args...)) +} + +func (g *loggerT) Errorln(args ...any) { + g.output(errorLog, fmt.Sprintln(args...)) +} + +func (g *loggerT) Errorf(format string, args ...any) { + g.output(errorLog, fmt.Sprintf(format, args...)) +} + +func (g *loggerT) Fatal(args ...any) { + g.output(fatalLog, fmt.Sprint(args...)) + os.Exit(1) +} + +func (g *loggerT) Fatalln(args ...any) { + g.output(fatalLog, fmt.Sprintln(args...)) + os.Exit(1) +} + +func (g *loggerT) Fatalf(format string, args ...any) { + g.output(fatalLog, fmt.Sprintf(format, args...)) + os.Exit(1) +} + +func (g *loggerT) V(l int) bool { + return l <= g.v +} + +// LoggerV2Config configures the LoggerV2 implementation. +type LoggerV2Config struct { + // Verbosity sets the verbosity level of the logger. + Verbosity int + // FormatJSON controls whether the logger should output logs in JSON format. + FormatJSON bool +} + +// NewLoggerV2 creates a new LoggerV2 instance with the provided configuration. +// The infoW, warningW, and errorW writers are used to write log messages of +// different severity levels. +func NewLoggerV2(infoW, warningW, errorW io.Writer, c LoggerV2Config) LoggerV2 { + var m []*log.Logger + flag := log.LstdFlags + if c.FormatJSON { + flag = 0 + } + m = append(m, log.New(infoW, "", flag)) + m = append(m, log.New(io.MultiWriter(infoW, warningW), "", flag)) + ew := io.MultiWriter(infoW, warningW, errorW) // ew will be used for error and fatal. + m = append(m, log.New(ew, "", flag)) + m = append(m, log.New(ew, "", flag)) + return &loggerT{m: m, v: c.Verbosity, jsonFormat: c.FormatJSON} +} diff --git a/vendor/google.golang.org/grpc/grpclog/logger.go b/vendor/google.golang.org/grpc/grpclog/logger.go index b1674d8267ca4..4b203585707af 100644 --- a/vendor/google.golang.org/grpc/grpclog/logger.go +++ b/vendor/google.golang.org/grpc/grpclog/logger.go @@ -18,70 +18,17 @@ package grpclog -import "google.golang.org/grpc/internal/grpclog" +import "google.golang.org/grpc/grpclog/internal" // Logger mimics golang's standard Logger as an interface. // // Deprecated: use LoggerV2. -type Logger interface { - Fatal(args ...any) - Fatalf(format string, args ...any) - Fatalln(args ...any) - Print(args ...any) - Printf(format string, args ...any) - Println(args ...any) -} +type Logger internal.Logger // SetLogger sets the logger that is used in grpc. Call only from // init() functions. // // Deprecated: use SetLoggerV2. func SetLogger(l Logger) { - grpclog.Logger = &loggerWrapper{Logger: l} -} - -// loggerWrapper wraps Logger into a LoggerV2. -type loggerWrapper struct { - Logger -} - -func (g *loggerWrapper) Info(args ...any) { - g.Logger.Print(args...) -} - -func (g *loggerWrapper) Infoln(args ...any) { - g.Logger.Println(args...) -} - -func (g *loggerWrapper) Infof(format string, args ...any) { - g.Logger.Printf(format, args...) -} - -func (g *loggerWrapper) Warning(args ...any) { - g.Logger.Print(args...) -} - -func (g *loggerWrapper) Warningln(args ...any) { - g.Logger.Println(args...) -} - -func (g *loggerWrapper) Warningf(format string, args ...any) { - g.Logger.Printf(format, args...) -} - -func (g *loggerWrapper) Error(args ...any) { - g.Logger.Print(args...) -} - -func (g *loggerWrapper) Errorln(args ...any) { - g.Logger.Println(args...) -} - -func (g *loggerWrapper) Errorf(format string, args ...any) { - g.Logger.Printf(format, args...) -} - -func (g *loggerWrapper) V(l int) bool { - // Returns true for all verbose level. - return true + internal.LoggerV2Impl = &internal.LoggerWrapper{Logger: l} } diff --git a/vendor/google.golang.org/grpc/grpclog/loggerv2.go b/vendor/google.golang.org/grpc/grpclog/loggerv2.go index ecfd36d713032..892dc13d164b9 100644 --- a/vendor/google.golang.org/grpc/grpclog/loggerv2.go +++ b/vendor/google.golang.org/grpc/grpclog/loggerv2.go @@ -19,52 +19,16 @@ package grpclog import ( - "encoding/json" - "fmt" "io" - "log" "os" "strconv" "strings" - "google.golang.org/grpc/internal/grpclog" + "google.golang.org/grpc/grpclog/internal" ) // LoggerV2 does underlying logging work for grpclog. -type LoggerV2 interface { - // Info logs to INFO log. Arguments are handled in the manner of fmt.Print. - Info(args ...any) - // Infoln logs to INFO log. Arguments are handled in the manner of fmt.Println. - Infoln(args ...any) - // Infof logs to INFO log. Arguments are handled in the manner of fmt.Printf. - Infof(format string, args ...any) - // Warning logs to WARNING log. Arguments are handled in the manner of fmt.Print. - Warning(args ...any) - // Warningln logs to WARNING log. Arguments are handled in the manner of fmt.Println. - Warningln(args ...any) - // Warningf logs to WARNING log. Arguments are handled in the manner of fmt.Printf. - Warningf(format string, args ...any) - // Error logs to ERROR log. Arguments are handled in the manner of fmt.Print. - Error(args ...any) - // Errorln logs to ERROR log. Arguments are handled in the manner of fmt.Println. - Errorln(args ...any) - // Errorf logs to ERROR log. Arguments are handled in the manner of fmt.Printf. - Errorf(format string, args ...any) - // Fatal logs to ERROR log. Arguments are handled in the manner of fmt.Print. - // gRPC ensures that all Fatal logs will exit with os.Exit(1). - // Implementations may also call os.Exit() with a non-zero exit code. - Fatal(args ...any) - // Fatalln logs to ERROR log. Arguments are handled in the manner of fmt.Println. - // gRPC ensures that all Fatal logs will exit with os.Exit(1). - // Implementations may also call os.Exit() with a non-zero exit code. - Fatalln(args ...any) - // Fatalf logs to ERROR log. Arguments are handled in the manner of fmt.Printf. - // gRPC ensures that all Fatal logs will exit with os.Exit(1). - // Implementations may also call os.Exit() with a non-zero exit code. - Fatalf(format string, args ...any) - // V reports whether verbosity level l is at least the requested verbose level. - V(l int) bool -} +type LoggerV2 internal.LoggerV2 // SetLoggerV2 sets logger that is used in grpc to a V2 logger. // Not mutex-protected, should be called before any gRPC functions. @@ -72,34 +36,8 @@ func SetLoggerV2(l LoggerV2) { if _, ok := l.(*componentData); ok { panic("cannot use component logger as grpclog logger") } - grpclog.Logger = l - grpclog.DepthLogger, _ = l.(grpclog.DepthLoggerV2) -} - -const ( - // infoLog indicates Info severity. - infoLog int = iota - // warningLog indicates Warning severity. - warningLog - // errorLog indicates Error severity. - errorLog - // fatalLog indicates Fatal severity. - fatalLog -) - -// severityName contains the string representation of each severity. -var severityName = []string{ - infoLog: "INFO", - warningLog: "WARNING", - errorLog: "ERROR", - fatalLog: "FATAL", -} - -// loggerT is the default logger used by grpclog. -type loggerT struct { - m []*log.Logger - v int - jsonFormat bool + internal.LoggerV2Impl = l + internal.DepthLoggerV2Impl, _ = l.(internal.DepthLoggerV2) } // NewLoggerV2 creates a loggerV2 with the provided writers. @@ -108,32 +46,13 @@ type loggerT struct { // Warning logs will be written to warningW and infoW. // Info logs will be written to infoW. func NewLoggerV2(infoW, warningW, errorW io.Writer) LoggerV2 { - return newLoggerV2WithConfig(infoW, warningW, errorW, loggerV2Config{}) + return internal.NewLoggerV2(infoW, warningW, errorW, internal.LoggerV2Config{}) } // NewLoggerV2WithVerbosity creates a loggerV2 with the provided writers and // verbosity level. func NewLoggerV2WithVerbosity(infoW, warningW, errorW io.Writer, v int) LoggerV2 { - return newLoggerV2WithConfig(infoW, warningW, errorW, loggerV2Config{verbose: v}) -} - -type loggerV2Config struct { - verbose int - jsonFormat bool -} - -func newLoggerV2WithConfig(infoW, warningW, errorW io.Writer, c loggerV2Config) LoggerV2 { - var m []*log.Logger - flag := log.LstdFlags - if c.jsonFormat { - flag = 0 - } - m = append(m, log.New(infoW, "", flag)) - m = append(m, log.New(io.MultiWriter(infoW, warningW), "", flag)) - ew := io.MultiWriter(infoW, warningW, errorW) // ew will be used for error and fatal. - m = append(m, log.New(ew, "", flag)) - m = append(m, log.New(ew, "", flag)) - return &loggerT{m: m, v: c.verbose, jsonFormat: c.jsonFormat} + return internal.NewLoggerV2(infoW, warningW, errorW, internal.LoggerV2Config{Verbosity: v}) } // newLoggerV2 creates a loggerV2 to be used as default logger. @@ -161,80 +80,10 @@ func newLoggerV2() LoggerV2 { jsonFormat := strings.EqualFold(os.Getenv("GRPC_GO_LOG_FORMATTER"), "json") - return newLoggerV2WithConfig(infoW, warningW, errorW, loggerV2Config{ - verbose: v, - jsonFormat: jsonFormat, - }) -} - -func (g *loggerT) output(severity int, s string) { - sevStr := severityName[severity] - if !g.jsonFormat { - g.m[severity].Output(2, fmt.Sprintf("%v: %v", sevStr, s)) - return - } - // TODO: we can also include the logging component, but that needs more - // (API) changes. - b, _ := json.Marshal(map[string]string{ - "severity": sevStr, - "message": s, + return internal.NewLoggerV2(infoW, warningW, errorW, internal.LoggerV2Config{ + Verbosity: v, + FormatJSON: jsonFormat, }) - g.m[severity].Output(2, string(b)) -} - -func (g *loggerT) Info(args ...any) { - g.output(infoLog, fmt.Sprint(args...)) -} - -func (g *loggerT) Infoln(args ...any) { - g.output(infoLog, fmt.Sprintln(args...)) -} - -func (g *loggerT) Infof(format string, args ...any) { - g.output(infoLog, fmt.Sprintf(format, args...)) -} - -func (g *loggerT) Warning(args ...any) { - g.output(warningLog, fmt.Sprint(args...)) -} - -func (g *loggerT) Warningln(args ...any) { - g.output(warningLog, fmt.Sprintln(args...)) -} - -func (g *loggerT) Warningf(format string, args ...any) { - g.output(warningLog, fmt.Sprintf(format, args...)) -} - -func (g *loggerT) Error(args ...any) { - g.output(errorLog, fmt.Sprint(args...)) -} - -func (g *loggerT) Errorln(args ...any) { - g.output(errorLog, fmt.Sprintln(args...)) -} - -func (g *loggerT) Errorf(format string, args ...any) { - g.output(errorLog, fmt.Sprintf(format, args...)) -} - -func (g *loggerT) Fatal(args ...any) { - g.output(fatalLog, fmt.Sprint(args...)) - os.Exit(1) -} - -func (g *loggerT) Fatalln(args ...any) { - g.output(fatalLog, fmt.Sprintln(args...)) - os.Exit(1) -} - -func (g *loggerT) Fatalf(format string, args ...any) { - g.output(fatalLog, fmt.Sprintf(format, args...)) - os.Exit(1) -} - -func (g *loggerT) V(l int) bool { - return l <= g.v } // DepthLoggerV2 logs at a specified call frame. If a LoggerV2 also implements @@ -245,14 +94,4 @@ func (g *loggerT) V(l int) bool { // // Notice: This type is EXPERIMENTAL and may be changed or removed in a // later release. -type DepthLoggerV2 interface { - LoggerV2 - // InfoDepth logs to INFO log at the specified depth. Arguments are handled in the manner of fmt.Println. - InfoDepth(depth int, args ...any) - // WarningDepth logs to WARNING log at the specified depth. Arguments are handled in the manner of fmt.Println. - WarningDepth(depth int, args ...any) - // ErrorDepth logs to ERROR log at the specified depth. Arguments are handled in the manner of fmt.Println. - ErrorDepth(depth int, args ...any) - // FatalDepth logs to FATAL log at the specified depth. Arguments are handled in the manner of fmt.Println. - FatalDepth(depth int, args ...any) -} +type DepthLoggerV2 internal.DepthLoggerV2 diff --git a/vendor/google.golang.org/grpc/health/grpc_health_v1/health.pb.go b/vendor/google.golang.org/grpc/health/grpc_health_v1/health.pb.go index 38b8835073502..d92335445f650 100644 --- a/vendor/google.golang.org/grpc/health/grpc_health_v1/health.pb.go +++ b/vendor/google.golang.org/grpc/health/grpc_health_v1/health.pb.go @@ -17,8 +17,8 @@ // Code generated by protoc-gen-go. DO NOT EDIT. // versions: -// protoc-gen-go v1.34.1 -// protoc v4.25.2 +// protoc-gen-go v1.34.2 +// protoc v5.27.1 // source: grpc/health/v1/health.proto package grpc_health_v1 @@ -237,7 +237,7 @@ func file_grpc_health_v1_health_proto_rawDescGZIP() []byte { var file_grpc_health_v1_health_proto_enumTypes = make([]protoimpl.EnumInfo, 1) var file_grpc_health_v1_health_proto_msgTypes = make([]protoimpl.MessageInfo, 2) -var file_grpc_health_v1_health_proto_goTypes = []interface{}{ +var file_grpc_health_v1_health_proto_goTypes = []any{ (HealthCheckResponse_ServingStatus)(0), // 0: grpc.health.v1.HealthCheckResponse.ServingStatus (*HealthCheckRequest)(nil), // 1: grpc.health.v1.HealthCheckRequest (*HealthCheckResponse)(nil), // 2: grpc.health.v1.HealthCheckResponse @@ -261,7 +261,7 @@ func file_grpc_health_v1_health_proto_init() { return } if !protoimpl.UnsafeEnabled { - file_grpc_health_v1_health_proto_msgTypes[0].Exporter = func(v interface{}, i int) interface{} { + file_grpc_health_v1_health_proto_msgTypes[0].Exporter = func(v any, i int) any { switch v := v.(*HealthCheckRequest); i { case 0: return &v.state @@ -273,7 +273,7 @@ func file_grpc_health_v1_health_proto_init() { return nil } } - file_grpc_health_v1_health_proto_msgTypes[1].Exporter = func(v interface{}, i int) interface{} { + file_grpc_health_v1_health_proto_msgTypes[1].Exporter = func(v any, i int) any { switch v := v.(*HealthCheckResponse); i { case 0: return &v.state diff --git a/vendor/google.golang.org/grpc/health/grpc_health_v1/health_grpc.pb.go b/vendor/google.golang.org/grpc/health/grpc_health_v1/health_grpc.pb.go index 51b736ba06e5f..f96b8ab4927e1 100644 --- a/vendor/google.golang.org/grpc/health/grpc_health_v1/health_grpc.pb.go +++ b/vendor/google.golang.org/grpc/health/grpc_health_v1/health_grpc.pb.go @@ -17,8 +17,8 @@ // Code generated by protoc-gen-go-grpc. DO NOT EDIT. // versions: -// - protoc-gen-go-grpc v1.4.0 -// - protoc v4.25.2 +// - protoc-gen-go-grpc v1.5.1 +// - protoc v5.27.1 // source: grpc/health/v1/health.proto package grpc_health_v1 @@ -32,8 +32,8 @@ import ( // This is a compile-time assertion to ensure that this generated file // is compatible with the grpc package it is being compiled against. -// Requires gRPC-Go v1.62.0 or later. -const _ = grpc.SupportPackageIsVersion8 +// Requires gRPC-Go v1.64.0 or later. +const _ = grpc.SupportPackageIsVersion9 const ( Health_Check_FullMethodName = "/grpc.health.v1.Health/Check" @@ -73,7 +73,7 @@ type HealthClient interface { // should assume this method is not supported and should not retry the // call. If the call terminates with any other status (including OK), // clients should retry the call with appropriate exponential backoff. - Watch(ctx context.Context, in *HealthCheckRequest, opts ...grpc.CallOption) (Health_WatchClient, error) + Watch(ctx context.Context, in *HealthCheckRequest, opts ...grpc.CallOption) (grpc.ServerStreamingClient[HealthCheckResponse], error) } type healthClient struct { @@ -94,13 +94,13 @@ func (c *healthClient) Check(ctx context.Context, in *HealthCheckRequest, opts . return out, nil } -func (c *healthClient) Watch(ctx context.Context, in *HealthCheckRequest, opts ...grpc.CallOption) (Health_WatchClient, error) { +func (c *healthClient) Watch(ctx context.Context, in *HealthCheckRequest, opts ...grpc.CallOption) (grpc.ServerStreamingClient[HealthCheckResponse], error) { cOpts := append([]grpc.CallOption{grpc.StaticMethod()}, opts...) stream, err := c.cc.NewStream(ctx, &Health_ServiceDesc.Streams[0], Health_Watch_FullMethodName, cOpts...) if err != nil { return nil, err } - x := &healthWatchClient{ClientStream: stream} + x := &grpc.GenericClientStream[HealthCheckRequest, HealthCheckResponse]{ClientStream: stream} if err := x.ClientStream.SendMsg(in); err != nil { return nil, err } @@ -110,26 +110,12 @@ func (c *healthClient) Watch(ctx context.Context, in *HealthCheckRequest, opts . return x, nil } -type Health_WatchClient interface { - Recv() (*HealthCheckResponse, error) - grpc.ClientStream -} - -type healthWatchClient struct { - grpc.ClientStream -} - -func (x *healthWatchClient) Recv() (*HealthCheckResponse, error) { - m := new(HealthCheckResponse) - if err := x.ClientStream.RecvMsg(m); err != nil { - return nil, err - } - return m, nil -} +// This type alias is provided for backwards compatibility with existing code that references the prior non-generic stream type by name. +type Health_WatchClient = grpc.ServerStreamingClient[HealthCheckResponse] // HealthServer is the server API for Health service. // All implementations should embed UnimplementedHealthServer -// for forward compatibility +// for forward compatibility. // // Health is gRPC's mechanism for checking whether a server is able to handle // RPCs. Its semantics are documented in @@ -160,19 +146,23 @@ type HealthServer interface { // should assume this method is not supported and should not retry the // call. If the call terminates with any other status (including OK), // clients should retry the call with appropriate exponential backoff. - Watch(*HealthCheckRequest, Health_WatchServer) error + Watch(*HealthCheckRequest, grpc.ServerStreamingServer[HealthCheckResponse]) error } -// UnimplementedHealthServer should be embedded to have forward compatible implementations. -type UnimplementedHealthServer struct { -} +// UnimplementedHealthServer should be embedded to have +// forward compatible implementations. +// +// NOTE: this should be embedded by value instead of pointer to avoid a nil +// pointer dereference when methods are called. +type UnimplementedHealthServer struct{} func (UnimplementedHealthServer) Check(context.Context, *HealthCheckRequest) (*HealthCheckResponse, error) { return nil, status.Errorf(codes.Unimplemented, "method Check not implemented") } -func (UnimplementedHealthServer) Watch(*HealthCheckRequest, Health_WatchServer) error { +func (UnimplementedHealthServer) Watch(*HealthCheckRequest, grpc.ServerStreamingServer[HealthCheckResponse]) error { return status.Errorf(codes.Unimplemented, "method Watch not implemented") } +func (UnimplementedHealthServer) testEmbeddedByValue() {} // UnsafeHealthServer may be embedded to opt out of forward compatibility for this service. // Use of this interface is not recommended, as added methods to HealthServer will @@ -182,6 +172,13 @@ type UnsafeHealthServer interface { } func RegisterHealthServer(s grpc.ServiceRegistrar, srv HealthServer) { + // If the following call panics, it indicates UnimplementedHealthServer was + // embedded by pointer and is nil. This will cause panics if an + // unimplemented method is ever invoked, so we test this at initialization + // time to prevent it from happening at runtime later due to I/O. + if t, ok := srv.(interface{ testEmbeddedByValue() }); ok { + t.testEmbeddedByValue() + } s.RegisterService(&Health_ServiceDesc, srv) } @@ -208,21 +205,11 @@ func _Health_Watch_Handler(srv interface{}, stream grpc.ServerStream) error { if err := stream.RecvMsg(m); err != nil { return err } - return srv.(HealthServer).Watch(m, &healthWatchServer{ServerStream: stream}) -} - -type Health_WatchServer interface { - Send(*HealthCheckResponse) error - grpc.ServerStream + return srv.(HealthServer).Watch(m, &grpc.GenericServerStream[HealthCheckRequest, HealthCheckResponse]{ServerStream: stream}) } -type healthWatchServer struct { - grpc.ServerStream -} - -func (x *healthWatchServer) Send(m *HealthCheckResponse) error { - return x.ServerStream.SendMsg(m) -} +// This type alias is provided for backwards compatibility with existing code that references the prior non-generic stream type by name. +type Health_WatchServer = grpc.ServerStreamingServer[HealthCheckResponse] // Health_ServiceDesc is the grpc.ServiceDesc for Health service. // It's only intended for direct use with grpc.RegisterService, diff --git a/vendor/google.golang.org/grpc/health/server.go b/vendor/google.golang.org/grpc/health/server.go index cce6312d77f9c..d4b4b7081590e 100644 --- a/vendor/google.golang.org/grpc/health/server.go +++ b/vendor/google.golang.org/grpc/health/server.go @@ -51,7 +51,7 @@ func NewServer() *Server { } // Check implements `service Health`. -func (s *Server) Check(ctx context.Context, in *healthpb.HealthCheckRequest) (*healthpb.HealthCheckResponse, error) { +func (s *Server) Check(_ context.Context, in *healthpb.HealthCheckRequest) (*healthpb.HealthCheckResponse, error) { s.mu.RLock() defer s.mu.RUnlock() if servingStatus, ok := s.statusMap[in.Service]; ok { diff --git a/vendor/google.golang.org/grpc/internal/binarylog/method_logger.go b/vendor/google.golang.org/grpc/internal/binarylog/method_logger.go index aa4505a871dfb..9669328914ad4 100644 --- a/vendor/google.golang.org/grpc/internal/binarylog/method_logger.go +++ b/vendor/google.golang.org/grpc/internal/binarylog/method_logger.go @@ -106,7 +106,7 @@ func (ml *TruncatingMethodLogger) Build(c LogEntryConfig) *binlogpb.GrpcLogEntry } // Log creates a proto binary log entry, and logs it to the sink. -func (ml *TruncatingMethodLogger) Log(ctx context.Context, c LogEntryConfig) { +func (ml *TruncatingMethodLogger) Log(_ context.Context, c LogEntryConfig) { ml.sink.Write(ml.Build(c)) } diff --git a/vendor/google.golang.org/grpc/internal/channelz/channelmap.go b/vendor/google.golang.org/grpc/internal/channelz/channelmap.go index dfe18b08925d9..64c791953d017 100644 --- a/vendor/google.golang.org/grpc/internal/channelz/channelmap.go +++ b/vendor/google.golang.org/grpc/internal/channelz/channelmap.go @@ -46,7 +46,7 @@ type entry interface { // channelMap is the storage data structure for channelz. // -// Methods of channelMap can be divided in two two categories with respect to +// Methods of channelMap can be divided into two categories with respect to // locking. // // 1. Methods acquire the global lock. @@ -234,13 +234,6 @@ func copyMap(m map[int64]string) map[int64]string { return n } -func min(a, b int) int { - if a < b { - return a - } - return b -} - func (c *channelMap) getTopChannels(id int64, maxResults int) ([]*Channel, bool) { if maxResults <= 0 { maxResults = EntriesPerPage diff --git a/vendor/google.golang.org/grpc/internal/channelz/funcs.go b/vendor/google.golang.org/grpc/internal/channelz/funcs.go index 03e24e1507aa6..078bb81238bc4 100644 --- a/vendor/google.golang.org/grpc/internal/channelz/funcs.go +++ b/vendor/google.golang.org/grpc/internal/channelz/funcs.go @@ -33,7 +33,7 @@ var ( // outside this package except by tests. IDGen IDGenerator - db *channelMap = newChannelMap() + db = newChannelMap() // EntriesPerPage defines the number of channelz entries to be shown on a web page. EntriesPerPage = 50 curState int32 diff --git a/vendor/google.golang.org/grpc/internal/channelz/syscall_nonlinux.go b/vendor/google.golang.org/grpc/internal/channelz/syscall_nonlinux.go index d1ed8df6a5186..0e6e18e185c7a 100644 --- a/vendor/google.golang.org/grpc/internal/channelz/syscall_nonlinux.go +++ b/vendor/google.golang.org/grpc/internal/channelz/syscall_nonlinux.go @@ -35,13 +35,13 @@ type SocketOptionData struct { // Getsockopt defines the function to get socket options requested by channelz. // It is to be passed to syscall.RawConn.Control(). // Windows OS doesn't support Socket Option -func (s *SocketOptionData) Getsockopt(fd uintptr) { +func (s *SocketOptionData) Getsockopt(uintptr) { once.Do(func() { logger.Warning("Channelz: socket options are not supported on non-linux environments") }) } // GetSocketOption gets the socket option info of the conn. -func GetSocketOption(c any) *SocketOptionData { +func GetSocketOption(any) *SocketOptionData { return nil } diff --git a/vendor/google.golang.org/grpc/internal/envconfig/envconfig.go b/vendor/google.golang.org/grpc/internal/envconfig/envconfig.go index d906487139445..452985f8d8f1b 100644 --- a/vendor/google.golang.org/grpc/internal/envconfig/envconfig.go +++ b/vendor/google.golang.org/grpc/internal/envconfig/envconfig.go @@ -45,7 +45,11 @@ var ( // option is present for backward compatibility. This option may be overridden // by setting the environment variable "GRPC_ENFORCE_ALPN_ENABLED" to "true" // or "false". - EnforceALPNEnabled = boolFromEnv("GRPC_ENFORCE_ALPN_ENABLED", false) + EnforceALPNEnabled = boolFromEnv("GRPC_ENFORCE_ALPN_ENABLED", true) + // XDSFallbackSupport is the env variable that controls whether support for + // xDS fallback is turned on. If this is unset or is false, only the first + // xDS server in the list of server configs will be used. + XDSFallbackSupport = boolFromEnv("GRPC_EXPERIMENTAL_XDS_FALLBACK", false) ) func boolFromEnv(envVar string, def bool) bool { diff --git a/vendor/google.golang.org/grpc/internal/experimental.go b/vendor/google.golang.org/grpc/internal/experimental.go index 7f7044e1731c8..7617be2158957 100644 --- a/vendor/google.golang.org/grpc/internal/experimental.go +++ b/vendor/google.golang.org/grpc/internal/experimental.go @@ -18,11 +18,11 @@ package internal var ( - // WithRecvBufferPool is implemented by the grpc package and returns a dial + // WithBufferPool is implemented by the grpc package and returns a dial // option to configure a shared buffer pool for a grpc.ClientConn. - WithRecvBufferPool any // func (grpc.SharedBufferPool) grpc.DialOption + WithBufferPool any // func (grpc.SharedBufferPool) grpc.DialOption - // RecvBufferPool is implemented by the grpc package and returns a server + // BufferPool is implemented by the grpc package and returns a server // option to configure a shared buffer pool for a grpc.Server. - RecvBufferPool any // func (grpc.SharedBufferPool) grpc.ServerOption + BufferPool any // func (grpc.SharedBufferPool) grpc.ServerOption ) diff --git a/vendor/google.golang.org/grpc/internal/grpclog/prefixLogger.go b/vendor/google.golang.org/grpc/internal/grpclog/prefix_logger.go similarity index 63% rename from vendor/google.golang.org/grpc/internal/grpclog/prefixLogger.go rename to vendor/google.golang.org/grpc/internal/grpclog/prefix_logger.go index faa998de7632b..092ad187a2c8d 100644 --- a/vendor/google.golang.org/grpc/internal/grpclog/prefixLogger.go +++ b/vendor/google.golang.org/grpc/internal/grpclog/prefix_logger.go @@ -16,17 +16,21 @@ * */ +// Package grpclog provides logging functionality for internal gRPC packages, +// outside of the functionality provided by the external `grpclog` package. package grpclog import ( "fmt" + + "google.golang.org/grpc/grpclog" ) // PrefixLogger does logging with a prefix. // // Logging method on a nil logs without any prefix. type PrefixLogger struct { - logger DepthLoggerV2 + logger grpclog.DepthLoggerV2 prefix string } @@ -38,7 +42,7 @@ func (pl *PrefixLogger) Infof(format string, args ...any) { pl.logger.InfoDepth(1, fmt.Sprintf(format, args...)) return } - InfoDepth(1, fmt.Sprintf(format, args...)) + grpclog.InfoDepth(1, fmt.Sprintf(format, args...)) } // Warningf does warning logging. @@ -48,7 +52,7 @@ func (pl *PrefixLogger) Warningf(format string, args ...any) { pl.logger.WarningDepth(1, fmt.Sprintf(format, args...)) return } - WarningDepth(1, fmt.Sprintf(format, args...)) + grpclog.WarningDepth(1, fmt.Sprintf(format, args...)) } // Errorf does error logging. @@ -58,36 +62,18 @@ func (pl *PrefixLogger) Errorf(format string, args ...any) { pl.logger.ErrorDepth(1, fmt.Sprintf(format, args...)) return } - ErrorDepth(1, fmt.Sprintf(format, args...)) -} - -// Debugf does info logging at verbose level 2. -func (pl *PrefixLogger) Debugf(format string, args ...any) { - // TODO(6044): Refactor interfaces LoggerV2 and DepthLogger, and maybe - // rewrite PrefixLogger a little to ensure that we don't use the global - // `Logger` here, and instead use the `logger` field. - if !Logger.V(2) { - return - } - if pl != nil { - // Handle nil, so the tests can pass in a nil logger. - format = pl.prefix + format - pl.logger.InfoDepth(1, fmt.Sprintf(format, args...)) - return - } - InfoDepth(1, fmt.Sprintf(format, args...)) - + grpclog.ErrorDepth(1, fmt.Sprintf(format, args...)) } // V reports whether verbosity level l is at least the requested verbose level. func (pl *PrefixLogger) V(l int) bool { - // TODO(6044): Refactor interfaces LoggerV2 and DepthLogger, and maybe - // rewrite PrefixLogger a little to ensure that we don't use the global - // `Logger` here, and instead use the `logger` field. - return Logger.V(l) + if pl != nil { + return pl.logger.V(l) + } + return true } // NewPrefixLogger creates a prefix logger with the given prefix. -func NewPrefixLogger(logger DepthLoggerV2, prefix string) *PrefixLogger { +func NewPrefixLogger(logger grpclog.DepthLoggerV2, prefix string) *PrefixLogger { return &PrefixLogger{logger: logger, prefix: prefix} } diff --git a/vendor/google.golang.org/grpc/internal/grpcsync/callback_serializer.go b/vendor/google.golang.org/grpc/internal/grpcsync/callback_serializer.go index f7f40a16acee5..19b9d639275a8 100644 --- a/vendor/google.golang.org/grpc/internal/grpcsync/callback_serializer.go +++ b/vendor/google.golang.org/grpc/internal/grpcsync/callback_serializer.go @@ -53,16 +53,28 @@ func NewCallbackSerializer(ctx context.Context) *CallbackSerializer { return cs } -// Schedule adds a callback to be scheduled after existing callbacks are run. +// TrySchedule tries to schedules the provided callback function f to be +// executed in the order it was added. This is a best-effort operation. If the +// context passed to NewCallbackSerializer was canceled before this method is +// called, the callback will not be scheduled. // // Callbacks are expected to honor the context when performing any blocking // operations, and should return early when the context is canceled. +func (cs *CallbackSerializer) TrySchedule(f func(ctx context.Context)) { + cs.callbacks.Put(f) +} + +// ScheduleOr schedules the provided callback function f to be executed in the +// order it was added. If the context passed to NewCallbackSerializer has been +// canceled before this method is called, the onFailure callback will be +// executed inline instead. // -// Return value indicates if the callback was successfully added to the list of -// callbacks to be executed by the serializer. It is not possible to add -// callbacks once the context passed to NewCallbackSerializer is cancelled. -func (cs *CallbackSerializer) Schedule(f func(ctx context.Context)) bool { - return cs.callbacks.Put(f) == nil +// Callbacks are expected to honor the context when performing any blocking +// operations, and should return early when the context is canceled. +func (cs *CallbackSerializer) ScheduleOr(f func(ctx context.Context), onFailure func()) { + if cs.callbacks.Put(f) != nil { + onFailure() + } } func (cs *CallbackSerializer) run(ctx context.Context) { diff --git a/vendor/google.golang.org/grpc/internal/grpcsync/pubsub.go b/vendor/google.golang.org/grpc/internal/grpcsync/pubsub.go index aef8cec1ab0cd..6d8c2f518dff7 100644 --- a/vendor/google.golang.org/grpc/internal/grpcsync/pubsub.go +++ b/vendor/google.golang.org/grpc/internal/grpcsync/pubsub.go @@ -77,7 +77,7 @@ func (ps *PubSub) Subscribe(sub Subscriber) (cancel func()) { if ps.msg != nil { msg := ps.msg - ps.cs.Schedule(func(context.Context) { + ps.cs.TrySchedule(func(context.Context) { ps.mu.Lock() defer ps.mu.Unlock() if !ps.subscribers[sub] { @@ -103,7 +103,7 @@ func (ps *PubSub) Publish(msg any) { ps.msg = msg for sub := range ps.subscribers { s := sub - ps.cs.Schedule(func(context.Context) { + ps.cs.TrySchedule(func(context.Context) { ps.mu.Lock() defer ps.mu.Unlock() if !ps.subscribers[s] { diff --git a/vendor/google.golang.org/grpc/internal/internal.go b/vendor/google.golang.org/grpc/internal/internal.go index 5d66539869232..7aae9240ffc07 100644 --- a/vendor/google.golang.org/grpc/internal/internal.go +++ b/vendor/google.golang.org/grpc/internal/internal.go @@ -183,7 +183,7 @@ var ( // GRPCResolverSchemeExtraMetadata determines when gRPC will add extra // metadata to RPCs. - GRPCResolverSchemeExtraMetadata string = "xds" + GRPCResolverSchemeExtraMetadata = "xds" // EnterIdleModeForTesting gets the ClientConn to enter IDLE mode. EnterIdleModeForTesting any // func(*grpc.ClientConn) @@ -203,11 +203,31 @@ var ( // UserSetDefaultScheme is set to true if the user has overridden the // default resolver scheme. - UserSetDefaultScheme bool = false + UserSetDefaultScheme = false // ShuffleAddressListForTesting pseudo-randomizes the order of addresses. n // is the number of elements. swap swaps the elements with indexes i and j. ShuffleAddressListForTesting any // func(n int, swap func(i, j int)) + + // ConnectedAddress returns the connected address for a SubConnState. The + // address is only valid if the state is READY. + ConnectedAddress any // func (scs SubConnState) resolver.Address + + // SetConnectedAddress sets the connected address for a SubConnState. + SetConnectedAddress any // func(scs *SubConnState, addr resolver.Address) + + // SnapshotMetricRegistryForTesting snapshots the global data of the metric + // registry. Returns a cleanup function that sets the metric registry to its + // original state. Only called in testing functions. + SnapshotMetricRegistryForTesting func() func() + + // SetDefaultBufferPoolForTesting updates the default buffer pool, for + // testing purposes. + SetDefaultBufferPoolForTesting any // func(mem.BufferPool) + + // SetBufferPoolingThresholdForTesting updates the buffer pooling threshold, for + // testing purposes. + SetBufferPoolingThresholdForTesting any // func(int) ) // HealthChecker defines the signature of the client-side LB channel health diff --git a/vendor/google.golang.org/grpc/internal/resolver/passthrough/passthrough.go b/vendor/google.golang.org/grpc/internal/resolver/passthrough/passthrough.go index afac56572ad55..b901c7bace506 100644 --- a/vendor/google.golang.org/grpc/internal/resolver/passthrough/passthrough.go +++ b/vendor/google.golang.org/grpc/internal/resolver/passthrough/passthrough.go @@ -55,7 +55,7 @@ func (r *passthroughResolver) start() { r.cc.UpdateState(resolver.State{Addresses: []resolver.Address{{Addr: r.target.Endpoint()}}}) } -func (*passthroughResolver) ResolveNow(o resolver.ResolveNowOptions) {} +func (*passthroughResolver) ResolveNow(resolver.ResolveNowOptions) {} func (*passthroughResolver) Close() {} diff --git a/vendor/google.golang.org/grpc/internal/stats/labels.go b/vendor/google.golang.org/grpc/internal/stats/labels.go new file mode 100644 index 0000000000000..fd33af51ae896 --- /dev/null +++ b/vendor/google.golang.org/grpc/internal/stats/labels.go @@ -0,0 +1,42 @@ +/* + * + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +// Package stats provides internal stats related functionality. +package stats + +import "context" + +// Labels are the labels for metrics. +type Labels struct { + // TelemetryLabels are the telemetry labels to record. + TelemetryLabels map[string]string +} + +type labelsKey struct{} + +// GetLabels returns the Labels stored in the context, or nil if there is one. +func GetLabels(ctx context.Context) *Labels { + labels, _ := ctx.Value(labelsKey{}).(*Labels) + return labels +} + +// SetLabels sets the Labels in the context. +func SetLabels(ctx context.Context, labels *Labels) context.Context { + // could also append + return context.WithValue(ctx, labelsKey{}, labels) +} diff --git a/vendor/google.golang.org/grpc/internal/stats/metrics_recorder_list.go b/vendor/google.golang.org/grpc/internal/stats/metrics_recorder_list.go new file mode 100644 index 0000000000000..be110d41f9a42 --- /dev/null +++ b/vendor/google.golang.org/grpc/internal/stats/metrics_recorder_list.go @@ -0,0 +1,95 @@ +/* + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package stats + +import ( + "fmt" + + estats "google.golang.org/grpc/experimental/stats" + "google.golang.org/grpc/stats" +) + +// MetricsRecorderList forwards Record calls to all of its metricsRecorders. +// +// It eats any record calls where the label values provided do not match the +// number of label keys. +type MetricsRecorderList struct { + // metricsRecorders are the metrics recorders this list will forward to. + metricsRecorders []estats.MetricsRecorder +} + +// NewMetricsRecorderList creates a new metric recorder list with all the stats +// handlers provided which implement the MetricsRecorder interface. +// If no stats handlers provided implement the MetricsRecorder interface, +// the MetricsRecorder list returned is a no-op. +func NewMetricsRecorderList(shs []stats.Handler) *MetricsRecorderList { + var mrs []estats.MetricsRecorder + for _, sh := range shs { + if mr, ok := sh.(estats.MetricsRecorder); ok { + mrs = append(mrs, mr) + } + } + return &MetricsRecorderList{ + metricsRecorders: mrs, + } +} + +func verifyLabels(desc *estats.MetricDescriptor, labelsRecv ...string) { + if got, want := len(labelsRecv), len(desc.Labels)+len(desc.OptionalLabels); got != want { + panic(fmt.Sprintf("Received %d labels in call to record metric %q, but expected %d.", got, desc.Name, want)) + } +} + +func (l *MetricsRecorderList) RecordInt64Count(handle *estats.Int64CountHandle, incr int64, labels ...string) { + verifyLabels(handle.Descriptor(), labels...) + + for _, metricRecorder := range l.metricsRecorders { + metricRecorder.RecordInt64Count(handle, incr, labels...) + } +} + +func (l *MetricsRecorderList) RecordFloat64Count(handle *estats.Float64CountHandle, incr float64, labels ...string) { + verifyLabels(handle.Descriptor(), labels...) + + for _, metricRecorder := range l.metricsRecorders { + metricRecorder.RecordFloat64Count(handle, incr, labels...) + } +} + +func (l *MetricsRecorderList) RecordInt64Histo(handle *estats.Int64HistoHandle, incr int64, labels ...string) { + verifyLabels(handle.Descriptor(), labels...) + + for _, metricRecorder := range l.metricsRecorders { + metricRecorder.RecordInt64Histo(handle, incr, labels...) + } +} + +func (l *MetricsRecorderList) RecordFloat64Histo(handle *estats.Float64HistoHandle, incr float64, labels ...string) { + verifyLabels(handle.Descriptor(), labels...) + + for _, metricRecorder := range l.metricsRecorders { + metricRecorder.RecordFloat64Histo(handle, incr, labels...) + } +} + +func (l *MetricsRecorderList) RecordInt64Gauge(handle *estats.Int64GaugeHandle, incr int64, labels ...string) { + verifyLabels(handle.Descriptor(), labels...) + + for _, metricRecorder := range l.metricsRecorders { + metricRecorder.RecordInt64Gauge(handle, incr, labels...) + } +} diff --git a/vendor/google.golang.org/grpc/internal/status/status.go b/vendor/google.golang.org/grpc/internal/status/status.go index c7dbc82059525..757925381fe75 100644 --- a/vendor/google.golang.org/grpc/internal/status/status.go +++ b/vendor/google.golang.org/grpc/internal/status/status.go @@ -138,11 +138,11 @@ func (s *Status) WithDetails(details ...protoadapt.MessageV1) (*Status, error) { // s.Code() != OK implies that s.Proto() != nil. p := s.Proto() for _, detail := range details { - any, err := anypb.New(protoadapt.MessageV2Of(detail)) + m, err := anypb.New(protoadapt.MessageV2Of(detail)) if err != nil { return nil, err } - p.Details = append(p.Details, any) + p.Details = append(p.Details, m) } return &Status{s: p}, nil } diff --git a/vendor/google.golang.org/grpc/internal/syscall/syscall_nonlinux.go b/vendor/google.golang.org/grpc/internal/syscall/syscall_nonlinux.go index 999f52cd75bdb..54c24c2ff3865 100644 --- a/vendor/google.golang.org/grpc/internal/syscall/syscall_nonlinux.go +++ b/vendor/google.golang.org/grpc/internal/syscall/syscall_nonlinux.go @@ -58,20 +58,20 @@ func GetRusage() *Rusage { // CPUTimeDiff returns the differences of user CPU time and system CPU time used // between two Rusage structs. It a no-op function for non-linux environments. -func CPUTimeDiff(first *Rusage, latest *Rusage) (float64, float64) { +func CPUTimeDiff(*Rusage, *Rusage) (float64, float64) { log() return 0, 0 } // SetTCPUserTimeout is a no-op function under non-linux environments. -func SetTCPUserTimeout(conn net.Conn, timeout time.Duration) error { +func SetTCPUserTimeout(net.Conn, time.Duration) error { log() return nil } // GetTCPUserTimeout is a no-op function under non-linux environments. // A negative return value indicates the operation is not supported -func GetTCPUserTimeout(conn net.Conn) (int, error) { +func GetTCPUserTimeout(net.Conn) (int, error) { log() return -1, nil } diff --git a/vendor/google.golang.org/grpc/internal/tcp_keepalive_unix.go b/vendor/google.golang.org/grpc/internal/tcp_keepalive_unix.go index 078137b7fd705..7e7aaa5463683 100644 --- a/vendor/google.golang.org/grpc/internal/tcp_keepalive_unix.go +++ b/vendor/google.golang.org/grpc/internal/tcp_keepalive_unix.go @@ -44,7 +44,7 @@ func NetDialerWithTCPKeepalive() *net.Dialer { // combination of unconditionally enabling TCP keepalives here, and // disabling the overriding of TCP keepalive parameters by setting the // KeepAlive field to a negative value above, results in OS defaults for - // the TCP keealive interval and time parameters. + // the TCP keepalive interval and time parameters. Control: func(_, _ string, c syscall.RawConn) error { return c.Control(func(fd uintptr) { unix.SetsockoptInt(int(fd), unix.SOL_SOCKET, unix.SO_KEEPALIVE, 1) diff --git a/vendor/google.golang.org/grpc/internal/tcp_keepalive_windows.go b/vendor/google.golang.org/grpc/internal/tcp_keepalive_windows.go index fd7d43a8907ba..d5c1085eeaecd 100644 --- a/vendor/google.golang.org/grpc/internal/tcp_keepalive_windows.go +++ b/vendor/google.golang.org/grpc/internal/tcp_keepalive_windows.go @@ -44,7 +44,7 @@ func NetDialerWithTCPKeepalive() *net.Dialer { // combination of unconditionally enabling TCP keepalives here, and // disabling the overriding of TCP keepalive parameters by setting the // KeepAlive field to a negative value above, results in OS defaults for - // the TCP keealive interval and time parameters. + // the TCP keepalive interval and time parameters. Control: func(_, _ string, c syscall.RawConn) error { return c.Control(func(fd uintptr) { windows.SetsockoptInt(windows.Handle(fd), windows.SOL_SOCKET, windows.SO_KEEPALIVE, 1) diff --git a/vendor/google.golang.org/grpc/internal/transport/controlbuf.go b/vendor/google.golang.org/grpc/internal/transport/controlbuf.go index 3deadfb4a20c9..ef72fbb3a0163 100644 --- a/vendor/google.golang.org/grpc/internal/transport/controlbuf.go +++ b/vendor/google.golang.org/grpc/internal/transport/controlbuf.go @@ -32,6 +32,7 @@ import ( "golang.org/x/net/http2/hpack" "google.golang.org/grpc/internal/grpclog" "google.golang.org/grpc/internal/grpcutil" + "google.golang.org/grpc/mem" "google.golang.org/grpc/status" ) @@ -148,9 +149,9 @@ type dataFrame struct { streamID uint32 endStream bool h []byte - d []byte + reader mem.Reader // onEachWrite is called every time - // a part of d is written out. + // a part of data is written out. onEachWrite func() } @@ -289,18 +290,22 @@ func (l *outStreamList) dequeue() *outStream { } // controlBuffer is a way to pass information to loopy. -// Information is passed as specific struct types called control frames. -// A control frame not only represents data, messages or headers to be sent out -// but can also be used to instruct loopy to update its internal state. -// It shouldn't be confused with an HTTP2 frame, although some of the control frames -// like dataFrame and headerFrame do go out on wire as HTTP2 frames. +// +// Information is passed as specific struct types called control frames. A +// control frame not only represents data, messages or headers to be sent out +// but can also be used to instruct loopy to update its internal state. It +// shouldn't be confused with an HTTP2 frame, although some of the control +// frames like dataFrame and headerFrame do go out on wire as HTTP2 frames. type controlBuffer struct { - ch chan struct{} - done <-chan struct{} + wakeupCh chan struct{} // Unblocks readers waiting for something to read. + done <-chan struct{} // Closed when the transport is done. + + // Mutex guards all the fields below, except trfChan which can be read + // atomically without holding mu. mu sync.Mutex - consumerWaiting bool - list *itemList - err error + consumerWaiting bool // True when readers are blocked waiting for new data. + closed bool // True when the controlbuf is finished. + list *itemList // List of queued control frames. // transportResponseFrames counts the number of queued items that represent // the response of an action initiated by the peer. trfChan is created @@ -308,47 +313,59 @@ type controlBuffer struct { // closed and nilled when transportResponseFrames drops below the // threshold. Both fields are protected by mu. transportResponseFrames int - trfChan atomic.Value // chan struct{} + trfChan atomic.Pointer[chan struct{}] } func newControlBuffer(done <-chan struct{}) *controlBuffer { return &controlBuffer{ - ch: make(chan struct{}, 1), - list: &itemList{}, - done: done, + wakeupCh: make(chan struct{}, 1), + list: &itemList{}, + done: done, } } -// throttle blocks if there are too many incomingSettings/cleanupStreams in the -// controlbuf. +// throttle blocks if there are too many frames in the control buf that +// represent the response of an action initiated by the peer, like +// incomingSettings cleanupStreams etc. func (c *controlBuffer) throttle() { - ch, _ := c.trfChan.Load().(chan struct{}) - if ch != nil { + if ch := c.trfChan.Load(); ch != nil { select { - case <-ch: + case <-(*ch): case <-c.done: } } } +// put adds an item to the controlbuf. func (c *controlBuffer) put(it cbItem) error { _, err := c.executeAndPut(nil, it) return err } +// executeAndPut runs f, and if the return value is true, adds the given item to +// the controlbuf. The item could be nil, in which case, this method simply +// executes f and does not add the item to the controlbuf. +// +// The first return value indicates whether the item was successfully added to +// the control buffer. A non-nil error, specifically ErrConnClosing, is returned +// if the control buffer is already closed. func (c *controlBuffer) executeAndPut(f func() bool, it cbItem) (bool, error) { - var wakeUp bool c.mu.Lock() - if c.err != nil { - c.mu.Unlock() - return false, c.err + defer c.mu.Unlock() + + if c.closed { + return false, ErrConnClosing } if f != nil { if !f() { // f wasn't successful - c.mu.Unlock() return false, nil } } + if it == nil { + return true, nil + } + + var wakeUp bool if c.consumerWaiting { wakeUp = true c.consumerWaiting = false @@ -359,98 +376,102 @@ func (c *controlBuffer) executeAndPut(f func() bool, it cbItem) (bool, error) { if c.transportResponseFrames == maxQueuedTransportResponseFrames { // We are adding the frame that puts us over the threshold; create // a throttling channel. - c.trfChan.Store(make(chan struct{})) + ch := make(chan struct{}) + c.trfChan.Store(&ch) } } - c.mu.Unlock() if wakeUp { select { - case c.ch <- struct{}{}: + case c.wakeupCh <- struct{}{}: default: } } return true, nil } -// Note argument f should never be nil. -func (c *controlBuffer) execute(f func(it any) bool, it any) (bool, error) { - c.mu.Lock() - if c.err != nil { - c.mu.Unlock() - return false, c.err - } - if !f(it) { // f wasn't successful - c.mu.Unlock() - return false, nil - } - c.mu.Unlock() - return true, nil -} - +// get returns the next control frame from the control buffer. If block is true +// **and** there are no control frames in the control buffer, the call blocks +// until one of the conditions is met: there is a frame to return or the +// transport is closed. func (c *controlBuffer) get(block bool) (any, error) { for { c.mu.Lock() - if c.err != nil { + frame, err := c.getOnceLocked() + if frame != nil || err != nil || !block { + // If we read a frame or an error, we can return to the caller. The + // call to getOnceLocked() returns a nil frame and a nil error if + // there is nothing to read, and in that case, if the caller asked + // us not to block, we can return now as well. c.mu.Unlock() - return nil, c.err - } - if !c.list.isEmpty() { - h := c.list.dequeue().(cbItem) - if h.isTransportResponseFrame() { - if c.transportResponseFrames == maxQueuedTransportResponseFrames { - // We are removing the frame that put us over the - // threshold; close and clear the throttling channel. - ch := c.trfChan.Load().(chan struct{}) - close(ch) - c.trfChan.Store((chan struct{})(nil)) - } - c.transportResponseFrames-- - } - c.mu.Unlock() - return h, nil - } - if !block { - c.mu.Unlock() - return nil, nil + return frame, err } c.consumerWaiting = true c.mu.Unlock() + + // Release the lock above and wait to be woken up. select { - case <-c.ch: + case <-c.wakeupCh: case <-c.done: return nil, errors.New("transport closed by client") } } } +// Callers must not use this method, but should instead use get(). +// +// Caller must hold c.mu. +func (c *controlBuffer) getOnceLocked() (any, error) { + if c.closed { + return false, ErrConnClosing + } + if c.list.isEmpty() { + return nil, nil + } + h := c.list.dequeue().(cbItem) + if h.isTransportResponseFrame() { + if c.transportResponseFrames == maxQueuedTransportResponseFrames { + // We are removing the frame that put us over the + // threshold; close and clear the throttling channel. + ch := c.trfChan.Swap(nil) + close(*ch) + } + c.transportResponseFrames-- + } + return h, nil +} + +// finish closes the control buffer, cleaning up any streams that have queued +// header frames. Once this method returns, no more frames can be added to the +// control buffer, and attempts to do so will return ErrConnClosing. func (c *controlBuffer) finish() { c.mu.Lock() - if c.err != nil { - c.mu.Unlock() + defer c.mu.Unlock() + + if c.closed { return } - c.err = ErrConnClosing + c.closed = true // There may be headers for streams in the control buffer. // These streams need to be cleaned out since the transport // is still not aware of these yet. for head := c.list.dequeueAll(); head != nil; head = head.next { - hdr, ok := head.it.(*headerFrame) - if !ok { - continue - } - if hdr.onOrphaned != nil { // It will be nil on the server-side. - hdr.onOrphaned(ErrConnClosing) + switch v := head.it.(type) { + case *headerFrame: + if v.onOrphaned != nil { // It will be nil on the server-side. + v.onOrphaned(ErrConnClosing) + } + case *dataFrame: + _ = v.reader.Close() } } + // In case throttle() is currently in flight, it needs to be unblocked. // Otherwise, the transport may not close, since the transport is closed by // the reader encountering the connection error. - ch, _ := c.trfChan.Load().(chan struct{}) + ch := c.trfChan.Swap(nil) if ch != nil { - close(ch) + close(*ch) } - c.trfChan.Store((chan struct{})(nil)) - c.mu.Unlock() } type side int @@ -466,7 +487,7 @@ const ( // stream maintains a queue of data frames; as loopy receives data frames // it gets added to the queue of the relevant stream. // Loopy goes over this list of active streams by processing one node every iteration, -// thereby closely resemebling to a round-robin scheduling over all streams. While +// thereby closely resembling a round-robin scheduling over all streams. While // processing a stream, loopy writes out data bytes from this stream capped by the min // of http2MaxFrameLen, connection-level flow control and stream-level flow control. type loopyWriter struct { @@ -490,12 +511,13 @@ type loopyWriter struct { draining bool conn net.Conn logger *grpclog.PrefixLogger + bufferPool mem.BufferPool // Side-specific handlers ssGoAwayHandler func(*goAway) (bool, error) } -func newLoopyWriter(s side, fr *framer, cbuf *controlBuffer, bdpEst *bdpEstimator, conn net.Conn, logger *grpclog.PrefixLogger, goAwayHandler func(*goAway) (bool, error)) *loopyWriter { +func newLoopyWriter(s side, fr *framer, cbuf *controlBuffer, bdpEst *bdpEstimator, conn net.Conn, logger *grpclog.PrefixLogger, goAwayHandler func(*goAway) (bool, error), bufferPool mem.BufferPool) *loopyWriter { var buf bytes.Buffer l := &loopyWriter{ side: s, @@ -511,6 +533,7 @@ func newLoopyWriter(s side, fr *framer, cbuf *controlBuffer, bdpEst *bdpEstimato conn: conn, logger: logger, ssGoAwayHandler: goAwayHandler, + bufferPool: bufferPool, } return l } @@ -768,6 +791,11 @@ func (l *loopyWriter) cleanupStreamHandler(c *cleanupStream) error { // not be established yet. delete(l.estdStreams, c.streamID) str.deleteSelf() + for head := str.itl.dequeueAll(); head != nil; head = head.next { + if df, ok := head.it.(*dataFrame); ok { + _ = df.reader.Close() + } + } } if c.rst { // If RST_STREAM needs to be sent. if err := l.framer.fr.WriteRSTStream(c.streamID, c.rstCode); err != nil { @@ -903,16 +931,18 @@ func (l *loopyWriter) processData() (bool, error) { dataItem := str.itl.peek().(*dataFrame) // Peek at the first data item this stream. // A data item is represented by a dataFrame, since it later translates into // multiple HTTP2 data frames. - // Every dataFrame has two buffers; h that keeps grpc-message header and d that is actual data. - // As an optimization to keep wire traffic low, data from d is copied to h to make as big as the - // maximum possible HTTP2 frame size. + // Every dataFrame has two buffers; h that keeps grpc-message header and data + // that is the actual message. As an optimization to keep wire traffic low, data + // from data is copied to h to make as big as the maximum possible HTTP2 frame + // size. - if len(dataItem.h) == 0 && len(dataItem.d) == 0 { // Empty data frame + if len(dataItem.h) == 0 && dataItem.reader.Remaining() == 0 { // Empty data frame // Client sends out empty data frame with endStream = true if err := l.framer.fr.WriteData(dataItem.streamID, dataItem.endStream, nil); err != nil { return false, err } str.itl.dequeue() // remove the empty data item from stream + _ = dataItem.reader.Close() if str.itl.isEmpty() { str.state = empty } else if trailer, ok := str.itl.peek().(*headerFrame); ok { // the next item is trailers. @@ -927,9 +957,7 @@ func (l *loopyWriter) processData() (bool, error) { } return false, nil } - var ( - buf []byte - ) + // Figure out the maximum size we can send maxSize := http2MaxFrameLen if strQuota := int(l.oiws) - str.bytesOutStanding; strQuota <= 0 { // stream-level flow control. @@ -943,43 +971,50 @@ func (l *loopyWriter) processData() (bool, error) { } // Compute how much of the header and data we can send within quota and max frame length hSize := min(maxSize, len(dataItem.h)) - dSize := min(maxSize-hSize, len(dataItem.d)) - if hSize != 0 { - if dSize == 0 { - buf = dataItem.h - } else { - // We can add some data to grpc message header to distribute bytes more equally across frames. - // Copy on the stack to avoid generating garbage - var localBuf [http2MaxFrameLen]byte - copy(localBuf[:hSize], dataItem.h) - copy(localBuf[hSize:], dataItem.d[:dSize]) - buf = localBuf[:hSize+dSize] - } + dSize := min(maxSize-hSize, dataItem.reader.Remaining()) + remainingBytes := len(dataItem.h) + dataItem.reader.Remaining() - hSize - dSize + size := hSize + dSize + + var buf *[]byte + + if hSize != 0 && dSize == 0 { + buf = &dataItem.h } else { - buf = dataItem.d - } + // Note: this is only necessary because the http2.Framer does not support + // partially writing a frame, so the sequence must be materialized into a buffer. + // TODO: Revisit once https://github.com/golang/go/issues/66655 is addressed. + pool := l.bufferPool + if pool == nil { + // Note that this is only supposed to be nil in tests. Otherwise, stream is + // always initialized with a BufferPool. + pool = mem.DefaultBufferPool() + } + buf = pool.Get(size) + defer pool.Put(buf) - size := hSize + dSize + copy((*buf)[:hSize], dataItem.h) + _, _ = dataItem.reader.Read((*buf)[hSize:]) + } // Now that outgoing flow controls are checked we can replenish str's write quota str.wq.replenish(size) var endStream bool // If this is the last data message on this stream and all of it can be written in this iteration. - if dataItem.endStream && len(dataItem.h)+len(dataItem.d) <= size { + if dataItem.endStream && remainingBytes == 0 { endStream = true } if dataItem.onEachWrite != nil { dataItem.onEachWrite() } - if err := l.framer.fr.WriteData(dataItem.streamID, endStream, buf[:size]); err != nil { + if err := l.framer.fr.WriteData(dataItem.streamID, endStream, (*buf)[:size]); err != nil { return false, err } str.bytesOutStanding += size l.sendQuota -= uint32(size) dataItem.h = dataItem.h[hSize:] - dataItem.d = dataItem.d[dSize:] - if len(dataItem.h) == 0 && len(dataItem.d) == 0 { // All the data from that message was written out. + if remainingBytes == 0 { // All the data from that message was written out. + _ = dataItem.reader.Close() str.itl.dequeue() } if str.itl.isEmpty() { @@ -998,10 +1033,3 @@ func (l *loopyWriter) processData() (bool, error) { } return false, nil } - -func min(a, b int) int { - if a < b { - return a - } - return b -} diff --git a/vendor/google.golang.org/grpc/internal/transport/handler_server.go b/vendor/google.golang.org/grpc/internal/transport/handler_server.go index 4a3ddce29a4e7..ce878693bd741 100644 --- a/vendor/google.golang.org/grpc/internal/transport/handler_server.go +++ b/vendor/google.golang.org/grpc/internal/transport/handler_server.go @@ -24,7 +24,6 @@ package transport import ( - "bytes" "context" "errors" "fmt" @@ -40,6 +39,7 @@ import ( "google.golang.org/grpc/credentials" "google.golang.org/grpc/internal/grpclog" "google.golang.org/grpc/internal/grpcutil" + "google.golang.org/grpc/mem" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" "google.golang.org/grpc/stats" @@ -50,7 +50,7 @@ import ( // NewServerHandlerTransport returns a ServerTransport handling gRPC from // inside an http.Handler, or writes an HTTP error to w and returns an error. // It requires that the http Server supports HTTP/2. -func NewServerHandlerTransport(w http.ResponseWriter, r *http.Request, stats []stats.Handler) (ServerTransport, error) { +func NewServerHandlerTransport(w http.ResponseWriter, r *http.Request, stats []stats.Handler, bufferPool mem.BufferPool) (ServerTransport, error) { if r.Method != http.MethodPost { w.Header().Set("Allow", http.MethodPost) msg := fmt.Sprintf("invalid gRPC request method %q", r.Method) @@ -98,6 +98,7 @@ func NewServerHandlerTransport(w http.ResponseWriter, r *http.Request, stats []s contentType: contentType, contentSubtype: contentSubtype, stats: stats, + bufferPool: bufferPool, } st.logger = prefixLoggerForServerHandlerTransport(st) @@ -171,6 +172,8 @@ type serverHandlerTransport struct { stats []stats.Handler logger *grpclog.PrefixLogger + + bufferPool mem.BufferPool } func (ht *serverHandlerTransport) Close(err error) { @@ -244,6 +247,7 @@ func (ht *serverHandlerTransport) WriteStatus(s *Stream, st *status.Status) erro } s.hdrMu.Lock() + defer s.hdrMu.Unlock() if p := st.Proto(); p != nil && len(p.Details) > 0 { delete(s.trailer, grpcStatusDetailsBinHeader) stBytes, err := proto.Marshal(p) @@ -268,7 +272,6 @@ func (ht *serverHandlerTransport) WriteStatus(s *Stream, st *status.Status) erro } } } - s.hdrMu.Unlock() }) if err == nil { // transport has not been closed @@ -330,16 +333,28 @@ func (ht *serverHandlerTransport) writeCustomHeaders(s *Stream) { s.hdrMu.Unlock() } -func (ht *serverHandlerTransport) Write(s *Stream, hdr []byte, data []byte, opts *Options) error { +func (ht *serverHandlerTransport) Write(s *Stream, hdr []byte, data mem.BufferSlice, _ *Options) error { + // Always take a reference because otherwise there is no guarantee the data will + // be available after this function returns. This is what callers to Write + // expect. + data.Ref() headersWritten := s.updateHeaderSent() - return ht.do(func() { + err := ht.do(func() { + defer data.Free() if !headersWritten { ht.writePendingHeaders(s) } ht.rw.Write(hdr) - ht.rw.Write(data) + for _, b := range data { + _, _ = ht.rw.Write(b.ReadOnlyData()) + } ht.rw.(http.Flusher).Flush() }) + if err != nil { + data.Free() + return err + } + return nil } func (ht *serverHandlerTransport) WriteHeader(s *Stream, md metadata.MD) error { @@ -406,7 +421,7 @@ func (ht *serverHandlerTransport) HandleStreams(ctx context.Context, startStream headerWireLength: 0, // won't have access to header wire length until golang/go#18997. } s.trReader = &transportReader{ - reader: &recvBufferReader{ctx: s.ctx, ctxDone: s.ctx.Done(), recv: s.buf, freeBuffer: func(*bytes.Buffer) {}}, + reader: &recvBufferReader{ctx: s.ctx, ctxDone: s.ctx.Done(), recv: s.buf}, windowHandler: func(int) {}, } @@ -415,21 +430,19 @@ func (ht *serverHandlerTransport) HandleStreams(ctx context.Context, startStream go func() { defer close(readerDone) - // TODO: minimize garbage, optimize recvBuffer code/ownership - const readSize = 8196 - for buf := make([]byte, readSize); ; { - n, err := req.Body.Read(buf) + for { + buf := ht.bufferPool.Get(http2MaxFrameLen) + n, err := req.Body.Read(*buf) if n > 0 { - s.buf.put(recvMsg{buffer: bytes.NewBuffer(buf[:n:n])}) - buf = buf[n:] + *buf = (*buf)[:n] + s.buf.put(recvMsg{buffer: mem.NewBuffer(buf, ht.bufferPool)}) + } else { + ht.bufferPool.Put(buf) } if err != nil { s.buf.put(recvMsg{err: mapRecvMsgError(err)}) return } - if len(buf) == 0 { - buf = make([]byte, readSize) - } } }() @@ -462,7 +475,7 @@ func (ht *serverHandlerTransport) IncrMsgSent() {} func (ht *serverHandlerTransport) IncrMsgRecv() {} -func (ht *serverHandlerTransport) Drain(debugData string) { +func (ht *serverHandlerTransport) Drain(string) { panic("Drain() is not implemented") } diff --git a/vendor/google.golang.org/grpc/internal/transport/http2_client.go b/vendor/google.golang.org/grpc/internal/transport/http2_client.go index 3c63c706986da..c769deab53c77 100644 --- a/vendor/google.golang.org/grpc/internal/transport/http2_client.go +++ b/vendor/google.golang.org/grpc/internal/transport/http2_client.go @@ -47,6 +47,7 @@ import ( isyscall "google.golang.org/grpc/internal/syscall" "google.golang.org/grpc/internal/transport/networktype" "google.golang.org/grpc/keepalive" + "google.golang.org/grpc/mem" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" "google.golang.org/grpc/resolver" @@ -59,6 +60,8 @@ import ( // atomically. var clientConnectionCounter uint64 +var goAwayLoopyWriterTimeout = 5 * time.Second + var metadataFromOutgoingContextRaw = internal.FromOutgoingContextRaw.(func(context.Context) (metadata.MD, [][]string, bool)) // http2Client implements the ClientTransport interface with HTTP2. @@ -144,7 +147,7 @@ type http2Client struct { onClose func(GoAwayReason) - bufferPool *bufferPool + bufferPool mem.BufferPool connectionID uint64 logger *grpclog.PrefixLogger @@ -229,7 +232,7 @@ func newHTTP2Client(connectCtx, ctx context.Context, addr resolver.Address, opts } }(conn) - // The following defer and goroutine monitor the connectCtx for cancelation + // The following defer and goroutine monitor the connectCtx for cancellation // and deadline. On context expiration, the connection is hard closed and // this function will naturally fail as a result. Otherwise, the defer // waits for the goroutine to exit to prevent the context from being @@ -346,7 +349,7 @@ func newHTTP2Client(connectCtx, ctx context.Context, addr resolver.Address, opts streamQuota: defaultMaxStreamsClient, streamsQuotaAvailable: make(chan struct{}, 1), keepaliveEnabled: keepaliveEnabled, - bufferPool: newBufferPool(), + bufferPool: opts.BufferPool, onClose: onClose, } var czSecurity credentials.ChannelzSecurityValue @@ -463,7 +466,7 @@ func newHTTP2Client(connectCtx, ctx context.Context, addr resolver.Address, opts return nil, err } go func() { - t.loopy = newLoopyWriter(clientSide, t.framer, t.controlBuf, t.bdpEst, t.conn, t.logger, t.outgoingGoAwayHandler) + t.loopy = newLoopyWriter(clientSide, t.framer, t.controlBuf, t.bdpEst, t.conn, t.logger, t.outgoingGoAwayHandler, t.bufferPool) if err := t.loopy.run(); !isIOError(err) { // Immediately close the connection, as the loopy writer returns // when there are no more active streams and we were draining (the @@ -504,7 +507,6 @@ func (t *http2Client) newStream(ctx context.Context, callHdr *CallHdr) *Stream { closeStream: func(err error) { t.CloseStream(s, err) }, - freeBuffer: t.bufferPool.put, }, windowHandler: func(n int) { t.updateWindow(s, uint32(n)) @@ -770,7 +772,7 @@ func (t *http2Client) NewStream(ctx context.Context, callHdr *CallHdr) (*Stream, hdr := &headerFrame{ hf: headerFields, endStream: false, - initStream: func(id uint32) error { + initStream: func(uint32) error { t.mu.Lock() // TODO: handle transport closure in loopy instead and remove this // initStream is never called when transport is draining. @@ -983,6 +985,7 @@ func (t *http2Client) closeStream(s *Stream, err error, rst bool, rstCode http2. // only once on a transport. Once it is called, the transport should not be // accessed anymore. func (t *http2Client) Close(err error) { + t.conn.SetWriteDeadline(time.Now().Add(time.Second * 10)) t.mu.Lock() // Make sure we only close once. if t.state == closing { @@ -1006,10 +1009,20 @@ func (t *http2Client) Close(err error) { t.kpDormancyCond.Signal() } t.mu.Unlock() + // Per HTTP/2 spec, a GOAWAY frame must be sent before closing the - // connection. See https://httpwg.org/specs/rfc7540.html#GOAWAY. + // connection. See https://httpwg.org/specs/rfc7540.html#GOAWAY. It + // also waits for loopyWriter to be closed with a timer to avoid the + // long blocking in case the connection is blackholed, i.e. TCP is + // just stuck. t.controlBuf.put(&goAway{code: http2.ErrCodeNo, debugData: []byte("client transport shutdown"), closeConn: err}) - <-t.writerDone + timer := time.NewTimer(goAwayLoopyWriterTimeout) + defer timer.Stop() + select { + case <-t.writerDone: // success + case <-timer.C: + t.logger.Infof("Failed to write a GOAWAY frame as part of connection close after %s. Giving up and closing the transport.", goAwayLoopyWriterTimeout) + } t.cancel() t.conn.Close() channelz.RemoveEntry(t.channelz.ID) @@ -1065,27 +1078,36 @@ func (t *http2Client) GracefulClose() { // Write formats the data into HTTP2 data frame(s) and sends it out. The caller // should proceed only if Write returns nil. -func (t *http2Client) Write(s *Stream, hdr []byte, data []byte, opts *Options) error { +func (t *http2Client) Write(s *Stream, hdr []byte, data mem.BufferSlice, opts *Options) error { + reader := data.Reader() + if opts.Last { // If it's the last message, update stream state. if !s.compareAndSwapState(streamActive, streamWriteDone) { + _ = reader.Close() return errStreamDone } } else if s.getState() != streamActive { + _ = reader.Close() return errStreamDone } df := &dataFrame{ streamID: s.id, endStream: opts.Last, h: hdr, - d: data, + reader: reader, } - if hdr != nil || data != nil { // If it's not an empty data frame, check quota. - if err := s.wq.get(int32(len(hdr) + len(data))); err != nil { + if hdr != nil || df.reader.Remaining() != 0 { // If it's not an empty data frame, check quota. + if err := s.wq.get(int32(len(hdr) + df.reader.Remaining())); err != nil { + _ = reader.Close() return err } } - return t.controlBuf.put(df) + if err := t.controlBuf.put(df); err != nil { + _ = reader.Close() + return err + } + return nil } func (t *http2Client) getStream(f http2.Frame) *Stream { @@ -1190,10 +1212,13 @@ func (t *http2Client) handleData(f *http2.DataFrame) { // guarantee f.Data() is consumed before the arrival of next frame. // Can this copy be eliminated? if len(f.Data()) > 0 { - buffer := t.bufferPool.get() - buffer.Reset() - buffer.Write(f.Data()) - s.write(recvMsg{buffer: buffer}) + pool := t.bufferPool + if pool == nil { + // Note that this is only supposed to be nil in tests. Otherwise, stream is + // always initialized with a BufferPool. + pool = mem.DefaultBufferPool() + } + s.write(recvMsg{buffer: mem.Copy(f.Data(), pool)}) } } // The server has closed the stream without sending trailers. Record that @@ -1222,7 +1247,7 @@ func (t *http2Client) handleRSTStream(f *http2.RSTStreamFrame) { if statusCode == codes.Canceled { if d, ok := s.ctx.Deadline(); ok && !d.After(time.Now()) { // Our deadline was already exceeded, and that was likely the cause - // of this cancelation. Alter the status code accordingly. + // of this cancellation. Alter the status code accordingly. statusCode = codes.DeadlineExceeded } } @@ -1307,7 +1332,7 @@ func (t *http2Client) handleGoAway(f *http2.GoAwayFrame) { id := f.LastStreamID if id > 0 && id%2 == 0 { t.mu.Unlock() - t.Close(connectionErrorf(true, nil, "received goaway with non-zero even-numbered numbered stream id: %v", id)) + t.Close(connectionErrorf(true, nil, "received goaway with non-zero even-numbered stream id: %v", id)) return } // A client can receive multiple GoAways from the server (see @@ -1642,11 +1667,10 @@ func (t *http2Client) reader(errCh chan<- error) { t.closeStream(s, status.Error(code, msg), true, http2.ErrCodeProtocol, status.New(code, msg), nil, false) } continue - } else { - // Transport error. - t.Close(connectionErrorf(true, err, "error reading from server: %v", err)) - return } + // Transport error. + t.Close(connectionErrorf(true, err, "error reading from server: %v", err)) + return } switch frame := frame.(type) { case *http2.MetaHeadersFrame: @@ -1671,13 +1695,6 @@ func (t *http2Client) reader(errCh chan<- error) { } } -func minTime(a, b time.Duration) time.Duration { - if a < b { - return a - } - return b -} - // keepalive running in a separate goroutine makes sure the connection is alive by sending pings. func (t *http2Client) keepalive() { p := &ping{data: [8]byte{}} @@ -1745,7 +1762,7 @@ func (t *http2Client) keepalive() { // timeoutLeft. This will ensure that we wait only for kp.Time // before sending out the next ping (for cases where the ping is // acked). - sleepDuration := minTime(t.kp.Time, timeoutLeft) + sleepDuration := min(t.kp.Time, timeoutLeft) timeoutLeft -= sleepDuration timer.Reset(sleepDuration) case <-t.ctx.Done(): diff --git a/vendor/google.golang.org/grpc/internal/transport/http2_server.go b/vendor/google.golang.org/grpc/internal/transport/http2_server.go index b7091165b5013..584b50fe55302 100644 --- a/vendor/google.golang.org/grpc/internal/transport/http2_server.go +++ b/vendor/google.golang.org/grpc/internal/transport/http2_server.go @@ -39,6 +39,7 @@ import ( "google.golang.org/grpc/internal/grpcutil" "google.golang.org/grpc/internal/pretty" "google.golang.org/grpc/internal/syscall" + "google.golang.org/grpc/mem" "google.golang.org/protobuf/proto" "google.golang.org/grpc/codes" @@ -119,7 +120,7 @@ type http2Server struct { // Fields below are for channelz metric collection. channelz *channelz.Socket - bufferPool *bufferPool + bufferPool mem.BufferPool connectionID uint64 @@ -261,7 +262,7 @@ func NewServerTransport(conn net.Conn, config *ServerConfig) (_ ServerTransport, idle: time.Now(), kep: kep, initialWindowSize: iwz, - bufferPool: newBufferPool(), + bufferPool: config.BufferPool, } var czSecurity credentials.ChannelzSecurityValue if au, ok := authInfo.(credentials.ChannelzSecurityInfo); ok { @@ -330,7 +331,7 @@ func NewServerTransport(conn net.Conn, config *ServerConfig) (_ ServerTransport, t.handleSettings(sf) go func() { - t.loopy = newLoopyWriter(serverSide, t.framer, t.controlBuf, t.bdpEst, t.conn, t.logger, t.outgoingGoAwayHandler) + t.loopy = newLoopyWriter(serverSide, t.framer, t.controlBuf, t.bdpEst, t.conn, t.logger, t.outgoingGoAwayHandler, t.bufferPool) err := t.loopy.run() close(t.loopyWriterDone) if !isIOError(err) { @@ -613,10 +614,9 @@ func (t *http2Server) operateHeaders(ctx context.Context, frame *http2.MetaHeade s.wq = newWriteQuota(defaultWriteQuota, s.ctxDone) s.trReader = &transportReader{ reader: &recvBufferReader{ - ctx: s.ctx, - ctxDone: s.ctxDone, - recv: s.buf, - freeBuffer: t.bufferPool.put, + ctx: s.ctx, + ctxDone: s.ctxDone, + recv: s.buf, }, windowHandler: func(n int) { t.updateWindow(s, uint32(n)) @@ -813,10 +813,13 @@ func (t *http2Server) handleData(f *http2.DataFrame) { // guarantee f.Data() is consumed before the arrival of next frame. // Can this copy be eliminated? if len(f.Data()) > 0 { - buffer := t.bufferPool.get() - buffer.Reset() - buffer.Write(f.Data()) - s.write(recvMsg{buffer: buffer}) + pool := t.bufferPool + if pool == nil { + // Note that this is only supposed to be nil in tests. Otherwise, stream is + // always initialized with a BufferPool. + pool = mem.DefaultBufferPool() + } + s.write(recvMsg{buffer: mem.Copy(f.Data(), pool)}) } } if f.StreamEnded() { @@ -1089,7 +1092,9 @@ func (t *http2Server) WriteStatus(s *Stream, st *status.Status) error { onWrite: t.setResetPingStrikes, } - success, err := t.controlBuf.execute(t.checkForHeaderListSize, trailingHeader) + success, err := t.controlBuf.executeAndPut(func() bool { + return t.checkForHeaderListSize(trailingHeader) + }, nil) if !success { if err != nil { return err @@ -1112,27 +1117,37 @@ func (t *http2Server) WriteStatus(s *Stream, st *status.Status) error { // Write converts the data into HTTP2 data frame and sends it out. Non-nil error // is returns if it fails (e.g., framing error, transport error). -func (t *http2Server) Write(s *Stream, hdr []byte, data []byte, opts *Options) error { +func (t *http2Server) Write(s *Stream, hdr []byte, data mem.BufferSlice, _ *Options) error { + reader := data.Reader() + if !s.isHeaderSent() { // Headers haven't been written yet. if err := t.WriteHeader(s, nil); err != nil { + _ = reader.Close() return err } } else { // Writing headers checks for this condition. if s.getState() == streamDone { + _ = reader.Close() return t.streamContextErr(s) } } + df := &dataFrame{ streamID: s.id, h: hdr, - d: data, + reader: reader, onEachWrite: t.setResetPingStrikes, } - if err := s.wq.get(int32(len(hdr) + len(data))); err != nil { + if err := s.wq.get(int32(len(hdr) + df.reader.Remaining())); err != nil { + _ = reader.Close() return t.streamContextErr(s) } - return t.controlBuf.put(df) + if err := t.controlBuf.put(df); err != nil { + _ = reader.Close() + return err + } + return nil } // keepalive running in a separate goroutine does the following: @@ -1223,7 +1238,7 @@ func (t *http2Server) keepalive() { // timeoutLeft. This will ensure that we wait only for kp.Time // before sending out the next ping (for cases where the ping is // acked). - sleepDuration := minTime(t.kp.Time, kpTimeoutLeft) + sleepDuration := min(t.kp.Time, kpTimeoutLeft) kpTimeoutLeft -= sleepDuration kpTimer.Reset(sleepDuration) case <-t.done: diff --git a/vendor/google.golang.org/grpc/internal/transport/http_util.go b/vendor/google.golang.org/grpc/internal/transport/http_util.go index 39cef3bd442eb..3613d7b64817d 100644 --- a/vendor/google.golang.org/grpc/internal/transport/http_util.go +++ b/vendor/google.golang.org/grpc/internal/transport/http_util.go @@ -317,28 +317,32 @@ func newBufWriter(conn net.Conn, batchSize int, pool *sync.Pool) *bufWriter { return w } -func (w *bufWriter) Write(b []byte) (n int, err error) { +func (w *bufWriter) Write(b []byte) (int, error) { if w.err != nil { return 0, w.err } if w.batchSize == 0 { // Buffer has been disabled. - n, err = w.conn.Write(b) + n, err := w.conn.Write(b) return n, toIOError(err) } if w.buf == nil { b := w.pool.Get().(*[]byte) w.buf = *b } + written := 0 for len(b) > 0 { - nn := copy(w.buf[w.offset:], b) - b = b[nn:] - w.offset += nn - n += nn - if w.offset >= w.batchSize { - err = w.flushKeepBuffer() + copied := copy(w.buf[w.offset:], b) + b = b[copied:] + written += copied + w.offset += copied + if w.offset < w.batchSize { + continue + } + if err := w.flushKeepBuffer(); err != nil { + return written, err } } - return n, err + return written, nil } func (w *bufWriter) Flush() error { @@ -389,7 +393,7 @@ type framer struct { fr *http2.Framer } -var writeBufferPoolMap map[int]*sync.Pool = make(map[int]*sync.Pool) +var writeBufferPoolMap = make(map[int]*sync.Pool) var writeBufferMutex sync.Mutex func newFramer(conn net.Conn, writeBufferSize, readBufferSize int, sharedWriteBuffer bool, maxHeaderListSize uint32) *framer { diff --git a/vendor/google.golang.org/grpc/internal/transport/proxy.go b/vendor/google.golang.org/grpc/internal/transport/proxy.go index 24fa1032574cb..54b2244365444 100644 --- a/vendor/google.golang.org/grpc/internal/transport/proxy.go +++ b/vendor/google.golang.org/grpc/internal/transport/proxy.go @@ -107,8 +107,14 @@ func doHTTPConnectHandshake(ctx context.Context, conn net.Conn, backendAddr stri } return nil, fmt.Errorf("failed to do connect handshake, response: %q", dump) } - - return &bufConn{Conn: conn, r: r}, nil + // The buffer could contain extra bytes from the target server, so we can't + // discard it. However, in many cases where the server waits for the client + // to send the first message (e.g. when TLS is being used), the buffer will + // be empty, so we can avoid the overhead of reading through this buffer. + if r.Buffered() != 0 { + return &bufConn{Conn: conn, r: r}, nil + } + return conn, nil } // proxyDial dials, connecting to a proxy first if necessary. Checks if a proxy diff --git a/vendor/google.golang.org/grpc/internal/transport/transport.go b/vendor/google.golang.org/grpc/internal/transport/transport.go index 4b39c0ade97c0..fdd6fa86cc15c 100644 --- a/vendor/google.golang.org/grpc/internal/transport/transport.go +++ b/vendor/google.golang.org/grpc/internal/transport/transport.go @@ -22,7 +22,6 @@ package transport import ( - "bytes" "context" "errors" "fmt" @@ -37,6 +36,7 @@ import ( "google.golang.org/grpc/credentials" "google.golang.org/grpc/internal/channelz" "google.golang.org/grpc/keepalive" + "google.golang.org/grpc/mem" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" "google.golang.org/grpc/resolver" @@ -47,32 +47,10 @@ import ( const logLevel = 2 -type bufferPool struct { - pool sync.Pool -} - -func newBufferPool() *bufferPool { - return &bufferPool{ - pool: sync.Pool{ - New: func() any { - return new(bytes.Buffer) - }, - }, - } -} - -func (p *bufferPool) get() *bytes.Buffer { - return p.pool.Get().(*bytes.Buffer) -} - -func (p *bufferPool) put(b *bytes.Buffer) { - p.pool.Put(b) -} - // recvMsg represents the received msg from the transport. All transport // protocol specific info has been removed. type recvMsg struct { - buffer *bytes.Buffer + buffer mem.Buffer // nil: received some data // io.EOF: stream is completed. data is nil. // other non-nil error: transport failure. data is nil. @@ -102,6 +80,9 @@ func newRecvBuffer() *recvBuffer { func (b *recvBuffer) put(r recvMsg) { b.mu.Lock() if b.err != nil { + // drop the buffer on the floor. Since b.err is not nil, any subsequent reads + // will always return an error, making this buffer inaccessible. + r.buffer.Free() b.mu.Unlock() // An error had occurred earlier, don't accept more // data or errors. @@ -148,45 +129,97 @@ type recvBufferReader struct { ctx context.Context ctxDone <-chan struct{} // cache of ctx.Done() (for performance). recv *recvBuffer - last *bytes.Buffer // Stores the remaining data in the previous calls. + last mem.Buffer // Stores the remaining data in the previous calls. err error - freeBuffer func(*bytes.Buffer) } -// Read reads the next len(p) bytes from last. If last is drained, it tries to -// read additional data from recv. It blocks if there no additional data available -// in recv. If Read returns any non-nil error, it will continue to return that error. -func (r *recvBufferReader) Read(p []byte) (n int, err error) { +func (r *recvBufferReader) ReadHeader(header []byte) (n int, err error) { if r.err != nil { return 0, r.err } if r.last != nil { - // Read remaining data left in last call. - copied, _ := r.last.Read(p) - if r.last.Len() == 0 { - r.freeBuffer(r.last) + n, r.last = mem.ReadUnsafe(header, r.last) + return n, nil + } + if r.closeStream != nil { + n, r.err = r.readHeaderClient(header) + } else { + n, r.err = r.readHeader(header) + } + return n, r.err +} + +// Read reads the next n bytes from last. If last is drained, it tries to read +// additional data from recv. It blocks if there no additional data available in +// recv. If Read returns any non-nil error, it will continue to return that +// error. +func (r *recvBufferReader) Read(n int) (buf mem.Buffer, err error) { + if r.err != nil { + return nil, r.err + } + if r.last != nil { + buf = r.last + if r.last.Len() > n { + buf, r.last = mem.SplitUnsafe(buf, n) + } else { r.last = nil } - return copied, nil + return buf, nil } if r.closeStream != nil { - n, r.err = r.readClient(p) + buf, r.err = r.readClient(n) } else { - n, r.err = r.read(p) + buf, r.err = r.read(n) } - return n, r.err + return buf, r.err } -func (r *recvBufferReader) read(p []byte) (n int, err error) { +func (r *recvBufferReader) readHeader(header []byte) (n int, err error) { select { case <-r.ctxDone: return 0, ContextErr(r.ctx.Err()) case m := <-r.recv.get(): - return r.readAdditional(m, p) + return r.readHeaderAdditional(m, header) + } +} + +func (r *recvBufferReader) read(n int) (buf mem.Buffer, err error) { + select { + case <-r.ctxDone: + return nil, ContextErr(r.ctx.Err()) + case m := <-r.recv.get(): + return r.readAdditional(m, n) + } +} + +func (r *recvBufferReader) readHeaderClient(header []byte) (n int, err error) { + // If the context is canceled, then closes the stream with nil metadata. + // closeStream writes its error parameter to r.recv as a recvMsg. + // r.readAdditional acts on that message and returns the necessary error. + select { + case <-r.ctxDone: + // Note that this adds the ctx error to the end of recv buffer, and + // reads from the head. This will delay the error until recv buffer is + // empty, thus will delay ctx cancellation in Recv(). + // + // It's done this way to fix a race between ctx cancel and trailer. The + // race was, stream.Recv() may return ctx error if ctxDone wins the + // race, but stream.Trailer() may return a non-nil md because the stream + // was not marked as done when trailer is received. This closeStream + // call will mark stream as done, thus fix the race. + // + // TODO: delaying ctx error seems like a unnecessary side effect. What + // we really want is to mark the stream as done, and return ctx error + // faster. + r.closeStream(ContextErr(r.ctx.Err())) + m := <-r.recv.get() + return r.readHeaderAdditional(m, header) + case m := <-r.recv.get(): + return r.readHeaderAdditional(m, header) } } -func (r *recvBufferReader) readClient(p []byte) (n int, err error) { +func (r *recvBufferReader) readClient(n int) (buf mem.Buffer, err error) { // If the context is canceled, then closes the stream with nil metadata. // closeStream writes its error parameter to r.recv as a recvMsg. // r.readAdditional acts on that message and returns the necessary error. @@ -207,25 +240,40 @@ func (r *recvBufferReader) readClient(p []byte) (n int, err error) { // faster. r.closeStream(ContextErr(r.ctx.Err())) m := <-r.recv.get() - return r.readAdditional(m, p) + return r.readAdditional(m, n) case m := <-r.recv.get(): - return r.readAdditional(m, p) + return r.readAdditional(m, n) } } -func (r *recvBufferReader) readAdditional(m recvMsg, p []byte) (n int, err error) { +func (r *recvBufferReader) readHeaderAdditional(m recvMsg, header []byte) (n int, err error) { r.recv.load() if m.err != nil { + if m.buffer != nil { + m.buffer.Free() + } return 0, m.err } - copied, _ := m.buffer.Read(p) - if m.buffer.Len() == 0 { - r.freeBuffer(m.buffer) - r.last = nil - } else { - r.last = m.buffer + + n, r.last = mem.ReadUnsafe(header, m.buffer) + + return n, nil +} + +func (r *recvBufferReader) readAdditional(m recvMsg, n int) (b mem.Buffer, err error) { + r.recv.load() + if m.err != nil { + if m.buffer != nil { + m.buffer.Free() + } + return nil, m.err + } + + if m.buffer.Len() > n { + m.buffer, r.last = mem.SplitUnsafe(m.buffer, n) } - return copied, nil + + return m.buffer, nil } type streamState uint32 @@ -241,7 +289,7 @@ const ( type Stream struct { id uint32 st ServerTransport // nil for client side Stream - ct *http2Client // nil for server side Stream + ct ClientTransport // nil for server side Stream ctx context.Context // the associated context of the stream cancel context.CancelFunc // always nil for client side Stream done chan struct{} // closed at the end of stream to unblock writers. On the client side. @@ -251,7 +299,7 @@ type Stream struct { recvCompress string sendCompress string buf *recvBuffer - trReader io.Reader + trReader *transportReader fc *inFlow wq *writeQuota @@ -408,7 +456,7 @@ func (s *Stream) TrailersOnly() bool { return s.noHeaders } -// Trailer returns the cached trailer metedata. Note that if it is not called +// Trailer returns the cached trailer metadata. Note that if it is not called // after the entire stream is done, it could return an empty MD. Client // side only. // It can be safely read only after stream has ended that is either read @@ -499,36 +547,87 @@ func (s *Stream) write(m recvMsg) { s.buf.put(m) } -// Read reads all p bytes from the wire for this stream. -func (s *Stream) Read(p []byte) (n int, err error) { +func (s *Stream) ReadHeader(header []byte) (err error) { + // Don't request a read if there was an error earlier + if er := s.trReader.er; er != nil { + return er + } + s.requestRead(len(header)) + for len(header) != 0 { + n, err := s.trReader.ReadHeader(header) + header = header[n:] + if len(header) == 0 { + err = nil + } + if err != nil { + if n > 0 && err == io.EOF { + err = io.ErrUnexpectedEOF + } + return err + } + } + return nil +} + +// Read reads n bytes from the wire for this stream. +func (s *Stream) Read(n int) (data mem.BufferSlice, err error) { // Don't request a read if there was an error earlier - if er := s.trReader.(*transportReader).er; er != nil { - return 0, er + if er := s.trReader.er; er != nil { + return nil, er } - s.requestRead(len(p)) - return io.ReadFull(s.trReader, p) + s.requestRead(n) + for n != 0 { + buf, err := s.trReader.Read(n) + var bufLen int + if buf != nil { + bufLen = buf.Len() + } + n -= bufLen + if n == 0 { + err = nil + } + if err != nil { + if bufLen > 0 && err == io.EOF { + err = io.ErrUnexpectedEOF + } + data.Free() + return nil, err + } + data = append(data, buf) + } + return data, nil } -// tranportReader reads all the data available for this Stream from the transport and +// transportReader reads all the data available for this Stream from the transport and // passes them into the decoder, which converts them into a gRPC message stream. // The error is io.EOF when the stream is done or another non-nil error if // the stream broke. type transportReader struct { - reader io.Reader + reader *recvBufferReader // The handler to control the window update procedure for both this // particular stream and the associated transport. windowHandler func(int) er error } -func (t *transportReader) Read(p []byte) (n int, err error) { - n, err = t.reader.Read(p) +func (t *transportReader) ReadHeader(header []byte) (int, error) { + n, err := t.reader.ReadHeader(header) if err != nil { t.er = err - return + return 0, err + } + t.windowHandler(len(header)) + return n, nil +} + +func (t *transportReader) Read(n int) (mem.Buffer, error) { + buf, err := t.reader.Read(n) + if err != nil { + t.er = err + return buf, err } - t.windowHandler(n) - return + t.windowHandler(buf.Len()) + return buf, nil } // BytesReceived indicates whether any bytes have been received on this stream. @@ -574,6 +673,7 @@ type ServerConfig struct { ChannelzParent *channelz.Server MaxHeaderListSize *uint32 HeaderTableSize *uint32 + BufferPool mem.BufferPool } // ConnectOptions covers all relevant options for communicating with the server. @@ -612,6 +712,8 @@ type ConnectOptions struct { MaxHeaderListSize *uint32 // UseProxy specifies if a proxy should be used. UseProxy bool + // The mem.BufferPool to use when reading/writing to the wire. + BufferPool mem.BufferPool } // NewClientTransport establishes the transport with the required ConnectOptions @@ -673,7 +775,7 @@ type ClientTransport interface { // Write sends the data for the given stream. A nil stream indicates // the write is to be performed on the transport as a whole. - Write(s *Stream, hdr []byte, data []byte, opts *Options) error + Write(s *Stream, hdr []byte, data mem.BufferSlice, opts *Options) error // NewStream creates a Stream for an RPC. NewStream(ctx context.Context, callHdr *CallHdr) (*Stream, error) @@ -725,7 +827,7 @@ type ServerTransport interface { // Write sends the data for the given stream. // Write may not be called on all streams. - Write(s *Stream, hdr []byte, data []byte, opts *Options) error + Write(s *Stream, hdr []byte, data mem.BufferSlice, opts *Options) error // WriteStatus sends the status of a stream to the client. WriteStatus is // the final call made on a stream and always occurs. @@ -798,7 +900,7 @@ var ( // connection is draining. This could be caused by goaway or balancer // removing the address. errStreamDrain = status.Error(codes.Unavailable, "the connection is draining") - // errStreamDone is returned from write at the client side to indiacte application + // errStreamDone is returned from write at the client side to indicate application // layer of an error. errStreamDone = errors.New("the stream is done") // StatusGoAway indicates that the server sent a GOAWAY that included this diff --git a/vendor/google.golang.org/grpc/keepalive/keepalive.go b/vendor/google.golang.org/grpc/keepalive/keepalive.go index 34d31b5e7d311..eb42b19fb99a1 100644 --- a/vendor/google.golang.org/grpc/keepalive/keepalive.go +++ b/vendor/google.golang.org/grpc/keepalive/keepalive.go @@ -34,15 +34,29 @@ type ClientParameters struct { // After a duration of this time if the client doesn't see any activity it // pings the server to see if the transport is still alive. // If set below 10s, a minimum value of 10s will be used instead. - Time time.Duration // The current default value is infinity. + // + // Note that gRPC servers have a default EnforcementPolicy.MinTime of 5 + // minutes (which means the client shouldn't ping more frequently than every + // 5 minutes). + // + // Though not ideal, it's not a strong requirement for Time to be less than + // EnforcementPolicy.MinTime. Time will automatically double if the server + // disconnects due to its enforcement policy. + // + // For more details, see + // https://github.com/grpc/proposal/blob/master/A8-client-side-keepalive.md + Time time.Duration // After having pinged for keepalive check, the client waits for a duration // of Timeout and if no activity is seen even after that the connection is // closed. - Timeout time.Duration // The current default value is 20 seconds. + // + // If keepalive is enabled, and this value is not explicitly set, the default + // is 20 seconds. + Timeout time.Duration // If true, client sends keepalive pings even with no active RPCs. If false, // when there are no active RPCs, Time and Timeout will be ignored and no // keepalive pings will be sent. - PermitWithoutStream bool // false by default. + PermitWithoutStream bool } // ServerParameters is used to set keepalive and max-age parameters on the diff --git a/vendor/google.golang.org/grpc/mem/buffer_pool.go b/vendor/google.golang.org/grpc/mem/buffer_pool.go new file mode 100644 index 0000000000000..c37c58c0233ec --- /dev/null +++ b/vendor/google.golang.org/grpc/mem/buffer_pool.go @@ -0,0 +1,194 @@ +/* + * + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package mem + +import ( + "sort" + "sync" + + "google.golang.org/grpc/internal" +) + +// BufferPool is a pool of buffers that can be shared and reused, resulting in +// decreased memory allocation. +type BufferPool interface { + // Get returns a buffer with specified length from the pool. + Get(length int) *[]byte + + // Put returns a buffer to the pool. + Put(*[]byte) +} + +var defaultBufferPoolSizes = []int{ + 256, + 4 << 10, // 4KB (go page size) + 16 << 10, // 16KB (max HTTP/2 frame size used by gRPC) + 32 << 10, // 32KB (default buffer size for io.Copy) + 1 << 20, // 1MB +} + +var defaultBufferPool BufferPool + +func init() { + defaultBufferPool = NewTieredBufferPool(defaultBufferPoolSizes...) + + internal.SetDefaultBufferPoolForTesting = func(pool BufferPool) { + defaultBufferPool = pool + } + + internal.SetBufferPoolingThresholdForTesting = func(threshold int) { + bufferPoolingThreshold = threshold + } +} + +// DefaultBufferPool returns the current default buffer pool. It is a BufferPool +// created with NewBufferPool that uses a set of default sizes optimized for +// expected workflows. +func DefaultBufferPool() BufferPool { + return defaultBufferPool +} + +// NewTieredBufferPool returns a BufferPool implementation that uses multiple +// underlying pools of the given pool sizes. +func NewTieredBufferPool(poolSizes ...int) BufferPool { + sort.Ints(poolSizes) + pools := make([]*sizedBufferPool, len(poolSizes)) + for i, s := range poolSizes { + pools[i] = newSizedBufferPool(s) + } + return &tieredBufferPool{ + sizedPools: pools, + } +} + +// tieredBufferPool implements the BufferPool interface with multiple tiers of +// buffer pools for different sizes of buffers. +type tieredBufferPool struct { + sizedPools []*sizedBufferPool + fallbackPool simpleBufferPool +} + +func (p *tieredBufferPool) Get(size int) *[]byte { + return p.getPool(size).Get(size) +} + +func (p *tieredBufferPool) Put(buf *[]byte) { + p.getPool(cap(*buf)).Put(buf) +} + +func (p *tieredBufferPool) getPool(size int) BufferPool { + poolIdx := sort.Search(len(p.sizedPools), func(i int) bool { + return p.sizedPools[i].defaultSize >= size + }) + + if poolIdx == len(p.sizedPools) { + return &p.fallbackPool + } + + return p.sizedPools[poolIdx] +} + +// sizedBufferPool is a BufferPool implementation that is optimized for specific +// buffer sizes. For example, HTTP/2 frames within gRPC have a default max size +// of 16kb and a sizedBufferPool can be configured to only return buffers with a +// capacity of 16kb. Note that however it does not support returning larger +// buffers and in fact panics if such a buffer is requested. Because of this, +// this BufferPool implementation is not meant to be used on its own and rather +// is intended to be embedded in a tieredBufferPool such that Get is only +// invoked when the required size is smaller than or equal to defaultSize. +type sizedBufferPool struct { + pool sync.Pool + defaultSize int +} + +func (p *sizedBufferPool) Get(size int) *[]byte { + buf := p.pool.Get().(*[]byte) + b := *buf + clear(b[:cap(b)]) + *buf = b[:size] + return buf +} + +func (p *sizedBufferPool) Put(buf *[]byte) { + if cap(*buf) < p.defaultSize { + // Ignore buffers that are too small to fit in the pool. Otherwise, when + // Get is called it will panic as it tries to index outside the bounds + // of the buffer. + return + } + p.pool.Put(buf) +} + +func newSizedBufferPool(size int) *sizedBufferPool { + return &sizedBufferPool{ + pool: sync.Pool{ + New: func() any { + buf := make([]byte, size) + return &buf + }, + }, + defaultSize: size, + } +} + +var _ BufferPool = (*simpleBufferPool)(nil) + +// simpleBufferPool is an implementation of the BufferPool interface that +// attempts to pool buffers with a sync.Pool. When Get is invoked, it tries to +// acquire a buffer from the pool but if that buffer is too small, it returns it +// to the pool and creates a new one. +type simpleBufferPool struct { + pool sync.Pool +} + +func (p *simpleBufferPool) Get(size int) *[]byte { + bs, ok := p.pool.Get().(*[]byte) + if ok && cap(*bs) >= size { + *bs = (*bs)[:size] + return bs + } + + // A buffer was pulled from the pool, but it is too small. Put it back in + // the pool and create one large enough. + if ok { + p.pool.Put(bs) + } + + b := make([]byte, size) + return &b +} + +func (p *simpleBufferPool) Put(buf *[]byte) { + p.pool.Put(buf) +} + +var _ BufferPool = NopBufferPool{} + +// NopBufferPool is a buffer pool that returns new buffers without pooling. +type NopBufferPool struct{} + +// Get returns a buffer with specified length from the pool. +func (NopBufferPool) Get(length int) *[]byte { + b := make([]byte, length) + return &b +} + +// Put returns a buffer to the pool. +func (NopBufferPool) Put(*[]byte) { +} diff --git a/vendor/google.golang.org/grpc/mem/buffer_slice.go b/vendor/google.golang.org/grpc/mem/buffer_slice.go new file mode 100644 index 0000000000000..228e9c2f20f26 --- /dev/null +++ b/vendor/google.golang.org/grpc/mem/buffer_slice.go @@ -0,0 +1,226 @@ +/* + * + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package mem + +import ( + "io" +) + +// BufferSlice offers a means to represent data that spans one or more Buffer +// instances. A BufferSlice is meant to be immutable after creation, and methods +// like Ref create and return copies of the slice. This is why all methods have +// value receivers rather than pointer receivers. +// +// Note that any of the methods that read the underlying buffers such as Ref, +// Len or CopyTo etc., will panic if any underlying buffers have already been +// freed. It is recommended to not directly interact with any of the underlying +// buffers directly, rather such interactions should be mediated through the +// various methods on this type. +// +// By convention, any APIs that return (mem.BufferSlice, error) should reduce +// the burden on the caller by never returning a mem.BufferSlice that needs to +// be freed if the error is non-nil, unless explicitly stated. +type BufferSlice []Buffer + +// Len returns the sum of the length of all the Buffers in this slice. +// +// # Warning +// +// Invoking the built-in len on a BufferSlice will return the number of buffers +// in the slice, and *not* the value returned by this function. +func (s BufferSlice) Len() int { + var length int + for _, b := range s { + length += b.Len() + } + return length +} + +// Ref invokes Ref on each buffer in the slice. +func (s BufferSlice) Ref() { + for _, b := range s { + b.Ref() + } +} + +// Free invokes Buffer.Free() on each Buffer in the slice. +func (s BufferSlice) Free() { + for _, b := range s { + b.Free() + } +} + +// CopyTo copies each of the underlying Buffer's data into the given buffer, +// returning the number of bytes copied. Has the same semantics as the copy +// builtin in that it will copy as many bytes as it can, stopping when either dst +// is full or s runs out of data, returning the minimum of s.Len() and len(dst). +func (s BufferSlice) CopyTo(dst []byte) int { + off := 0 + for _, b := range s { + off += copy(dst[off:], b.ReadOnlyData()) + } + return off +} + +// Materialize concatenates all the underlying Buffer's data into a single +// contiguous buffer using CopyTo. +func (s BufferSlice) Materialize() []byte { + l := s.Len() + if l == 0 { + return nil + } + out := make([]byte, l) + s.CopyTo(out) + return out +} + +// MaterializeToBuffer functions like Materialize except that it writes the data +// to a single Buffer pulled from the given BufferPool. +// +// As a special case, if the input BufferSlice only actually has one Buffer, this +// function simply increases the refcount before returning said Buffer. Freeing this +// buffer won't release it until the BufferSlice is itself released. +func (s BufferSlice) MaterializeToBuffer(pool BufferPool) Buffer { + if len(s) == 1 { + s[0].Ref() + return s[0] + } + sLen := s.Len() + if sLen == 0 { + return emptyBuffer{} + } + buf := pool.Get(sLen) + s.CopyTo(*buf) + return NewBuffer(buf, pool) +} + +// Reader returns a new Reader for the input slice after taking references to +// each underlying buffer. +func (s BufferSlice) Reader() Reader { + s.Ref() + return &sliceReader{ + data: s, + len: s.Len(), + } +} + +// Reader exposes a BufferSlice's data as an io.Reader, allowing it to interface +// with other parts systems. It also provides an additional convenience method +// Remaining(), which returns the number of unread bytes remaining in the slice. +// Buffers will be freed as they are read. +type Reader interface { + io.Reader + io.ByteReader + // Close frees the underlying BufferSlice and never returns an error. Subsequent + // calls to Read will return (0, io.EOF). + Close() error + // Remaining returns the number of unread bytes remaining in the slice. + Remaining() int +} + +type sliceReader struct { + data BufferSlice + len int + // The index into data[0].ReadOnlyData(). + bufferIdx int +} + +func (r *sliceReader) Remaining() int { + return r.len +} + +func (r *sliceReader) Close() error { + r.data.Free() + r.data = nil + r.len = 0 + return nil +} + +func (r *sliceReader) freeFirstBufferIfEmpty() bool { + if len(r.data) == 0 || r.bufferIdx != len(r.data[0].ReadOnlyData()) { + return false + } + + r.data[0].Free() + r.data = r.data[1:] + r.bufferIdx = 0 + return true +} + +func (r *sliceReader) Read(buf []byte) (n int, _ error) { + if r.len == 0 { + return 0, io.EOF + } + + for len(buf) != 0 && r.len != 0 { + // Copy as much as possible from the first Buffer in the slice into the + // given byte slice. + data := r.data[0].ReadOnlyData() + copied := copy(buf, data[r.bufferIdx:]) + r.len -= copied // Reduce len by the number of bytes copied. + r.bufferIdx += copied // Increment the buffer index. + n += copied // Increment the total number of bytes read. + buf = buf[copied:] // Shrink the given byte slice. + + // If we have copied all the data from the first Buffer, free it and advance to + // the next in the slice. + r.freeFirstBufferIfEmpty() + } + + return n, nil +} + +func (r *sliceReader) ReadByte() (byte, error) { + if r.len == 0 { + return 0, io.EOF + } + + // There may be any number of empty buffers in the slice, clear them all until a + // non-empty buffer is reached. This is guaranteed to exit since r.len is not 0. + for r.freeFirstBufferIfEmpty() { + } + + b := r.data[0].ReadOnlyData()[r.bufferIdx] + r.len-- + r.bufferIdx++ + // Free the first buffer in the slice if the last byte was read + r.freeFirstBufferIfEmpty() + return b, nil +} + +var _ io.Writer = (*writer)(nil) + +type writer struct { + buffers *BufferSlice + pool BufferPool +} + +func (w *writer) Write(p []byte) (n int, err error) { + b := Copy(p, w.pool) + *w.buffers = append(*w.buffers, b) + return b.Len(), nil +} + +// NewWriter wraps the given BufferSlice and BufferPool to implement the +// io.Writer interface. Every call to Write copies the contents of the given +// buffer into a new Buffer pulled from the given pool and the Buffer is added to +// the given BufferSlice. +func NewWriter(buffers *BufferSlice, pool BufferPool) io.Writer { + return &writer{buffers: buffers, pool: pool} +} diff --git a/vendor/google.golang.org/grpc/mem/buffers.go b/vendor/google.golang.org/grpc/mem/buffers.go new file mode 100644 index 0000000000000..4d66b2ccc2be6 --- /dev/null +++ b/vendor/google.golang.org/grpc/mem/buffers.go @@ -0,0 +1,252 @@ +/* + * + * Copyright 2024 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +// Package mem provides utilities that facilitate memory reuse in byte slices +// that are used as buffers. +// +// # Experimental +// +// Notice: All APIs in this package are EXPERIMENTAL and may be changed or +// removed in a later release. +package mem + +import ( + "fmt" + "sync" + "sync/atomic" +) + +// A Buffer represents a reference counted piece of data (in bytes) that can be +// acquired by a call to NewBuffer() or Copy(). A reference to a Buffer may be +// released by calling Free(), which invokes the free function given at creation +// only after all references are released. +// +// Note that a Buffer is not safe for concurrent access and instead each +// goroutine should use its own reference to the data, which can be acquired via +// a call to Ref(). +// +// Attempts to access the underlying data after releasing the reference to the +// Buffer will panic. +type Buffer interface { + // ReadOnlyData returns the underlying byte slice. Note that it is undefined + // behavior to modify the contents of this slice in any way. + ReadOnlyData() []byte + // Ref increases the reference counter for this Buffer. + Ref() + // Free decrements this Buffer's reference counter and frees the underlying + // byte slice if the counter reaches 0 as a result of this call. + Free() + // Len returns the Buffer's size. + Len() int + + split(n int) (left, right Buffer) + read(buf []byte) (int, Buffer) +} + +var ( + bufferPoolingThreshold = 1 << 10 + + bufferObjectPool = sync.Pool{New: func() any { return new(buffer) }} + refObjectPool = sync.Pool{New: func() any { return new(atomic.Int32) }} +) + +func IsBelowBufferPoolingThreshold(size int) bool { + return size <= bufferPoolingThreshold +} + +type buffer struct { + origData *[]byte + data []byte + refs *atomic.Int32 + pool BufferPool +} + +func newBuffer() *buffer { + return bufferObjectPool.Get().(*buffer) +} + +// NewBuffer creates a new Buffer from the given data, initializing the reference +// counter to 1. The data will then be returned to the given pool when all +// references to the returned Buffer are released. As a special case to avoid +// additional allocations, if the given buffer pool is nil, the returned buffer +// will be a "no-op" Buffer where invoking Buffer.Free() does nothing and the +// underlying data is never freed. +// +// Note that the backing array of the given data is not copied. +func NewBuffer(data *[]byte, pool BufferPool) Buffer { + if pool == nil || IsBelowBufferPoolingThreshold(len(*data)) { + return (SliceBuffer)(*data) + } + b := newBuffer() + b.origData = data + b.data = *data + b.pool = pool + b.refs = refObjectPool.Get().(*atomic.Int32) + b.refs.Add(1) + return b +} + +// Copy creates a new Buffer from the given data, initializing the reference +// counter to 1. +// +// It acquires a []byte from the given pool and copies over the backing array +// of the given data. The []byte acquired from the pool is returned to the +// pool when all references to the returned Buffer are released. +func Copy(data []byte, pool BufferPool) Buffer { + if IsBelowBufferPoolingThreshold(len(data)) { + buf := make(SliceBuffer, len(data)) + copy(buf, data) + return buf + } + + buf := pool.Get(len(data)) + copy(*buf, data) + return NewBuffer(buf, pool) +} + +func (b *buffer) ReadOnlyData() []byte { + if b.refs == nil { + panic("Cannot read freed buffer") + } + return b.data +} + +func (b *buffer) Ref() { + if b.refs == nil { + panic("Cannot ref freed buffer") + } + b.refs.Add(1) +} + +func (b *buffer) Free() { + if b.refs == nil { + panic("Cannot free freed buffer") + } + + refs := b.refs.Add(-1) + switch { + case refs > 0: + return + case refs == 0: + if b.pool != nil { + b.pool.Put(b.origData) + } + + refObjectPool.Put(b.refs) + b.origData = nil + b.data = nil + b.refs = nil + b.pool = nil + bufferObjectPool.Put(b) + default: + panic("Cannot free freed buffer") + } +} + +func (b *buffer) Len() int { + return len(b.ReadOnlyData()) +} + +func (b *buffer) split(n int) (Buffer, Buffer) { + if b.refs == nil { + panic("Cannot split freed buffer") + } + + b.refs.Add(1) + split := newBuffer() + split.origData = b.origData + split.data = b.data[n:] + split.refs = b.refs + split.pool = b.pool + + b.data = b.data[:n] + + return b, split +} + +func (b *buffer) read(buf []byte) (int, Buffer) { + if b.refs == nil { + panic("Cannot read freed buffer") + } + + n := copy(buf, b.data) + if n == len(b.data) { + b.Free() + return n, nil + } + + b.data = b.data[n:] + return n, b +} + +// String returns a string representation of the buffer. May be used for +// debugging purposes. +func (b *buffer) String() string { + return fmt.Sprintf("mem.Buffer(%p, data: %p, length: %d)", b, b.ReadOnlyData(), len(b.ReadOnlyData())) +} + +func ReadUnsafe(dst []byte, buf Buffer) (int, Buffer) { + return buf.read(dst) +} + +// SplitUnsafe modifies the receiver to point to the first n bytes while it +// returns a new reference to the remaining bytes. The returned Buffer functions +// just like a normal reference acquired using Ref(). +func SplitUnsafe(buf Buffer, n int) (left, right Buffer) { + return buf.split(n) +} + +type emptyBuffer struct{} + +func (e emptyBuffer) ReadOnlyData() []byte { + return nil +} + +func (e emptyBuffer) Ref() {} +func (e emptyBuffer) Free() {} + +func (e emptyBuffer) Len() int { + return 0 +} + +func (e emptyBuffer) split(int) (left, right Buffer) { + return e, e +} + +func (e emptyBuffer) read([]byte) (int, Buffer) { + return 0, e +} + +type SliceBuffer []byte + +func (s SliceBuffer) ReadOnlyData() []byte { return s } +func (s SliceBuffer) Ref() {} +func (s SliceBuffer) Free() {} +func (s SliceBuffer) Len() int { return len(s) } + +func (s SliceBuffer) split(n int) (left, right Buffer) { + return s[:n], s[n:] +} + +func (s SliceBuffer) read(buf []byte) (int, Buffer) { + n := copy(buf, s) + if n == len(s) { + return n, nil + } + return n, s[n:] +} diff --git a/vendor/google.golang.org/grpc/metadata/metadata.go b/vendor/google.golang.org/grpc/metadata/metadata.go index 1e9485fd6e268..d2e15253bbfbc 100644 --- a/vendor/google.golang.org/grpc/metadata/metadata.go +++ b/vendor/google.golang.org/grpc/metadata/metadata.go @@ -213,11 +213,6 @@ func FromIncomingContext(ctx context.Context) (MD, bool) { // ValueFromIncomingContext returns the metadata value corresponding to the metadata // key from the incoming metadata if it exists. Keys are matched in a case insensitive // manner. -// -// # Experimental -// -// Notice: This API is EXPERIMENTAL and may be changed or removed in a -// later release. func ValueFromIncomingContext(ctx context.Context, key string) []string { md, ok := ctx.Value(mdIncomingKey{}).(MD) if !ok { @@ -228,7 +223,7 @@ func ValueFromIncomingContext(ctx context.Context, key string) []string { return copyOf(v) } for k, v := range md { - // Case insenitive comparison: MD is a map, and there's no guarantee + // Case insensitive comparison: MD is a map, and there's no guarantee // that the MD attached to the context is created using our helper // functions. if strings.EqualFold(k, key) { diff --git a/vendor/google.golang.org/grpc/preloader.go b/vendor/google.golang.org/grpc/preloader.go index 73bd63364335e..e87a17f36a50b 100644 --- a/vendor/google.golang.org/grpc/preloader.go +++ b/vendor/google.golang.org/grpc/preloader.go @@ -20,6 +20,7 @@ package grpc import ( "google.golang.org/grpc/codes" + "google.golang.org/grpc/mem" "google.golang.org/grpc/status" ) @@ -31,9 +32,10 @@ import ( // later release. type PreparedMsg struct { // Struct for preparing msg before sending them - encodedData []byte + encodedData mem.BufferSlice hdr []byte - payload []byte + payload mem.BufferSlice + pf payloadFormat } // Encode marshalls and compresses the message using the codec and compressor for the stream. @@ -57,11 +59,27 @@ func (p *PreparedMsg) Encode(s Stream, msg any) error { if err != nil { return err } - p.encodedData = data - compData, err := compress(data, rpcInfo.preloaderInfo.cp, rpcInfo.preloaderInfo.comp) + + materializedData := data.Materialize() + data.Free() + p.encodedData = mem.BufferSlice{mem.NewBuffer(&materializedData, nil)} + + // TODO: it should be possible to grab the bufferPool from the underlying + // stream implementation with a type cast to its actual type (such as + // addrConnStream) and accessing the buffer pool directly. + var compData mem.BufferSlice + compData, p.pf, err = compress(p.encodedData, rpcInfo.preloaderInfo.cp, rpcInfo.preloaderInfo.comp, mem.DefaultBufferPool()) if err != nil { return err } - p.hdr, p.payload = msgHeader(data, compData) + + if p.pf.isCompressed() { + materializedCompData := compData.Materialize() + compData.Free() + compData = mem.BufferSlice{mem.NewBuffer(&materializedCompData, nil)} + } + + p.hdr, p.payload = msgHeader(p.encodedData, compData, p.pf) + return nil } diff --git a/vendor/google.golang.org/grpc/regenerate.sh b/vendor/google.golang.org/grpc/regenerate.sh deleted file mode 100644 index 3edca296c224c..0000000000000 --- a/vendor/google.golang.org/grpc/regenerate.sh +++ /dev/null @@ -1,123 +0,0 @@ -#!/bin/bash -# Copyright 2020 gRPC authors. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -set -eu -o pipefail - -WORKDIR=$(mktemp -d) - -function finish { - rm -rf "$WORKDIR" -} -trap finish EXIT - -export GOBIN=${WORKDIR}/bin -export PATH=${GOBIN}:${PATH} -mkdir -p ${GOBIN} - -echo "remove existing generated files" -# grpc_testing_not_regenerate/*.pb.go is not re-generated, -# see grpc_testing_not_regenerate/README.md for details. -rm -f $(find . -name '*.pb.go' | grep -v 'grpc_testing_not_regenerate') - -echo "go install google.golang.org/protobuf/cmd/protoc-gen-go" -(cd test/tools && go install google.golang.org/protobuf/cmd/protoc-gen-go) - -echo "go install cmd/protoc-gen-go-grpc" -(cd cmd/protoc-gen-go-grpc && go install .) - -echo "git clone https://github.com/grpc/grpc-proto" -git clone --quiet https://github.com/grpc/grpc-proto ${WORKDIR}/grpc-proto - -echo "git clone https://github.com/protocolbuffers/protobuf" -git clone --quiet https://github.com/protocolbuffers/protobuf ${WORKDIR}/protobuf - -# Pull in code.proto as a proto dependency -mkdir -p ${WORKDIR}/googleapis/google/rpc -echo "curl https://raw.githubusercontent.com/googleapis/googleapis/master/google/rpc/code.proto" -curl --silent https://raw.githubusercontent.com/googleapis/googleapis/master/google/rpc/code.proto > ${WORKDIR}/googleapis/google/rpc/code.proto - -mkdir -p ${WORKDIR}/out - -# Generates sources without the embed requirement -LEGACY_SOURCES=( - ${WORKDIR}/grpc-proto/grpc/binlog/v1/binarylog.proto - ${WORKDIR}/grpc-proto/grpc/channelz/v1/channelz.proto - ${WORKDIR}/grpc-proto/grpc/health/v1/health.proto - ${WORKDIR}/grpc-proto/grpc/lb/v1/load_balancer.proto - profiling/proto/service.proto - ${WORKDIR}/grpc-proto/grpc/reflection/v1alpha/reflection.proto - ${WORKDIR}/grpc-proto/grpc/reflection/v1/reflection.proto -) - -# Generates only the new gRPC Service symbols -SOURCES=( - $(git ls-files --exclude-standard --cached --others "*.proto" | grep -v '^profiling/proto/service.proto$') - ${WORKDIR}/grpc-proto/grpc/gcp/altscontext.proto - ${WORKDIR}/grpc-proto/grpc/gcp/handshaker.proto - ${WORKDIR}/grpc-proto/grpc/gcp/transport_security_common.proto - ${WORKDIR}/grpc-proto/grpc/lookup/v1/rls.proto - ${WORKDIR}/grpc-proto/grpc/lookup/v1/rls_config.proto - ${WORKDIR}/grpc-proto/grpc/testing/*.proto - ${WORKDIR}/grpc-proto/grpc/core/*.proto -) - -# These options of the form 'Mfoo.proto=bar' instruct the codegen to use an -# import path of 'bar' in the generated code when 'foo.proto' is imported in -# one of the sources. -# -# Note that the protos listed here are all for testing purposes. All protos to -# be used externally should have a go_package option (and they don't need to be -# listed here). -OPTS=Mgrpc/core/stats.proto=google.golang.org/grpc/interop/grpc_testing/core,\ -Mgrpc/testing/benchmark_service.proto=google.golang.org/grpc/interop/grpc_testing,\ -Mgrpc/testing/stats.proto=google.golang.org/grpc/interop/grpc_testing,\ -Mgrpc/testing/report_qps_scenario_service.proto=google.golang.org/grpc/interop/grpc_testing,\ -Mgrpc/testing/messages.proto=google.golang.org/grpc/interop/grpc_testing,\ -Mgrpc/testing/worker_service.proto=google.golang.org/grpc/interop/grpc_testing,\ -Mgrpc/testing/control.proto=google.golang.org/grpc/interop/grpc_testing,\ -Mgrpc/testing/test.proto=google.golang.org/grpc/interop/grpc_testing,\ -Mgrpc/testing/payloads.proto=google.golang.org/grpc/interop/grpc_testing,\ -Mgrpc/testing/empty.proto=google.golang.org/grpc/interop/grpc_testing - -for src in ${SOURCES[@]}; do - echo "protoc ${src}" - protoc --go_out=${OPTS}:${WORKDIR}/out --go-grpc_out=${OPTS},use_generic_streams_experimental=true:${WORKDIR}/out \ - -I"." \ - -I${WORKDIR}/grpc-proto \ - -I${WORKDIR}/googleapis \ - -I${WORKDIR}/protobuf/src \ - ${src} -done - -for src in ${LEGACY_SOURCES[@]}; do - echo "protoc ${src}" - protoc --go_out=${OPTS}:${WORKDIR}/out --go-grpc_out=${OPTS},require_unimplemented_servers=false:${WORKDIR}/out \ - -I"." \ - -I${WORKDIR}/grpc-proto \ - -I${WORKDIR}/googleapis \ - -I${WORKDIR}/protobuf/src \ - ${src} -done - -# The go_package option in grpc/lookup/v1/rls.proto doesn't match the -# current location. Move it into the right place. -mkdir -p ${WORKDIR}/out/google.golang.org/grpc/internal/proto/grpc_lookup_v1 -mv ${WORKDIR}/out/google.golang.org/grpc/lookup/grpc_lookup_v1/* ${WORKDIR}/out/google.golang.org/grpc/internal/proto/grpc_lookup_v1 - -# grpc_testing_not_regenerate/*.pb.go are not re-generated, -# see grpc_testing_not_regenerate/README.md for details. -rm ${WORKDIR}/out/google.golang.org/grpc/reflection/test/grpc_testing_not_regenerate/*.pb.go - -cp -R ${WORKDIR}/out/google.golang.org/grpc/* . diff --git a/vendor/google.golang.org/grpc/resolver/manual/manual.go b/vendor/google.golang.org/grpc/resolver/manual/manual.go index f2efa2a2cb5a9..09e864a89d35b 100644 --- a/vendor/google.golang.org/grpc/resolver/manual/manual.go +++ b/vendor/google.golang.org/grpc/resolver/manual/manual.go @@ -76,9 +76,11 @@ func (r *Resolver) InitialState(s resolver.State) { // Build returns itself for Resolver, because it's both a builder and a resolver. func (r *Resolver) Build(target resolver.Target, cc resolver.ClientConn, opts resolver.BuildOptions) (resolver.Resolver, error) { - r.BuildCallback(target, cc, opts) r.mu.Lock() defer r.mu.Unlock() + // Call BuildCallback after locking to avoid a race when UpdateState + // or ReportError is called before Build returns. + r.BuildCallback(target, cc, opts) r.CC = cc if r.lastSeenState != nil { err := r.CC.UpdateState(*r.lastSeenState) diff --git a/vendor/google.golang.org/grpc/resolver_wrapper.go b/vendor/google.golang.org/grpc/resolver_wrapper.go index c5fb45236faf6..23bb3fb258240 100644 --- a/vendor/google.golang.org/grpc/resolver_wrapper.go +++ b/vendor/google.golang.org/grpc/resolver_wrapper.go @@ -66,7 +66,7 @@ func newCCResolverWrapper(cc *ClientConn) *ccResolverWrapper { // any newly created ccResolverWrapper, except that close may be called instead. func (ccr *ccResolverWrapper) start() error { errCh := make(chan error) - ccr.serializer.Schedule(func(ctx context.Context) { + ccr.serializer.TrySchedule(func(ctx context.Context) { if ctx.Err() != nil { return } @@ -85,7 +85,7 @@ func (ccr *ccResolverWrapper) start() error { } func (ccr *ccResolverWrapper) resolveNow(o resolver.ResolveNowOptions) { - ccr.serializer.Schedule(func(ctx context.Context) { + ccr.serializer.TrySchedule(func(ctx context.Context) { if ctx.Err() != nil || ccr.resolver == nil { return } @@ -102,7 +102,7 @@ func (ccr *ccResolverWrapper) close() { ccr.closed = true ccr.mu.Unlock() - ccr.serializer.Schedule(func(context.Context) { + ccr.serializer.TrySchedule(func(context.Context) { if ccr.resolver == nil { return } @@ -177,6 +177,9 @@ func (ccr *ccResolverWrapper) ParseServiceConfig(scJSON string) *serviceconfig.P // addChannelzTraceEvent adds a channelz trace event containing the new // state received from resolver implementations. func (ccr *ccResolverWrapper) addChannelzTraceEvent(s resolver.State) { + if !logger.V(0) && !channelz.IsOn() { + return + } var updates []string var oldSC, newSC *ServiceConfig var oldOK, newOK bool diff --git a/vendor/google.golang.org/grpc/rpc_util.go b/vendor/google.golang.org/grpc/rpc_util.go index fdd49e6e91510..2d96f1405e8da 100644 --- a/vendor/google.golang.org/grpc/rpc_util.go +++ b/vendor/google.golang.org/grpc/rpc_util.go @@ -19,7 +19,6 @@ package grpc import ( - "bytes" "compress/gzip" "context" "encoding/binary" @@ -35,6 +34,7 @@ import ( "google.golang.org/grpc/encoding" "google.golang.org/grpc/encoding/proto" "google.golang.org/grpc/internal/transport" + "google.golang.org/grpc/mem" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" "google.golang.org/grpc/stats" @@ -220,8 +220,8 @@ type HeaderCallOption struct { HeaderAddr *metadata.MD } -func (o HeaderCallOption) before(c *callInfo) error { return nil } -func (o HeaderCallOption) after(c *callInfo, attempt *csAttempt) { +func (o HeaderCallOption) before(*callInfo) error { return nil } +func (o HeaderCallOption) after(_ *callInfo, attempt *csAttempt) { *o.HeaderAddr, _ = attempt.s.Header() } @@ -242,8 +242,8 @@ type TrailerCallOption struct { TrailerAddr *metadata.MD } -func (o TrailerCallOption) before(c *callInfo) error { return nil } -func (o TrailerCallOption) after(c *callInfo, attempt *csAttempt) { +func (o TrailerCallOption) before(*callInfo) error { return nil } +func (o TrailerCallOption) after(_ *callInfo, attempt *csAttempt) { *o.TrailerAddr = attempt.s.Trailer() } @@ -264,24 +264,20 @@ type PeerCallOption struct { PeerAddr *peer.Peer } -func (o PeerCallOption) before(c *callInfo) error { return nil } -func (o PeerCallOption) after(c *callInfo, attempt *csAttempt) { +func (o PeerCallOption) before(*callInfo) error { return nil } +func (o PeerCallOption) after(_ *callInfo, attempt *csAttempt) { if x, ok := peer.FromContext(attempt.s.Context()); ok { *o.PeerAddr = *x } } -// WaitForReady configures the action to take when an RPC is attempted on broken -// connections or unreachable servers. If waitForReady is false and the -// connection is in the TRANSIENT_FAILURE state, the RPC will fail -// immediately. Otherwise, the RPC client will block the call until a -// connection is available (or the call is canceled or times out) and will -// retry the call if it fails due to a transient error. gRPC will not retry if -// data was written to the wire unless the server indicates it did not process -// the data. Please refer to -// https://github.com/grpc/grpc/blob/master/doc/wait-for-ready.md. +// WaitForReady configures the RPC's behavior when the client is in +// TRANSIENT_FAILURE, which occurs when all addresses fail to connect. If +// waitForReady is false, the RPC will fail immediately. Otherwise, the client +// will wait until a connection becomes available or the RPC's deadline is +// reached. // -// By default, RPCs don't "wait for ready". +// By default, RPCs do not "wait for ready". func WaitForReady(waitForReady bool) CallOption { return FailFastCallOption{FailFast: !waitForReady} } @@ -308,7 +304,7 @@ func (o FailFastCallOption) before(c *callInfo) error { c.failFast = o.FailFast return nil } -func (o FailFastCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o FailFastCallOption) after(*callInfo, *csAttempt) {} // OnFinish returns a CallOption that configures a callback to be called when // the call completes. The error passed to the callback is the status of the @@ -343,7 +339,7 @@ func (o OnFinishCallOption) before(c *callInfo) error { return nil } -func (o OnFinishCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o OnFinishCallOption) after(*callInfo, *csAttempt) {} // MaxCallRecvMsgSize returns a CallOption which sets the maximum message size // in bytes the client can receive. If this is not set, gRPC uses the default @@ -367,7 +363,7 @@ func (o MaxRecvMsgSizeCallOption) before(c *callInfo) error { c.maxReceiveMessageSize = &o.MaxRecvMsgSize return nil } -func (o MaxRecvMsgSizeCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o MaxRecvMsgSizeCallOption) after(*callInfo, *csAttempt) {} // MaxCallSendMsgSize returns a CallOption which sets the maximum message size // in bytes the client can send. If this is not set, gRPC uses the default @@ -391,7 +387,7 @@ func (o MaxSendMsgSizeCallOption) before(c *callInfo) error { c.maxSendMessageSize = &o.MaxSendMsgSize return nil } -func (o MaxSendMsgSizeCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o MaxSendMsgSizeCallOption) after(*callInfo, *csAttempt) {} // PerRPCCredentials returns a CallOption that sets credentials.PerRPCCredentials // for a call. @@ -414,7 +410,7 @@ func (o PerRPCCredsCallOption) before(c *callInfo) error { c.creds = o.Creds return nil } -func (o PerRPCCredsCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o PerRPCCredsCallOption) after(*callInfo, *csAttempt) {} // UseCompressor returns a CallOption which sets the compressor used when // sending the request. If WithCompressor is also set, UseCompressor has @@ -442,7 +438,7 @@ func (o CompressorCallOption) before(c *callInfo) error { c.compressorType = o.CompressorType return nil } -func (o CompressorCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o CompressorCallOption) after(*callInfo, *csAttempt) {} // CallContentSubtype returns a CallOption that will set the content-subtype // for a call. For example, if content-subtype is "json", the Content-Type over @@ -479,7 +475,7 @@ func (o ContentSubtypeCallOption) before(c *callInfo) error { c.contentSubtype = o.ContentSubtype return nil } -func (o ContentSubtypeCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o ContentSubtypeCallOption) after(*callInfo, *csAttempt) {} // ForceCodec returns a CallOption that will set codec to be used for all // request and response messages for a call. The result of calling Name() will @@ -515,10 +511,50 @@ type ForceCodecCallOption struct { } func (o ForceCodecCallOption) before(c *callInfo) error { - c.codec = o.Codec + c.codec = newCodecV1Bridge(o.Codec) return nil } -func (o ForceCodecCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o ForceCodecCallOption) after(*callInfo, *csAttempt) {} + +// ForceCodecV2 returns a CallOption that will set codec to be used for all +// request and response messages for a call. The result of calling Name() will +// be used as the content-subtype after converting to lowercase, unless +// CallContentSubtype is also used. +// +// See Content-Type on +// https://github.com/grpc/grpc/blob/master/doc/PROTOCOL-HTTP2.md#requests for +// more details. Also see the documentation on RegisterCodec and +// CallContentSubtype for more details on the interaction between Codec and +// content-subtype. +// +// This function is provided for advanced users; prefer to use only +// CallContentSubtype to select a registered codec instead. +// +// # Experimental +// +// Notice: This API is EXPERIMENTAL and may be changed or removed in a +// later release. +func ForceCodecV2(codec encoding.CodecV2) CallOption { + return ForceCodecV2CallOption{CodecV2: codec} +} + +// ForceCodecV2CallOption is a CallOption that indicates the codec used for +// marshaling messages. +// +// # Experimental +// +// Notice: This type is EXPERIMENTAL and may be changed or removed in a +// later release. +type ForceCodecV2CallOption struct { + CodecV2 encoding.CodecV2 +} + +func (o ForceCodecV2CallOption) before(c *callInfo) error { + c.codec = o.CodecV2 + return nil +} + +func (o ForceCodecV2CallOption) after(*callInfo, *csAttempt) {} // CallCustomCodec behaves like ForceCodec, but accepts a grpc.Codec instead of // an encoding.Codec. @@ -540,10 +576,10 @@ type CustomCodecCallOption struct { } func (o CustomCodecCallOption) before(c *callInfo) error { - c.codec = o.Codec + c.codec = newCodecV0Bridge(o.Codec) return nil } -func (o CustomCodecCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o CustomCodecCallOption) after(*callInfo, *csAttempt) {} // MaxRetryRPCBufferSize returns a CallOption that limits the amount of memory // used for buffering this RPC's requests for retry purposes. @@ -571,7 +607,7 @@ func (o MaxRetryRPCBufferSizeCallOption) before(c *callInfo) error { c.maxRetryRPCBufferSize = o.MaxRetryRPCBufferSize return nil } -func (o MaxRetryRPCBufferSizeCallOption) after(c *callInfo, attempt *csAttempt) {} +func (o MaxRetryRPCBufferSizeCallOption) after(*callInfo, *csAttempt) {} // The format of the payload: compressed or not? type payloadFormat uint8 @@ -581,19 +617,28 @@ const ( compressionMade payloadFormat = 1 // compressed ) +func (pf payloadFormat) isCompressed() bool { + return pf == compressionMade +} + +type streamReader interface { + ReadHeader(header []byte) error + Read(n int) (mem.BufferSlice, error) +} + // parser reads complete gRPC messages from the underlying reader. type parser struct { // r is the underlying reader. // See the comment on recvMsg for the permissible // error types. - r io.Reader + r streamReader // The header of a gRPC message. Find more detail at // https://github.com/grpc/grpc/blob/master/doc/PROTOCOL-HTTP2.md header [5]byte - // recvBufferPool is the pool of shared receive buffers. - recvBufferPool SharedBufferPool + // bufferPool is the pool of shared receive buffers. + bufferPool mem.BufferPool } // recvMsg reads a complete gRPC message from the stream. @@ -608,14 +653,15 @@ type parser struct { // - an error from the status package // // No other error values or types must be returned, which also means -// that the underlying io.Reader must not return an incompatible +// that the underlying streamReader must not return an incompatible // error. -func (p *parser) recvMsg(maxReceiveMessageSize int) (pf payloadFormat, msg []byte, err error) { - if _, err := p.r.Read(p.header[:]); err != nil { +func (p *parser) recvMsg(maxReceiveMessageSize int) (payloadFormat, mem.BufferSlice, error) { + err := p.r.ReadHeader(p.header[:]) + if err != nil { return 0, nil, err } - pf = payloadFormat(p.header[0]) + pf := payloadFormat(p.header[0]) length := binary.BigEndian.Uint32(p.header[1:]) if length == 0 { @@ -627,20 +673,21 @@ func (p *parser) recvMsg(maxReceiveMessageSize int) (pf payloadFormat, msg []byt if int(length) > maxReceiveMessageSize { return 0, nil, status.Errorf(codes.ResourceExhausted, "grpc: received message larger than max (%d vs. %d)", length, maxReceiveMessageSize) } - msg = p.recvBufferPool.Get(int(length)) - if _, err := p.r.Read(msg); err != nil { + + data, err := p.r.Read(int(length)) + if err != nil { if err == io.EOF { err = io.ErrUnexpectedEOF } return 0, nil, err } - return pf, msg, nil + return pf, data, nil } // encode serializes msg and returns a buffer containing the message, or an // error if it is too large to be transmitted by grpc. If msg is nil, it // generates an empty message. -func encode(c baseCodec, msg any) ([]byte, error) { +func encode(c baseCodec, msg any) (mem.BufferSlice, error) { if msg == nil { // NOTE: typed nils will not be caught by this check return nil, nil } @@ -648,7 +695,8 @@ func encode(c baseCodec, msg any) ([]byte, error) { if err != nil { return nil, status.Errorf(codes.Internal, "grpc: error while marshaling: %v", err.Error()) } - if uint(len(b)) > math.MaxUint32 { + if uint(b.Len()) > math.MaxUint32 { + b.Free() return nil, status.Errorf(codes.ResourceExhausted, "grpc: message too large (%d bytes)", len(b)) } return b, nil @@ -659,34 +707,41 @@ func encode(c baseCodec, msg any) ([]byte, error) { // indicating no compression was done. // // TODO(dfawley): eliminate cp parameter by wrapping Compressor in an encoding.Compressor. -func compress(in []byte, cp Compressor, compressor encoding.Compressor) ([]byte, error) { - if compressor == nil && cp == nil { - return nil, nil - } - if len(in) == 0 { - return nil, nil +func compress(in mem.BufferSlice, cp Compressor, compressor encoding.Compressor, pool mem.BufferPool) (mem.BufferSlice, payloadFormat, error) { + if (compressor == nil && cp == nil) || in.Len() == 0 { + return nil, compressionNone, nil } + var out mem.BufferSlice + w := mem.NewWriter(&out, pool) wrapErr := func(err error) error { + out.Free() return status.Errorf(codes.Internal, "grpc: error while compressing: %v", err.Error()) } - cbuf := &bytes.Buffer{} if compressor != nil { - z, err := compressor.Compress(cbuf) + z, err := compressor.Compress(w) if err != nil { - return nil, wrapErr(err) + return nil, 0, wrapErr(err) } - if _, err := z.Write(in); err != nil { - return nil, wrapErr(err) + for _, b := range in { + if _, err := z.Write(b.ReadOnlyData()); err != nil { + return nil, 0, wrapErr(err) + } } if err := z.Close(); err != nil { - return nil, wrapErr(err) + return nil, 0, wrapErr(err) } } else { - if err := cp.Do(cbuf, in); err != nil { - return nil, wrapErr(err) + // This is obviously really inefficient since it fully materializes the data, but + // there is no way around this with the old Compressor API. At least it attempts + // to return the buffer to the provider, in the hopes it can be reused (maybe + // even by a subsequent call to this very function). + buf := in.MaterializeToBuffer(pool) + defer buf.Free() + if err := cp.Do(w, buf.ReadOnlyData()); err != nil { + return nil, 0, wrapErr(err) } } - return cbuf.Bytes(), nil + return out, compressionMade, nil } const ( @@ -697,33 +752,36 @@ const ( // msgHeader returns a 5-byte header for the message being transmitted and the // payload, which is compData if non-nil or data otherwise. -func msgHeader(data, compData []byte) (hdr []byte, payload []byte) { +func msgHeader(data, compData mem.BufferSlice, pf payloadFormat) (hdr []byte, payload mem.BufferSlice) { hdr = make([]byte, headerLen) - if compData != nil { - hdr[0] = byte(compressionMade) - data = compData + hdr[0] = byte(pf) + + var length uint32 + if pf.isCompressed() { + length = uint32(compData.Len()) + payload = compData } else { - hdr[0] = byte(compressionNone) + length = uint32(data.Len()) + payload = data } // Write length of payload into buf - binary.BigEndian.PutUint32(hdr[payloadLen:], uint32(len(data))) - return hdr, data + binary.BigEndian.PutUint32(hdr[payloadLen:], length) + return hdr, payload } -func outPayload(client bool, msg any, data, payload []byte, t time.Time) *stats.OutPayload { +func outPayload(client bool, msg any, dataLength, payloadLength int, t time.Time) *stats.OutPayload { return &stats.OutPayload{ Client: client, Payload: msg, - Data: data, - Length: len(data), - WireLength: len(payload) + headerLen, - CompressedLength: len(payload), + Length: dataLength, + WireLength: payloadLength + headerLen, + CompressedLength: payloadLength, SentTime: t, } } -func checkRecvPayload(pf payloadFormat, recvCompress string, haveCompressor bool) *status.Status { +func checkRecvPayload(pf payloadFormat, recvCompress string, haveCompressor bool, isServer bool) *status.Status { switch pf { case compressionNone: case compressionMade: @@ -731,7 +789,11 @@ func checkRecvPayload(pf payloadFormat, recvCompress string, haveCompressor bool return status.New(codes.Internal, "grpc: compressed flag set with identity or empty encoding") } if !haveCompressor { - return status.Newf(codes.Unimplemented, "grpc: Decompressor is not installed for grpc-encoding %q", recvCompress) + if isServer { + return status.Newf(codes.Unimplemented, "grpc: Decompressor is not installed for grpc-encoding %q", recvCompress) + } else { + return status.Newf(codes.Internal, "grpc: Decompressor is not installed for grpc-encoding %q", recvCompress) + } } default: return status.Newf(codes.Internal, "grpc: received unexpected payload format %d", pf) @@ -741,104 +803,129 @@ func checkRecvPayload(pf payloadFormat, recvCompress string, haveCompressor bool type payloadInfo struct { compressedLength int // The compressed length got from wire. - uncompressedBytes []byte + uncompressedBytes mem.BufferSlice +} + +func (p *payloadInfo) free() { + if p != nil && p.uncompressedBytes != nil { + p.uncompressedBytes.Free() + } } // recvAndDecompress reads a message from the stream, decompressing it if necessary. // // Cancelling the returned cancel function releases the buffer back to the pool. So the caller should cancel as soon as // the buffer is no longer needed. -func recvAndDecompress(p *parser, s *transport.Stream, dc Decompressor, maxReceiveMessageSize int, payInfo *payloadInfo, compressor encoding.Compressor, -) (uncompressedBuf []byte, cancel func(), err error) { - pf, compressedBuf, err := p.recvMsg(maxReceiveMessageSize) +// TODO: Refactor this function to reduce the number of arguments. +// See: https://google.github.io/styleguide/go/best-practices.html#function-argument-lists +func recvAndDecompress(p *parser, s *transport.Stream, dc Decompressor, maxReceiveMessageSize int, payInfo *payloadInfo, compressor encoding.Compressor, isServer bool, +) (out mem.BufferSlice, err error) { + pf, compressed, err := p.recvMsg(maxReceiveMessageSize) if err != nil { - return nil, nil, err + return nil, err } - if st := checkRecvPayload(pf, s.RecvCompress(), compressor != nil || dc != nil); st != nil { - return nil, nil, st.Err() + compressedLength := compressed.Len() + + if st := checkRecvPayload(pf, s.RecvCompress(), compressor != nil || dc != nil, isServer); st != nil { + compressed.Free() + return nil, st.Err() } var size int - if pf == compressionMade { + if pf.isCompressed() { + defer compressed.Free() + // To match legacy behavior, if the decompressor is set by WithDecompressor or RPCDecompressor, // use this decompressor as the default. if dc != nil { - uncompressedBuf, err = dc.Do(bytes.NewReader(compressedBuf)) + var uncompressedBuf []byte + uncompressedBuf, err = dc.Do(compressed.Reader()) + if err == nil { + out = mem.BufferSlice{mem.NewBuffer(&uncompressedBuf, nil)} + } size = len(uncompressedBuf) } else { - uncompressedBuf, size, err = decompress(compressor, compressedBuf, maxReceiveMessageSize) + out, size, err = decompress(compressor, compressed, maxReceiveMessageSize, p.bufferPool) } if err != nil { - return nil, nil, status.Errorf(codes.Internal, "grpc: failed to decompress the received message: %v", err) + return nil, status.Errorf(codes.Internal, "grpc: failed to decompress the received message: %v", err) } if size > maxReceiveMessageSize { + out.Free() // TODO: Revisit the error code. Currently keep it consistent with java // implementation. - return nil, nil, status.Errorf(codes.ResourceExhausted, "grpc: received message after decompression larger than max (%d vs. %d)", size, maxReceiveMessageSize) + return nil, status.Errorf(codes.ResourceExhausted, "grpc: received message after decompression larger than max (%d vs. %d)", size, maxReceiveMessageSize) } } else { - uncompressedBuf = compressedBuf + out = compressed } if payInfo != nil { - payInfo.compressedLength = len(compressedBuf) - payInfo.uncompressedBytes = uncompressedBuf - - cancel = func() {} - } else { - cancel = func() { - p.recvBufferPool.Put(&compressedBuf) - } + payInfo.compressedLength = compressedLength + out.Ref() + payInfo.uncompressedBytes = out } - return uncompressedBuf, cancel, nil + return out, nil } // Using compressor, decompress d, returning data and size. // Optionally, if data will be over maxReceiveMessageSize, just return the size. -func decompress(compressor encoding.Compressor, d []byte, maxReceiveMessageSize int) ([]byte, int, error) { - dcReader, err := compressor.Decompress(bytes.NewReader(d)) +func decompress(compressor encoding.Compressor, d mem.BufferSlice, maxReceiveMessageSize int, pool mem.BufferPool) (mem.BufferSlice, int, error) { + dcReader, err := compressor.Decompress(d.Reader()) if err != nil { return nil, 0, err } - if sizer, ok := compressor.(interface { - DecompressedSize(compressedBytes []byte) int - }); ok { - if size := sizer.DecompressedSize(d); size >= 0 { - if size > maxReceiveMessageSize { - return nil, size, nil - } - // size is used as an estimate to size the buffer, but we - // will read more data if available. - // +MinRead so ReadFrom will not reallocate if size is correct. - // - // TODO: If we ensure that the buffer size is the same as the DecompressedSize, - // we can also utilize the recv buffer pool here. - buf := bytes.NewBuffer(make([]byte, 0, size+bytes.MinRead)) - bytesRead, err := buf.ReadFrom(io.LimitReader(dcReader, int64(maxReceiveMessageSize)+1)) - return buf.Bytes(), int(bytesRead), err - } + + // TODO: Can/should this still be preserved with the new BufferSlice API? Are + // there any actual benefits to allocating a single large buffer instead of + // multiple smaller ones? + //if sizer, ok := compressor.(interface { + // DecompressedSize(compressedBytes []byte) int + //}); ok { + // if size := sizer.DecompressedSize(d); size >= 0 { + // if size > maxReceiveMessageSize { + // return nil, size, nil + // } + // // size is used as an estimate to size the buffer, but we + // // will read more data if available. + // // +MinRead so ReadFrom will not reallocate if size is correct. + // // + // // TODO: If we ensure that the buffer size is the same as the DecompressedSize, + // // we can also utilize the recv buffer pool here. + // buf := bytes.NewBuffer(make([]byte, 0, size+bytes.MinRead)) + // bytesRead, err := buf.ReadFrom(io.LimitReader(dcReader, int64(maxReceiveMessageSize)+1)) + // return buf.Bytes(), int(bytesRead), err + // } + //} + + var out mem.BufferSlice + _, err = io.Copy(mem.NewWriter(&out, pool), io.LimitReader(dcReader, int64(maxReceiveMessageSize)+1)) + if err != nil { + out.Free() + return nil, 0, err } - // Read from LimitReader with limit max+1. So if the underlying - // reader is over limit, the result will be bigger than max. - d, err = io.ReadAll(io.LimitReader(dcReader, int64(maxReceiveMessageSize)+1)) - return d, len(d), err + return out, out.Len(), nil } // For the two compressor parameters, both should not be set, but if they are, // dc takes precedence over compressor. // TODO(dfawley): wrap the old compressor/decompressor using the new API? -func recv(p *parser, c baseCodec, s *transport.Stream, dc Decompressor, m any, maxReceiveMessageSize int, payInfo *payloadInfo, compressor encoding.Compressor) error { - buf, cancel, err := recvAndDecompress(p, s, dc, maxReceiveMessageSize, payInfo, compressor) +func recv(p *parser, c baseCodec, s *transport.Stream, dc Decompressor, m any, maxReceiveMessageSize int, payInfo *payloadInfo, compressor encoding.Compressor, isServer bool) error { + data, err := recvAndDecompress(p, s, dc, maxReceiveMessageSize, payInfo, compressor, isServer) if err != nil { return err } - defer cancel() - if err := c.Unmarshal(buf, m); err != nil { + // If the codec wants its own reference to the data, it can get it. Otherwise, always + // free the buffers. + defer data.Free() + + if err := c.Unmarshal(data, m); err != nil { return status.Errorf(codes.Internal, "grpc: failed to unmarshal the received message: %v", err) } + return nil } @@ -941,7 +1028,7 @@ func setCallInfoCodec(c *callInfo) error { // encoding.Codec (Name vs. String method name). We only support // setting content subtype from encoding.Codec to avoid a behavior // change with the deprecated version. - if ec, ok := c.codec.(encoding.Codec); ok { + if ec, ok := c.codec.(encoding.CodecV2); ok { c.contentSubtype = strings.ToLower(ec.Name()) } } @@ -950,12 +1037,12 @@ func setCallInfoCodec(c *callInfo) error { if c.contentSubtype == "" { // No codec specified in CallOptions; use proto by default. - c.codec = encoding.GetCodec(proto.Name) + c.codec = getCodec(proto.Name) return nil } // c.contentSubtype is already lowercased in CallContentSubtype - c.codec = encoding.GetCodec(c.contentSubtype) + c.codec = getCodec(c.contentSubtype) if c.codec == nil { return status.Errorf(codes.Internal, "no codec registered for content-subtype %s", c.contentSubtype) } diff --git a/vendor/google.golang.org/grpc/server.go b/vendor/google.golang.org/grpc/server.go index 89f8e4792bf15..d1e1415a40f9b 100644 --- a/vendor/google.golang.org/grpc/server.go +++ b/vendor/google.golang.org/grpc/server.go @@ -45,6 +45,7 @@ import ( "google.golang.org/grpc/internal/grpcutil" "google.golang.org/grpc/internal/transport" "google.golang.org/grpc/keepalive" + "google.golang.org/grpc/mem" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" "google.golang.org/grpc/stats" @@ -80,7 +81,7 @@ func init() { } internal.BinaryLogger = binaryLogger internal.JoinServerOptions = newJoinServerOption - internal.RecvBufferPool = recvBufferPool + internal.BufferPool = bufferPool } var statusOK = status.New(codes.OK, "") @@ -170,7 +171,7 @@ type serverOptions struct { maxHeaderListSize *uint32 headerTableSize *uint32 numServerWorkers uint32 - recvBufferPool SharedBufferPool + bufferPool mem.BufferPool waitForHandlers bool } @@ -181,7 +182,7 @@ var defaultServerOptions = serverOptions{ connectionTimeout: 120 * time.Second, writeBufferSize: defaultWriteBufSize, readBufferSize: defaultReadBufSize, - recvBufferPool: nopBufferPool{}, + bufferPool: mem.DefaultBufferPool(), } var globalServerOptions []ServerOption @@ -313,7 +314,7 @@ func KeepaliveEnforcementPolicy(kep keepalive.EnforcementPolicy) ServerOption { // Will be supported throughout 1.x. func CustomCodec(codec Codec) ServerOption { return newFuncServerOption(func(o *serverOptions) { - o.codec = codec + o.codec = newCodecV0Bridge(codec) }) } @@ -342,7 +343,22 @@ func CustomCodec(codec Codec) ServerOption { // later release. func ForceServerCodec(codec encoding.Codec) ServerOption { return newFuncServerOption(func(o *serverOptions) { - o.codec = codec + o.codec = newCodecV1Bridge(codec) + }) +} + +// ForceServerCodecV2 is the equivalent of ForceServerCodec, but for the new +// CodecV2 interface. +// +// Will be supported throughout 1.x. +// +// # Experimental +// +// Notice: This API is EXPERIMENTAL and may be changed or removed in a +// later release. +func ForceServerCodecV2(codecV2 encoding.CodecV2) ServerOption { + return newFuncServerOption(func(o *serverOptions) { + o.codec = codecV2 }) } @@ -592,26 +608,9 @@ func WaitForHandlers(w bool) ServerOption { }) } -// RecvBufferPool returns a ServerOption that configures the server -// to use the provided shared buffer pool for parsing incoming messages. Depending -// on the application's workload, this could result in reduced memory allocation. -// -// If you are unsure about how to implement a memory pool but want to utilize one, -// begin with grpc.NewSharedBufferPool. -// -// Note: The shared buffer pool feature will not be active if any of the following -// options are used: StatsHandler, EnableTracing, or binary logging. In such -// cases, the shared buffer pool will be ignored. -// -// Deprecated: use experimental.WithRecvBufferPool instead. Will be deleted in -// v1.60.0 or later. -func RecvBufferPool(bufferPool SharedBufferPool) ServerOption { - return recvBufferPool(bufferPool) -} - -func recvBufferPool(bufferPool SharedBufferPool) ServerOption { +func bufferPool(bufferPool mem.BufferPool) ServerOption { return newFuncServerOption(func(o *serverOptions) { - o.recvBufferPool = bufferPool + o.bufferPool = bufferPool }) } @@ -622,7 +621,7 @@ func recvBufferPool(bufferPool SharedBufferPool) ServerOption { // workload (assuming a QPS of a few thousand requests/sec). const serverWorkerResetThreshold = 1 << 16 -// serverWorkers blocks on a *transport.Stream channel forever and waits for +// serverWorker blocks on a *transport.Stream channel forever and waits for // data to be fed by serveStreams. This allows multiple requests to be // processed by the same goroutine, removing the need for expensive stack // re-allocations (see the runtime.morestack problem [1]). @@ -980,6 +979,7 @@ func (s *Server) newHTTP2Transport(c net.Conn) transport.ServerTransport { ChannelzParent: s.channelz, MaxHeaderListSize: s.opts.maxHeaderListSize, HeaderTableSize: s.opts.headerTableSize, + BufferPool: s.opts.bufferPool, } st, err := transport.NewServerTransport(c, config) if err != nil { @@ -1072,7 +1072,7 @@ var _ http.Handler = (*Server)(nil) // Notice: This API is EXPERIMENTAL and may be changed or removed in a // later release. func (s *Server) ServeHTTP(w http.ResponseWriter, r *http.Request) { - st, err := transport.NewServerHandlerTransport(w, r, s.opts.statsHandlers) + st, err := transport.NewServerHandlerTransport(w, r, s.opts.statsHandlers, s.opts.bufferPool) if err != nil { // Errors returned from transport.NewServerHandlerTransport have // already been written to w. @@ -1142,20 +1142,35 @@ func (s *Server) sendResponse(ctx context.Context, t transport.ServerTransport, channelz.Error(logger, s.channelz, "grpc: server failed to encode response: ", err) return err } - compData, err := compress(data, cp, comp) + + compData, pf, err := compress(data, cp, comp, s.opts.bufferPool) if err != nil { + data.Free() channelz.Error(logger, s.channelz, "grpc: server failed to compress response: ", err) return err } - hdr, payload := msgHeader(data, compData) + + hdr, payload := msgHeader(data, compData, pf) + + defer func() { + compData.Free() + data.Free() + // payload does not need to be freed here, it is either data or compData, both of + // which are already freed. + }() + + dataLen := data.Len() + payloadLen := payload.Len() // TODO(dfawley): should we be checking len(data) instead? - if len(payload) > s.opts.maxSendMessageSize { - return status.Errorf(codes.ResourceExhausted, "grpc: trying to send message larger than max (%d vs. %d)", len(payload), s.opts.maxSendMessageSize) + if payloadLen > s.opts.maxSendMessageSize { + return status.Errorf(codes.ResourceExhausted, "grpc: trying to send message larger than max (%d vs. %d)", payloadLen, s.opts.maxSendMessageSize) } err = t.Write(stream, hdr, payload, opts) if err == nil { - for _, sh := range s.opts.statsHandlers { - sh.HandleRPC(ctx, outPayload(false, msg, data, payload, time.Now())) + if len(s.opts.statsHandlers) != 0 { + for _, sh := range s.opts.statsHandlers { + sh.HandleRPC(ctx, outPayload(false, msg, dataLen, payloadLen, time.Now())) + } } } return err @@ -1334,37 +1349,37 @@ func (s *Server) processUnaryRPC(ctx context.Context, t transport.ServerTranspor var payInfo *payloadInfo if len(shs) != 0 || len(binlogs) != 0 { payInfo = &payloadInfo{} + defer payInfo.free() } - d, cancel, err := recvAndDecompress(&parser{r: stream, recvBufferPool: s.opts.recvBufferPool}, stream, dc, s.opts.maxReceiveMessageSize, payInfo, decomp) + d, err := recvAndDecompress(&parser{r: stream, bufferPool: s.opts.bufferPool}, stream, dc, s.opts.maxReceiveMessageSize, payInfo, decomp, true) if err != nil { if e := t.WriteStatus(stream, status.Convert(err)); e != nil { channelz.Warningf(logger, s.channelz, "grpc: Server.processUnaryRPC failed to write status: %v", e) } return err } + defer d.Free() if channelz.IsOn() { t.IncrMsgRecv() } df := func(v any) error { - defer cancel() - if err := s.getCodec(stream.ContentSubtype()).Unmarshal(d, v); err != nil { return status.Errorf(codes.Internal, "grpc: error unmarshalling request: %v", err) } + for _, sh := range shs { sh.HandleRPC(ctx, &stats.InPayload{ RecvTime: time.Now(), Payload: v, - Length: len(d), + Length: d.Len(), WireLength: payInfo.compressedLength + headerLen, CompressedLength: payInfo.compressedLength, - Data: d, }) } if len(binlogs) != 0 { cm := &binarylog.ClientMessage{ - Message: d, + Message: d.Materialize(), } for _, binlog := range binlogs { binlog.Log(ctx, cm) @@ -1548,7 +1563,7 @@ func (s *Server) processStreamingRPC(ctx context.Context, t transport.ServerTran ctx: ctx, t: t, s: stream, - p: &parser{r: stream, recvBufferPool: s.opts.recvBufferPool}, + p: &parser{r: stream, bufferPool: s.opts.bufferPool}, codec: s.getCodec(stream.ContentSubtype()), maxReceiveMessageSize: s.opts.maxReceiveMessageSize, maxSendMessageSize: s.opts.maxSendMessageSize, @@ -1963,12 +1978,12 @@ func (s *Server) getCodec(contentSubtype string) baseCodec { return s.opts.codec } if contentSubtype == "" { - return encoding.GetCodec(proto.Name) + return getCodec(proto.Name) } - codec := encoding.GetCodec(contentSubtype) + codec := getCodec(contentSubtype) if codec == nil { logger.Warningf("Unsupported codec %q. Defaulting to %q for now. This will start to fail in future releases.", contentSubtype, proto.Name) - return encoding.GetCodec(proto.Name) + return getCodec(proto.Name) } return codec } diff --git a/vendor/google.golang.org/grpc/shared_buffer_pool.go b/vendor/google.golang.org/grpc/shared_buffer_pool.go deleted file mode 100644 index 48a64cfe8e256..0000000000000 --- a/vendor/google.golang.org/grpc/shared_buffer_pool.go +++ /dev/null @@ -1,154 +0,0 @@ -/* - * - * Copyright 2023 gRPC authors. - * - * Licensed under the Apache License, Version 2.0 (the "License"); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - * - */ - -package grpc - -import "sync" - -// SharedBufferPool is a pool of buffers that can be shared, resulting in -// decreased memory allocation. Currently, in gRPC-go, it is only utilized -// for parsing incoming messages. -// -// # Experimental -// -// Notice: This API is EXPERIMENTAL and may be changed or removed in a -// later release. -type SharedBufferPool interface { - // Get returns a buffer with specified length from the pool. - // - // The returned byte slice may be not zero initialized. - Get(length int) []byte - - // Put returns a buffer to the pool. - Put(*[]byte) -} - -// NewSharedBufferPool creates a simple SharedBufferPool with buckets -// of different sizes to optimize memory usage. This prevents the pool from -// wasting large amounts of memory, even when handling messages of varying sizes. -// -// # Experimental -// -// Notice: This API is EXPERIMENTAL and may be changed or removed in a -// later release. -func NewSharedBufferPool() SharedBufferPool { - return &simpleSharedBufferPool{ - pools: [poolArraySize]simpleSharedBufferChildPool{ - newBytesPool(level0PoolMaxSize), - newBytesPool(level1PoolMaxSize), - newBytesPool(level2PoolMaxSize), - newBytesPool(level3PoolMaxSize), - newBytesPool(level4PoolMaxSize), - newBytesPool(0), - }, - } -} - -// simpleSharedBufferPool is a simple implementation of SharedBufferPool. -type simpleSharedBufferPool struct { - pools [poolArraySize]simpleSharedBufferChildPool -} - -func (p *simpleSharedBufferPool) Get(size int) []byte { - return p.pools[p.poolIdx(size)].Get(size) -} - -func (p *simpleSharedBufferPool) Put(bs *[]byte) { - p.pools[p.poolIdx(cap(*bs))].Put(bs) -} - -func (p *simpleSharedBufferPool) poolIdx(size int) int { - switch { - case size <= level0PoolMaxSize: - return level0PoolIdx - case size <= level1PoolMaxSize: - return level1PoolIdx - case size <= level2PoolMaxSize: - return level2PoolIdx - case size <= level3PoolMaxSize: - return level3PoolIdx - case size <= level4PoolMaxSize: - return level4PoolIdx - default: - return levelMaxPoolIdx - } -} - -const ( - level0PoolMaxSize = 16 // 16 B - level1PoolMaxSize = level0PoolMaxSize * 16 // 256 B - level2PoolMaxSize = level1PoolMaxSize * 16 // 4 KB - level3PoolMaxSize = level2PoolMaxSize * 16 // 64 KB - level4PoolMaxSize = level3PoolMaxSize * 16 // 1 MB -) - -const ( - level0PoolIdx = iota - level1PoolIdx - level2PoolIdx - level3PoolIdx - level4PoolIdx - levelMaxPoolIdx - poolArraySize -) - -type simpleSharedBufferChildPool interface { - Get(size int) []byte - Put(any) -} - -type bufferPool struct { - sync.Pool - - defaultSize int -} - -func (p *bufferPool) Get(size int) []byte { - bs := p.Pool.Get().(*[]byte) - - if cap(*bs) < size { - p.Pool.Put(bs) - - return make([]byte, size) - } - - return (*bs)[:size] -} - -func newBytesPool(size int) simpleSharedBufferChildPool { - return &bufferPool{ - Pool: sync.Pool{ - New: func() any { - bs := make([]byte, size) - return &bs - }, - }, - defaultSize: size, - } -} - -// nopBufferPool is a buffer pool just makes new buffer without pooling. -type nopBufferPool struct { -} - -func (nopBufferPool) Get(length int) []byte { - return make([]byte, length) -} - -func (nopBufferPool) Put(*[]byte) { -} diff --git a/vendor/google.golang.org/grpc/stats/stats.go b/vendor/google.golang.org/grpc/stats/stats.go index fdb0bd65182c5..71195c4943d7a 100644 --- a/vendor/google.golang.org/grpc/stats/stats.go +++ b/vendor/google.golang.org/grpc/stats/stats.go @@ -77,9 +77,6 @@ type InPayload struct { // the call to HandleRPC which provides the InPayload returns and must be // copied if needed later. Payload any - // Data is the serialized message payload. - // Deprecated: Data will be removed in the next release. - Data []byte // Length is the size of the uncompressed payload data. Does not include any // framing (gRPC or HTTP/2). @@ -150,9 +147,6 @@ type OutPayload struct { // the call to HandleRPC which provides the OutPayload returns and must be // copied if needed later. Payload any - // Data is the serialized message payload. - // Deprecated: Data will be removed in the next release. - Data []byte // Length is the size of the uncompressed payload data. Does not include any // framing (gRPC or HTTP/2). Length int diff --git a/vendor/google.golang.org/grpc/stream.go b/vendor/google.golang.org/grpc/stream.go index 8051ef5b514a3..bb2b2a216ce24 100644 --- a/vendor/google.golang.org/grpc/stream.go +++ b/vendor/google.golang.org/grpc/stream.go @@ -41,6 +41,7 @@ import ( "google.golang.org/grpc/internal/serviceconfig" istatus "google.golang.org/grpc/internal/status" "google.golang.org/grpc/internal/transport" + "google.golang.org/grpc/mem" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" "google.golang.org/grpc/stats" @@ -359,7 +360,7 @@ func newClientStreamWithParams(ctx context.Context, desc *StreamDesc, cc *Client cs.attempt = a return nil } - if err := cs.withRetry(op, func() { cs.bufferForRetryLocked(0, op) }); err != nil { + if err := cs.withRetry(op, func() { cs.bufferForRetryLocked(0, op, nil) }); err != nil { return nil, err } @@ -517,7 +518,7 @@ func (a *csAttempt) newStream() error { } a.s = s a.ctx = s.Context() - a.p = &parser{r: s, recvBufferPool: a.cs.cc.dopts.recvBufferPool} + a.p = &parser{r: s, bufferPool: a.cs.cc.dopts.copts.BufferPool} return nil } @@ -566,10 +567,15 @@ type clientStream struct { // place where we need to check if the attempt is nil. attempt *csAttempt // TODO(hedging): hedging will have multiple attempts simultaneously. - committed bool // active attempt committed for retry? - onCommit func() - buffer []func(a *csAttempt) error // operations to replay on retry - bufferSize int // current size of buffer + committed bool // active attempt committed for retry? + onCommit func() + replayBuffer []replayOp // operations to replay on retry + replayBufferSize int // current size of replayBuffer +} + +type replayOp struct { + op func(a *csAttempt) error + cleanup func() } // csAttempt implements a single transport stream attempt within a @@ -607,7 +613,12 @@ func (cs *clientStream) commitAttemptLocked() { cs.onCommit() } cs.committed = true - cs.buffer = nil + for _, op := range cs.replayBuffer { + if op.cleanup != nil { + op.cleanup() + } + } + cs.replayBuffer = nil } func (cs *clientStream) commitAttempt() { @@ -732,7 +743,7 @@ func (cs *clientStream) retryLocked(attempt *csAttempt, lastErr error) error { // the stream is canceled. return err } - // Note that the first op in the replay buffer always sets cs.attempt + // Note that the first op in replayBuffer always sets cs.attempt // if it is able to pick a transport and create a stream. if lastErr = cs.replayBufferLocked(attempt); lastErr == nil { return nil @@ -761,7 +772,7 @@ func (cs *clientStream) withRetry(op func(a *csAttempt) error, onSuccess func()) // already be status errors. return toRPCErr(op(cs.attempt)) } - if len(cs.buffer) == 0 { + if len(cs.replayBuffer) == 0 { // For the first op, which controls creation of the stream and // assigns cs.attempt, we need to create a new attempt inline // before executing the first op. On subsequent ops, the attempt @@ -851,25 +862,26 @@ func (cs *clientStream) Trailer() metadata.MD { } func (cs *clientStream) replayBufferLocked(attempt *csAttempt) error { - for _, f := range cs.buffer { - if err := f(attempt); err != nil { + for _, f := range cs.replayBuffer { + if err := f.op(attempt); err != nil { return err } } return nil } -func (cs *clientStream) bufferForRetryLocked(sz int, op func(a *csAttempt) error) { +func (cs *clientStream) bufferForRetryLocked(sz int, op func(a *csAttempt) error, cleanup func()) { // Note: we still will buffer if retry is disabled (for transparent retries). if cs.committed { return } - cs.bufferSize += sz - if cs.bufferSize > cs.callInfo.maxRetryRPCBufferSize { + cs.replayBufferSize += sz + if cs.replayBufferSize > cs.callInfo.maxRetryRPCBufferSize { cs.commitAttemptLocked() + cleanup() return } - cs.buffer = append(cs.buffer, op) + cs.replayBuffer = append(cs.replayBuffer, replayOp{op: op, cleanup: cleanup}) } func (cs *clientStream) SendMsg(m any) (err error) { @@ -891,23 +903,50 @@ func (cs *clientStream) SendMsg(m any) (err error) { } // load hdr, payload, data - hdr, payload, data, err := prepareMsg(m, cs.codec, cs.cp, cs.comp) + hdr, data, payload, pf, err := prepareMsg(m, cs.codec, cs.cp, cs.comp, cs.cc.dopts.copts.BufferPool) if err != nil { return err } + defer func() { + data.Free() + // only free payload if compression was made, and therefore it is a different set + // of buffers from data. + if pf.isCompressed() { + payload.Free() + } + }() + + dataLen := data.Len() + payloadLen := payload.Len() // TODO(dfawley): should we be checking len(data) instead? - if len(payload) > *cs.callInfo.maxSendMessageSize { - return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", len(payload), *cs.callInfo.maxSendMessageSize) + if payloadLen > *cs.callInfo.maxSendMessageSize { + return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", payloadLen, *cs.callInfo.maxSendMessageSize) } + + // always take an extra ref in case data == payload (i.e. when the data isn't + // compressed). The original ref will always be freed by the deferred free above. + payload.Ref() op := func(a *csAttempt) error { - return a.sendMsg(m, hdr, payload, data) + return a.sendMsg(m, hdr, payload, dataLen, payloadLen) + } + + // onSuccess is invoked when the op is captured for a subsequent retry. If the + // stream was established by a previous message and therefore retries are + // disabled, onSuccess will not be invoked, and payloadRef can be freed + // immediately. + onSuccessCalled := false + err = cs.withRetry(op, func() { + cs.bufferForRetryLocked(len(hdr)+payloadLen, op, payload.Free) + onSuccessCalled = true + }) + if !onSuccessCalled { + payload.Free() } - err = cs.withRetry(op, func() { cs.bufferForRetryLocked(len(hdr)+len(payload), op) }) if len(cs.binlogs) != 0 && err == nil { cm := &binarylog.ClientMessage{ OnClientSide: true, - Message: data, + Message: data.Materialize(), } for _, binlog := range cs.binlogs { binlog.Log(cs.ctx, cm) @@ -924,6 +963,7 @@ func (cs *clientStream) RecvMsg(m any) error { var recvInfo *payloadInfo if len(cs.binlogs) != 0 { recvInfo = &payloadInfo{} + defer recvInfo.free() } err := cs.withRetry(func(a *csAttempt) error { return a.recvMsg(m, recvInfo) @@ -931,7 +971,7 @@ func (cs *clientStream) RecvMsg(m any) error { if len(cs.binlogs) != 0 && err == nil { sm := &binarylog.ServerMessage{ OnClientSide: true, - Message: recvInfo.uncompressedBytes, + Message: recvInfo.uncompressedBytes.Materialize(), } for _, binlog := range cs.binlogs { binlog.Log(cs.ctx, sm) @@ -958,7 +998,7 @@ func (cs *clientStream) CloseSend() error { // RecvMsg. This also matches historical behavior. return nil } - cs.withRetry(op, func() { cs.bufferForRetryLocked(0, op) }) + cs.withRetry(op, func() { cs.bufferForRetryLocked(0, op, nil) }) if len(cs.binlogs) != 0 { chc := &binarylog.ClientHalfClose{ OnClientSide: true, @@ -1034,7 +1074,7 @@ func (cs *clientStream) finish(err error) { cs.cancel() } -func (a *csAttempt) sendMsg(m any, hdr, payld, data []byte) error { +func (a *csAttempt) sendMsg(m any, hdr []byte, payld mem.BufferSlice, dataLength, payloadLength int) error { cs := a.cs if a.trInfo != nil { a.mu.Lock() @@ -1052,8 +1092,10 @@ func (a *csAttempt) sendMsg(m any, hdr, payld, data []byte) error { } return io.EOF } - for _, sh := range a.statsHandlers { - sh.HandleRPC(a.ctx, outPayload(true, m, data, payld, time.Now())) + if len(a.statsHandlers) != 0 { + for _, sh := range a.statsHandlers { + sh.HandleRPC(a.ctx, outPayload(true, m, dataLength, payloadLength, time.Now())) + } } if channelz.IsOn() { a.t.IncrMsgSent() @@ -1065,6 +1107,7 @@ func (a *csAttempt) recvMsg(m any, payInfo *payloadInfo) (err error) { cs := a.cs if len(a.statsHandlers) != 0 && payInfo == nil { payInfo = &payloadInfo{} + defer payInfo.free() } if !a.decompSet { @@ -1083,8 +1126,7 @@ func (a *csAttempt) recvMsg(m any, payInfo *payloadInfo) (err error) { // Only initialize this state once per stream. a.decompSet = true } - err = recv(a.p, cs.codec, a.s, a.dc, m, *cs.callInfo.maxReceiveMessageSize, payInfo, a.decomp) - if err != nil { + if err := recv(a.p, cs.codec, a.s, a.dc, m, *cs.callInfo.maxReceiveMessageSize, payInfo, a.decomp, false); err != nil { if err == io.EOF { if statusErr := a.s.Status().Err(); statusErr != nil { return statusErr @@ -1103,14 +1145,12 @@ func (a *csAttempt) recvMsg(m any, payInfo *payloadInfo) (err error) { } for _, sh := range a.statsHandlers { sh.HandleRPC(a.ctx, &stats.InPayload{ - Client: true, - RecvTime: time.Now(), - Payload: m, - // TODO truncate large payload. - Data: payInfo.uncompressedBytes, + Client: true, + RecvTime: time.Now(), + Payload: m, WireLength: payInfo.compressedLength + headerLen, CompressedLength: payInfo.compressedLength, - Length: len(payInfo.uncompressedBytes), + Length: payInfo.uncompressedBytes.Len(), }) } if channelz.IsOn() { @@ -1122,14 +1162,12 @@ func (a *csAttempt) recvMsg(m any, payInfo *payloadInfo) (err error) { } // Special handling for non-server-stream rpcs. // This recv expects EOF or errors, so we don't collect inPayload. - err = recv(a.p, cs.codec, a.s, a.dc, m, *cs.callInfo.maxReceiveMessageSize, nil, a.decomp) - if err == nil { - return toRPCErr(errors.New("grpc: client streaming protocol violation: get , want ")) - } - if err == io.EOF { + if err := recv(a.p, cs.codec, a.s, a.dc, m, *cs.callInfo.maxReceiveMessageSize, nil, a.decomp, false); err == io.EOF { return a.s.Status().Err() // non-server streaming Recv returns nil on success + } else if err != nil { + return toRPCErr(err) } - return toRPCErr(err) + return toRPCErr(errors.New("grpc: client streaming protocol violation: get , want ")) } func (a *csAttempt) finish(err error) { @@ -1185,12 +1223,12 @@ func (a *csAttempt) finish(err error) { a.mu.Unlock() } -// newClientStream creates a ClientStream with the specified transport, on the +// newNonRetryClientStream creates a ClientStream with the specified transport, on the // given addrConn. // // It's expected that the given transport is either the same one in addrConn, or // is already closed. To avoid race, transport is specified separately, instead -// of using ac.transpot. +// of using ac.transport. // // Main difference between this and ClientConn.NewStream: // - no retry @@ -1276,7 +1314,7 @@ func newNonRetryClientStream(ctx context.Context, desc *StreamDesc, method strin return nil, err } as.s = s - as.p = &parser{r: s, recvBufferPool: ac.dopts.recvBufferPool} + as.p = &parser{r: s, bufferPool: ac.dopts.copts.BufferPool} ac.incrCallsStarted() if desc != unaryStreamDesc { // Listen on stream context to cleanup when the stream context is @@ -1373,17 +1411,26 @@ func (as *addrConnStream) SendMsg(m any) (err error) { } // load hdr, payload, data - hdr, payld, _, err := prepareMsg(m, as.codec, as.cp, as.comp) + hdr, data, payload, pf, err := prepareMsg(m, as.codec, as.cp, as.comp, as.ac.dopts.copts.BufferPool) if err != nil { return err } + defer func() { + data.Free() + // only free payload if compression was made, and therefore it is a different set + // of buffers from data. + if pf.isCompressed() { + payload.Free() + } + }() + // TODO(dfawley): should we be checking len(data) instead? - if len(payld) > *as.callInfo.maxSendMessageSize { - return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", len(payld), *as.callInfo.maxSendMessageSize) + if payload.Len() > *as.callInfo.maxSendMessageSize { + return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", payload.Len(), *as.callInfo.maxSendMessageSize) } - if err := as.t.Write(as.s, hdr, payld, &transport.Options{Last: !as.desc.ClientStreams}); err != nil { + if err := as.t.Write(as.s, hdr, payload, &transport.Options{Last: !as.desc.ClientStreams}); err != nil { if !as.desc.ClientStreams { // For non-client-streaming RPCs, we return nil instead of EOF on error // because the generated code requires it. finish is not called; RecvMsg() @@ -1423,8 +1470,7 @@ func (as *addrConnStream) RecvMsg(m any) (err error) { // Only initialize this state once per stream. as.decompSet = true } - err = recv(as.p, as.codec, as.s, as.dc, m, *as.callInfo.maxReceiveMessageSize, nil, as.decomp) - if err != nil { + if err := recv(as.p, as.codec, as.s, as.dc, m, *as.callInfo.maxReceiveMessageSize, nil, as.decomp, false); err != nil { if err == io.EOF { if statusErr := as.s.Status().Err(); statusErr != nil { return statusErr @@ -1444,14 +1490,12 @@ func (as *addrConnStream) RecvMsg(m any) (err error) { // Special handling for non-server-stream rpcs. // This recv expects EOF or errors, so we don't collect inPayload. - err = recv(as.p, as.codec, as.s, as.dc, m, *as.callInfo.maxReceiveMessageSize, nil, as.decomp) - if err == nil { - return toRPCErr(errors.New("grpc: client streaming protocol violation: get , want ")) - } - if err == io.EOF { + if err := recv(as.p, as.codec, as.s, as.dc, m, *as.callInfo.maxReceiveMessageSize, nil, as.decomp, false); err == io.EOF { return as.s.Status().Err() // non-server streaming Recv returns nil on success + } else if err != nil { + return toRPCErr(err) } - return toRPCErr(err) + return toRPCErr(errors.New("grpc: client streaming protocol violation: get , want ")) } func (as *addrConnStream) finish(err error) { @@ -1645,18 +1689,31 @@ func (ss *serverStream) SendMsg(m any) (err error) { } // load hdr, payload, data - hdr, payload, data, err := prepareMsg(m, ss.codec, ss.cp, ss.comp) + hdr, data, payload, pf, err := prepareMsg(m, ss.codec, ss.cp, ss.comp, ss.p.bufferPool) if err != nil { return err } + defer func() { + data.Free() + // only free payload if compression was made, and therefore it is a different set + // of buffers from data. + if pf.isCompressed() { + payload.Free() + } + }() + + dataLen := data.Len() + payloadLen := payload.Len() + // TODO(dfawley): should we be checking len(data) instead? - if len(payload) > ss.maxSendMessageSize { - return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", len(payload), ss.maxSendMessageSize) + if payloadLen > ss.maxSendMessageSize { + return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", payloadLen, ss.maxSendMessageSize) } if err := ss.t.Write(ss.s, hdr, payload, &transport.Options{Last: false}); err != nil { return toRPCErr(err) } + if len(ss.binlogs) != 0 { if !ss.serverHeaderBinlogged { h, _ := ss.s.Header() @@ -1669,7 +1726,7 @@ func (ss *serverStream) SendMsg(m any) (err error) { } } sm := &binarylog.ServerMessage{ - Message: data, + Message: data.Materialize(), } for _, binlog := range ss.binlogs { binlog.Log(ss.ctx, sm) @@ -1677,7 +1734,7 @@ func (ss *serverStream) SendMsg(m any) (err error) { } if len(ss.statsHandler) != 0 { for _, sh := range ss.statsHandler { - sh.HandleRPC(ss.s.Context(), outPayload(false, m, data, payload, time.Now())) + sh.HandleRPC(ss.s.Context(), outPayload(false, m, dataLen, payloadLen, time.Now())) } } return nil @@ -1714,8 +1771,9 @@ func (ss *serverStream) RecvMsg(m any) (err error) { var payInfo *payloadInfo if len(ss.statsHandler) != 0 || len(ss.binlogs) != 0 { payInfo = &payloadInfo{} + defer payInfo.free() } - if err := recv(ss.p, ss.codec, ss.s, ss.dc, m, ss.maxReceiveMessageSize, payInfo, ss.decomp); err != nil { + if err := recv(ss.p, ss.codec, ss.s, ss.dc, m, ss.maxReceiveMessageSize, payInfo, ss.decomp, true); err != nil { if err == io.EOF { if len(ss.binlogs) != 0 { chc := &binarylog.ClientHalfClose{} @@ -1733,11 +1791,9 @@ func (ss *serverStream) RecvMsg(m any) (err error) { if len(ss.statsHandler) != 0 { for _, sh := range ss.statsHandler { sh.HandleRPC(ss.s.Context(), &stats.InPayload{ - RecvTime: time.Now(), - Payload: m, - // TODO truncate large payload. - Data: payInfo.uncompressedBytes, - Length: len(payInfo.uncompressedBytes), + RecvTime: time.Now(), + Payload: m, + Length: payInfo.uncompressedBytes.Len(), WireLength: payInfo.compressedLength + headerLen, CompressedLength: payInfo.compressedLength, }) @@ -1745,7 +1801,7 @@ func (ss *serverStream) RecvMsg(m any) (err error) { } if len(ss.binlogs) != 0 { cm := &binarylog.ClientMessage{ - Message: payInfo.uncompressedBytes, + Message: payInfo.uncompressedBytes.Materialize(), } for _, binlog := range ss.binlogs { binlog.Log(ss.ctx, cm) @@ -1760,23 +1816,26 @@ func MethodFromServerStream(stream ServerStream) (string, bool) { return Method(stream.Context()) } -// prepareMsg returns the hdr, payload and data -// using the compressors passed or using the -// passed preparedmsg -func prepareMsg(m any, codec baseCodec, cp Compressor, comp encoding.Compressor) (hdr, payload, data []byte, err error) { +// prepareMsg returns the hdr, payload and data using the compressors passed or +// using the passed preparedmsg. The returned boolean indicates whether +// compression was made and therefore whether the payload needs to be freed in +// addition to the returned data. Freeing the payload if the returned boolean is +// false can lead to undefined behavior. +func prepareMsg(m any, codec baseCodec, cp Compressor, comp encoding.Compressor, pool mem.BufferPool) (hdr []byte, data, payload mem.BufferSlice, pf payloadFormat, err error) { if preparedMsg, ok := m.(*PreparedMsg); ok { - return preparedMsg.hdr, preparedMsg.payload, preparedMsg.encodedData, nil + return preparedMsg.hdr, preparedMsg.encodedData, preparedMsg.payload, preparedMsg.pf, nil } // The input interface is not a prepared msg. // Marshal and Compress the data at this point data, err = encode(codec, m) if err != nil { - return nil, nil, nil, err + return nil, nil, nil, 0, err } - compData, err := compress(data, cp, comp) + compData, pf, err := compress(data, cp, comp, pool) if err != nil { - return nil, nil, nil, err + data.Free() + return nil, nil, nil, 0, err } - hdr, payload = msgHeader(data, compData) - return hdr, payload, data, nil + hdr, payload = msgHeader(data, compData, pf) + return hdr, data, payload, pf, nil } diff --git a/vendor/google.golang.org/grpc/stream_interfaces.go b/vendor/google.golang.org/grpc/stream_interfaces.go index 8b813529c0cc5..0037fee0bd71a 100644 --- a/vendor/google.golang.org/grpc/stream_interfaces.go +++ b/vendor/google.golang.org/grpc/stream_interfaces.go @@ -22,15 +22,35 @@ package grpc // request, many responses) RPC. It is generic over the type of the response // message. It is used in generated code. type ServerStreamingClient[Res any] interface { + // Recv receives the next response message from the server. The client may + // repeatedly call Recv to read messages from the response stream. If + // io.EOF is returned, the stream has terminated with an OK status. Any + // other error is compatible with the status package and indicates the + // RPC's status code and message. Recv() (*Res, error) + + // ClientStream is embedded to provide Context, Header, and Trailer + // functionality. No other methods in the ClientStream should be called + // directly. ClientStream } // ServerStreamingServer represents the server side of a server-streaming (one // request, many responses) RPC. It is generic over the type of the response // message. It is used in generated code. +// +// To terminate the response stream, return from the handler method and return +// an error from the status package, or use nil to indicate an OK status code. type ServerStreamingServer[Res any] interface { + // Send sends a response message to the client. The server handler may + // call Send multiple times to send multiple messages to the client. An + // error is returned if the stream was terminated unexpectedly, and the + // handler method should return, as the stream is no longer usable. Send(*Res) error + + // ServerStream is embedded to provide Context, SetHeader, SendHeader, and + // SetTrailer functionality. No other methods in the ServerStream should + // be called directly. ServerStream } @@ -39,8 +59,22 @@ type ServerStreamingServer[Res any] interface { // message stream and the type of the unary response message. It is used in // generated code. type ClientStreamingClient[Req any, Res any] interface { + // Send sends a request message to the server. The client may call Send + // multiple times to send multiple messages to the server. On error, Send + // aborts the stream. If the error was generated by the client, the status + // is returned directly. Otherwise, io.EOF is returned, and the status of + // the stream may be discovered using CloseAndRecv(). Send(*Req) error + + // CloseAndRecv closes the request stream and waits for the server's + // response. This method must be called once and only once after sending + // all request messages. Any error returned is implemented by the status + // package. CloseAndRecv() (*Res, error) + + // ClientStream is embedded to provide Context, Header, and Trailer + // functionality. No other methods in the ClientStream should be called + // directly. ClientStream } @@ -48,9 +82,28 @@ type ClientStreamingClient[Req any, Res any] interface { // requests, one response) RPC. It is generic over both the type of the request // message stream and the type of the unary response message. It is used in // generated code. +// +// To terminate the RPC, call SendAndClose and return nil from the method +// handler or do not call SendAndClose and return an error from the status +// package. type ClientStreamingServer[Req any, Res any] interface { + // Recv receives the next request message from the client. The server may + // repeatedly call Recv to read messages from the request stream. If + // io.EOF is returned, it indicates the client called CloseAndRecv on its + // ClientStreamingClient. Any other error indicates the stream was + // terminated unexpectedly, and the handler method should return, as the + // stream is no longer usable. Recv() (*Req, error) + + // SendAndClose sends a single response message to the client and closes + // the stream. This method must be called once and only once after all + // request messages have been processed. Recv should not be called after + // calling SendAndClose. SendAndClose(*Res) error + + // ServerStream is embedded to provide Context, SetHeader, SendHeader, and + // SetTrailer functionality. No other methods in the ServerStream should + // be called directly. ServerStream } @@ -59,8 +112,23 @@ type ClientStreamingServer[Req any, Res any] interface { // request message stream and the type of the response message stream. It is // used in generated code. type BidiStreamingClient[Req any, Res any] interface { + // Send sends a request message to the server. The client may call Send + // multiple times to send multiple messages to the server. On error, Send + // aborts the stream. If the error was generated by the client, the status + // is returned directly. Otherwise, io.EOF is returned, and the status of + // the stream may be discovered using Recv(). Send(*Req) error + + // Recv receives the next response message from the server. The client may + // repeatedly call Recv to read messages from the response stream. If + // io.EOF is returned, the stream has terminated with an OK status. Any + // other error is compatible with the status package and indicates the + // RPC's status code and message. Recv() (*Res, error) + + // ClientStream is embedded to provide Context, Header, Trailer, and + // CloseSend functionality. No other methods in the ClientStream should be + // called directly. ClientStream } @@ -68,9 +136,27 @@ type BidiStreamingClient[Req any, Res any] interface { // (many requests, many responses) RPC. It is generic over both the type of the // request message stream and the type of the response message stream. It is // used in generated code. +// +// To terminate the stream, return from the handler method and return +// an error from the status package, or use nil to indicate an OK status code. type BidiStreamingServer[Req any, Res any] interface { + // Recv receives the next request message from the client. The server may + // repeatedly call Recv to read messages from the request stream. If + // io.EOF is returned, it indicates the client called CloseSend on its + // BidiStreamingClient. Any other error indicates the stream was + // terminated unexpectedly, and the handler method should return, as the + // stream is no longer usable. Recv() (*Req, error) + + // Send sends a response message to the client. The server handler may + // call Send multiple times to send multiple messages to the client. An + // error is returned if the stream was terminated unexpectedly, and the + // handler method should return, as the stream is no longer usable. Send(*Res) error + + // ServerStream is embedded to provide Context, SetHeader, SendHeader, and + // SetTrailer functionality. No other methods in the ServerStream should + // be called directly. ServerStream } diff --git a/vendor/google.golang.org/grpc/version.go b/vendor/google.golang.org/grpc/version.go index bafaef99be989..187fbf1195141 100644 --- a/vendor/google.golang.org/grpc/version.go +++ b/vendor/google.golang.org/grpc/version.go @@ -19,4 +19,4 @@ package grpc // Version is the current grpc version. -const Version = "1.65.0" +const Version = "1.67.0" diff --git a/vendor/modules.txt b/vendor/modules.txt index 0c46312e7a9b1..0387d146e09ac 100644 --- a/vendor/modules.txt +++ b/vendor/modules.txt @@ -346,8 +346,8 @@ github.com/google/go-cmp/cmp/internal/value ## explicit; go 1.12 github.com/google/gofuzz github.com/google/gofuzz/bytesource -# github.com/google/pprof v0.0.0-20240727154555-813a5fbdbec8 -## explicit; go 1.19 +# github.com/google/pprof v0.0.0-20240827171923-fa2c70bbbfe5 +## explicit; go 1.22 github.com/google/pprof/profile # github.com/google/shlex v0.0.0-20191202100458-e7afc7fbc510 ## explicit; go 1.13 @@ -463,7 +463,7 @@ github.com/munnerz/goautoneg # github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f ## explicit github.com/mxk/go-flowrate/flowrate -# github.com/onsi/ginkgo/v2 v2.17.2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 +# github.com/onsi/ginkgo/v2 v2.20.2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44 ## explicit; go 1.20 github.com/onsi/ginkgo/v2 github.com/onsi/ginkgo/v2/config @@ -540,6 +540,19 @@ github.com/opencontainers/runtime-spec/specs-go github.com/opencontainers/selinux/go-selinux github.com/opencontainers/selinux/go-selinux/label github.com/opencontainers/selinux/pkg/pwalkdir +# github.com/openshift-eng/openshift-tests-extension v0.0.0-20241028125658-76ee2532d0bf +## explicit; go 1.22.0 +github.com/openshift-eng/openshift-tests-extension/pkg/cmd +github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdinfo +github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdlist +github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdrun +github.com/openshift-eng/openshift-tests-extension/pkg/cmd/cmdupdate +github.com/openshift-eng/openshift-tests-extension/pkg/dbtime +github.com/openshift-eng/openshift-tests-extension/pkg/extension +github.com/openshift-eng/openshift-tests-extension/pkg/extension/extensiontests +github.com/openshift-eng/openshift-tests-extension/pkg/flags +github.com/openshift-eng/openshift-tests-extension/pkg/ginkgo +github.com/openshift-eng/openshift-tests-extension/pkg/version # github.com/openshift/api v0.0.0-20241001152557-e415140e5d5f ## explicit; go 1.22.0 github.com/openshift/api/apiserver/v1 @@ -1083,7 +1096,7 @@ golang.org/x/crypto/ssh/internal/bcrypt_pbkdf ## explicit; go 1.20 golang.org/x/exp/constraints golang.org/x/exp/slices -# golang.org/x/mod v0.17.0 +# golang.org/x/mod v0.20.0 ## explicit; go 1.18 golang.org/x/mod/internal/lazyregexp golang.org/x/mod/module @@ -1104,7 +1117,7 @@ golang.org/x/net/internal/timeseries golang.org/x/net/proxy golang.org/x/net/trace golang.org/x/net/websocket -# golang.org/x/oauth2 v0.21.0 +# golang.org/x/oauth2 v0.22.0 ## explicit; go 1.18 golang.org/x/oauth2 golang.org/x/oauth2/internal @@ -1159,7 +1172,7 @@ golang.org/x/text/width # golang.org/x/time v0.3.0 ## explicit golang.org/x/time/rate -# golang.org/x/tools v0.21.1-0.20240508182429-e35e4ccd0d2d +# golang.org/x/tools v0.24.0 ## explicit; go 1.19 golang.org/x/tools/benchmark/parse golang.org/x/tools/cmd/stringer @@ -1168,7 +1181,6 @@ golang.org/x/tools/cover golang.org/x/tools/go/ast/astutil golang.org/x/tools/go/ast/inspector golang.org/x/tools/go/gcexportdata -golang.org/x/tools/go/internal/packagesdriver golang.org/x/tools/go/packages golang.org/x/tools/go/types/objectpath golang.org/x/tools/imports @@ -1190,17 +1202,17 @@ golang.org/x/tools/internal/versions # google.golang.org/genproto v0.0.0-20230822172742-b8732ec3820d ## explicit; go 1.19 google.golang.org/genproto/protobuf/field_mask -# google.golang.org/genproto/googleapis/api v0.0.0-20240528184218-531527333157 -## explicit; go 1.20 +# google.golang.org/genproto/googleapis/api v0.0.0-20240930140551-af27646dc61f +## explicit; go 1.21 google.golang.org/genproto/googleapis/api google.golang.org/genproto/googleapis/api/annotations google.golang.org/genproto/googleapis/api/expr/v1alpha1 google.golang.org/genproto/googleapis/api/httpbody -# google.golang.org/genproto/googleapis/rpc v0.0.0-20240701130421-f6361c86f094 -## explicit; go 1.20 +# google.golang.org/genproto/googleapis/rpc v0.0.0-20240924160255-9d4c2d233b61 +## explicit; go 1.21 google.golang.org/genproto/googleapis/rpc/errdetails google.golang.org/genproto/googleapis/rpc/status -# google.golang.org/grpc v1.65.0 +# google.golang.org/grpc v1.67.0 ## explicit; go 1.21 google.golang.org/grpc google.golang.org/grpc/attributes @@ -1219,7 +1231,9 @@ google.golang.org/grpc/credentials/insecure google.golang.org/grpc/encoding google.golang.org/grpc/encoding/gzip google.golang.org/grpc/encoding/proto +google.golang.org/grpc/experimental/stats google.golang.org/grpc/grpclog +google.golang.org/grpc/grpclog/internal google.golang.org/grpc/health google.golang.org/grpc/health/grpc_health_v1 google.golang.org/grpc/internal @@ -1243,11 +1257,13 @@ google.golang.org/grpc/internal/resolver/dns/internal google.golang.org/grpc/internal/resolver/passthrough google.golang.org/grpc/internal/resolver/unix google.golang.org/grpc/internal/serviceconfig +google.golang.org/grpc/internal/stats google.golang.org/grpc/internal/status google.golang.org/grpc/internal/syscall google.golang.org/grpc/internal/transport google.golang.org/grpc/internal/transport/networktype google.golang.org/grpc/keepalive +google.golang.org/grpc/mem google.golang.org/grpc/metadata google.golang.org/grpc/peer google.golang.org/grpc/resolver @@ -1569,4 +1585,4 @@ sigs.k8s.io/yaml sigs.k8s.io/yaml/goyaml.v2 sigs.k8s.io/yaml/goyaml.v3 # github.com/google/cadvisor => github.com/openshift/google-cadvisor v0.49.0-openshift-4.17-2 -# github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20240806135314-3946b2b7b2a8 +# github.com/onsi/ginkgo/v2 => github.com/openshift/onsi-ginkgo/v2 v2.6.1-0.20241008152707-25bf9f14db44