--Task-- name: Deploy_all_forgerock_components enabled: True class_name: DeployComponentsTask source_name: controller source_namespace: >default< target_name: controller target_namespace: >default< start: 0 stop: None timeout: no timeout loop: False interval: None dependencies: ['Enable_prometheus_admin_api'] wait_for: [] options: {} group_name: None Current dir: /mnt/disks/data/xslou/lodestar-fork/pyrock ________________________________________________________________________________ [28/May/2023 05:49:19] Deploy_all_forgerock_components pre : Initialising task parameters ________________________________________________________________________________ task will be executed on controller (localhost) ________________________________________________________________________________ [28/May/2023 05:49:19] Deploy_all_forgerock_components step1 : Deploy components ________________________________________________________________________________ ******************************** Cleaning up existing namespace ******************************** ------------------ Deleting secret agent controller ------------------ [loop_until]: kubectl --namespace=xlou delete sac --all [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- secretagentconfiguration.secret-agent.secrets.forgerock.io "forgerock-sac" deleted --- stderr --- ----------------------- Deleting all resources ----------------------- [loop_until]: kubectl --namespace=xlou delete all --all --grace-period=0 --force [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- pod "admin-ui-7c7859f797-c85wt" force deleted pod "am-7849cf7bdb-c97w4" force deleted pod "am-7849cf7bdb-xrf2f" force deleted pod "am-7849cf7bdb-zjgd9" force deleted pod "amster-j57d2" force deleted pod "ds-cts-0" force deleted pod "ds-cts-1" force deleted pod "ds-cts-2" force deleted pod "ds-idrepo-0" force deleted pod "ds-idrepo-1" force deleted pod "ds-idrepo-2" force deleted pod "end-user-ui-fdf99b7d7-mxgbj" force deleted pod "idm-58dc667486-hst96" force deleted pod "idm-58dc667486-rprmj" force deleted pod "ldif-importer-cjxt5" force deleted pod "lodemon-76d7c4b79c-mvf4c" force deleted pod "login-ui-7fdfb4d89f-5ptbp" force deleted pod "overseer-0-79894fbbd7-5b74c" force deleted service "admin-ui" force deleted service "am" force deleted service "ds-cts" force deleted service "ds-idrepo" force deleted service "end-user-ui" force deleted service "idm" force deleted service "login-ui" force deleted service "overseer-0" force deleted deployment.apps "admin-ui" force deleted deployment.apps "am" force deleted deployment.apps "end-user-ui" force deleted deployment.apps "idm" force deleted deployment.apps "lodemon" force deleted deployment.apps "login-ui" force deleted deployment.apps "overseer-0" force deleted statefulset.apps "ds-cts" force deleted statefulset.apps "ds-idrepo" force deleted job.batch "amster" force deleted job.batch "ldif-importer" force deleted --- stderr --- warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. [loop_until]: kubectl --namespace xlou get pods | grep "No resources found" [loop_until]: (max_time=360, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 10s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 21s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 31s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 42s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 52s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 1m 03s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 1m 13s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 1m 24s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 1m 34s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 1m 44s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 1m 55s (rc=0) - failed to find expected output: No resources found - retry [loop_until]: Function succeeded after 2m 05s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- No resources found in xlou namespace. ------------------------- Deleting configmap ------------------------- [loop_until]: kubectl --namespace=xlou get configmap --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- amster-files amster-retain dev-utils idm idm-logging-properties kube-root-ca.crt lodemon-config overseer-config-0 platform-config --- stderr --- [loop_until]: kubectl --namespace=xlou delete configmap amster-files --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- configmap "amster-files" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete configmap amster-retain --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- configmap "amster-retain" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete configmap dev-utils --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- configmap "dev-utils" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete configmap idm --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- configmap "idm" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete configmap idm-logging-properties --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- configmap "idm-logging-properties" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete configmap kube-root-ca.crt --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- configmap "kube-root-ca.crt" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete configmap lodemon-config --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- configmap "lodemon-config" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete configmap overseer-config-0 --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- configmap "overseer-config-0" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete configmap platform-config --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- configmap "platform-config" deleted --- stderr --- --------------------------- Deleting secret --------------------------- [loop_until]: kubectl --namespace=xlou get secret --output jsonpath='{.items[?(@.type=="Opaque")].metadata.name}' [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- cloud-storage-credentials-cts cloud-storage-credentials-idrepo --- stderr --- [loop_until]: kubectl --namespace=xlou delete secret cloud-storage-credentials-cts --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- secret "cloud-storage-credentials-cts" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete secret cloud-storage-credentials-idrepo --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- secret "cloud-storage-credentials-idrepo" deleted --- stderr --- -------------------------- Deleting ingress -------------------------- [loop_until]: kubectl --namespace=xlou get ingress --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- forgerock ig overseer-0 --- stderr --- [loop_until]: kubectl --namespace=xlou delete ingress forgerock --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- ingress.networking.k8s.io "forgerock" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete ingress ig --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- ingress.networking.k8s.io "ig" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete ingress overseer-0 --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- ingress.networking.k8s.io "overseer-0" deleted --- stderr --- ---------------------------- Deleting pvc ---------------------------- [loop_until]: kubectl --namespace=xlou get pvc --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- data-ds-cts-0 data-ds-cts-1 data-ds-cts-2 data-ds-idrepo-0 data-ds-idrepo-1 data-ds-idrepo-2 overseer-0 --- stderr --- [loop_until]: kubectl --namespace=xlou delete pvc data-ds-cts-0 --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- persistentvolumeclaim "data-ds-cts-0" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete pvc data-ds-cts-1 --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- persistentvolumeclaim "data-ds-cts-1" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete pvc data-ds-cts-2 --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- persistentvolumeclaim "data-ds-cts-2" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete pvc data-ds-idrepo-0 --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- persistentvolumeclaim "data-ds-idrepo-0" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete pvc data-ds-idrepo-1 --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- persistentvolumeclaim "data-ds-idrepo-1" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete pvc data-ds-idrepo-2 --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- persistentvolumeclaim "data-ds-idrepo-2" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete pvc overseer-0 --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- persistentvolumeclaim "overseer-0" deleted --- stderr --- [loop_until]: kubectl --namespace=xlou delete pv ds-backup-xlou --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- warning: deleting cluster-scoped resources, not scoped to the provided namespace ----------------- Deleting admin clusterrolebindings ----------------- [loop_until]: kubectl get clusterrolebinding --output jsonpath="{range .items[?(@.subjects[0].namespace=='xlou')]}{.metadata.name} {end}" [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- k8s-svc-acct-crb-xlou k8s-svc-acct-crb-xlou-0 --- stderr --- Deleting clusterrolebinding k8s-svc-acct-crb-xlou associated with xlou namespace [loop_until]: kubectl delete clusterrolebinding k8s-svc-acct-crb-xlou [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- clusterrolebinding.rbac.authorization.k8s.io "k8s-svc-acct-crb-xlou" deleted --- stderr --- Deleting clusterrolebinding k8s-svc-acct-crb-xlou-0 associated with xlou namespace [loop_until]: kubectl delete clusterrolebinding k8s-svc-acct-crb-xlou-0 [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- clusterrolebinding.rbac.authorization.k8s.io "k8s-svc-acct-crb-xlou-0" deleted --- stderr --- ------------------------- Deleting namespace ------------------------- [loop_until]: kubectl delete namespaces xlou --ignore-not-found --grace-period=0 --force [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- namespace "xlou" force deleted --- stderr --- warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. [loop_until]: awk -F" " "{print NF}" <<< `kubectl get namespace xlou --ignore-not-found` | grep 0 [loop_until]: (max_time=600, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- ************************************* Creating deployment ************************************* Creating normal (forgeops) type deployment for deployment: stack ------- Custom component configuration present. Loading values ------- ------------------ Deleting secret agent controller ------------------ [loop_until]: kubectl --namespace=xlou delete sac --all [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- No resources found --- stderr --- ----------------------- Deleting all resources ----------------------- [loop_until]: kubectl --namespace=xlou delete all --all --grace-period=0 --force [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- No resources found --- stderr --- warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. [loop_until]: kubectl --namespace xlou get pods | grep "No resources found" [loop_until]: (max_time=360, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- No resources found in xlou namespace. ------------------------- Deleting configmap ------------------------- [loop_until]: kubectl --namespace=xlou get configmap --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- --------------------------- Deleting secret --------------------------- [loop_until]: kubectl --namespace=xlou get secret --output jsonpath='{.items[?(@.type=="Opaque")].metadata.name}' [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- -------------------------- Deleting ingress -------------------------- [loop_until]: kubectl --namespace=xlou get ingress --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- ---------------------------- Deleting pvc ---------------------------- [loop_until]: kubectl --namespace=xlou get pvc --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- [loop_until]: kubectl --namespace=xlou delete pv ds-backup-xlou --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- warning: deleting cluster-scoped resources, not scoped to the provided namespace ----------------- Deleting admin clusterrolebindings ----------------- [loop_until]: kubectl get clusterrolebinding --output jsonpath="{range .items[?(@.subjects[0].namespace=='xlou')]}{.metadata.name} {end}" [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- ------------------------- Deleting namespace ------------------------- [loop_until]: kubectl delete namespaces xlou --ignore-not-found --grace-period=0 --force [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. [loop_until]: awk -F" " "{print NF}" <<< `kubectl get namespace xlou --ignore-not-found` | grep 0 [loop_until]: (max_time=600, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- [loop_until]: kubectl create namespace xlou [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- namespace/xlou created --- stderr --- [loop_until]: kubectl label namespace xlou self-service=false timeout=48 [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- namespace/xlou labeled --- stderr --- ************************************ Configuring components ************************************ Applying custom configuration, dockerfiles to deployment and custom lodestar component configuration Copying /mnt/disks/data/xslou/lodestar-fork/shared/config/custom/simplemanageduser/docker/idm to /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm No custom features provided. Nothing to do. ---- Updating components image tag/repo from platform-images repo ---- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --clean [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Cleaning up. [WARNING] Found nothing to clean. --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref master-ready-for-dev-pipelines --products ds [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Repo is at 72bd4e063f0fd5698b05b974fdd8664f01d2f11b on branch HEAD [INFO] Updating products ds [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/dsutil/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/ds-new/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/proxy/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref master-ready-for-dev-pipelines --products am [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Found existing files, attempting to not clone. [INFO] Repo is at 72bd4e063f0fd5698b05b974fdd8664f01d2f11b on branch HEAD [INFO] Updating products am [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am/Dockerfile to use gcr.io/forgerock-io/am-cdk/pit1:7.4.0-043c211b139ffe96bb9eb3a723a409253d1abf1c --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref master-ready-for-dev-pipelines --products amster [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Found existing files, attempting to not clone. [INFO] Repo is at 72bd4e063f0fd5698b05b974fdd8664f01d2f11b on branch HEAD [INFO] Updating products amster [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster/Dockerfile to use gcr.io/forgerock-io/amster/pit1:7.4.0-043c211b139ffe96bb9eb3a723a409253d1abf1c --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref master-ready-for-dev-pipelines --products idm [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Found existing files, attempting to not clone. [INFO] Repo is at 72bd4e063f0fd5698b05b974fdd8664f01d2f11b on branch HEAD [INFO] Updating products idm [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm/Dockerfile to use gcr.io/forgerock-io/idm-cdk/pit1:7.4.0-292ff61a9416bbbef0d5118acbe9ed7d498d0f1a --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref master-ready-for-dev-pipelines --products ui [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Found existing files, attempting to not clone. [INFO] Repo is at 72bd4e063f0fd5698b05b974fdd8664f01d2f11b on branch HEAD [INFO] Updating products ui [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui to use gcr.io/forgerock-io/platform-admin-ui/docker-build:7.3.0-cc26db93fcf78082e97775a28257668396d32787 [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui to use gcr.io/forgerock-io/platform-enduser-ui/docker-build:7.3.0-cc26db93fcf78082e97775a28257668396d32787 [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui to use gcr.io/forgerock-io/platform-login-ui/docker-build:7.3.0-cc26db93fcf78082e97775a28257668396d32787 --- stderr --- - Checking if component Dockerfile/kustomize needs additional update - [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker ds cts [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts --- stderr --- Checking if ds-cts dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts/Dockerfile: gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker ds idrepo [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo --- stderr --- Checking if ds-idrepo dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo/Dockerfile: gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker am [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am --- stderr --- Checking if am dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am/Dockerfile: gcr.io/forgerock-io/am-cdk/pit1:7.4.0-043c211b139ffe96bb9eb3a723a409253d1abf1c No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker amster [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster --- stderr --- Checking if amster dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster/Dockerfile: gcr.io/forgerock-io/amster/pit1:7.4.0-043c211b139ffe96bb9eb3a723a409253d1abf1c No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker idm [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm --- stderr --- Checking if idm dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm/Dockerfile: gcr.io/forgerock-io/idm-cdk/pit1:7.4.0-292ff61a9416bbbef0d5118acbe9ed7d498d0f1a No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path kustomize base end-user-ui [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui --- stderr --- Checking if end-user-ui kustomize file /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui/kustomization.yaml needs additional update (custom tag/repo from config.yaml or image name resolution) Read image line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui/kustomization.yaml: gcr.io/forgerock-io/platform-enduser-ui/docker-build:7.3.0-cc26db93fcf78082e97775a28257668396d32787 No need to update the image line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui/kustomization.yaml [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path kustomize base login-ui [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui --- stderr --- Checking if login-ui kustomize file /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui/kustomization.yaml needs additional update (custom tag/repo from config.yaml or image name resolution) Read image line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui/kustomization.yaml: gcr.io/forgerock-io/platform-login-ui/docker-build:7.3.0-cc26db93fcf78082e97775a28257668396d32787 No need to update the image line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui/kustomization.yaml [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path kustomize base admin-ui [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui --- stderr --- Checking if admin-ui kustomize file /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui/kustomization.yaml needs additional update (custom tag/repo from config.yaml or image name resolution) Read image line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui/kustomization.yaml: gcr.io/forgerock-io/platform-admin-ui/docker-build:7.3.0-cc26db93fcf78082e97775a28257668396d32787 No need to update the image line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui/kustomization.yaml [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path kustomize overlay medium [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/overlay/medium --- stderr --- [loop_until]: kubectl --namespace=xlou delete --filename /tmp/tmp2kq272r9 [loop_until]: (max_time=180, interval=5, expected_rc=[0, 1] [loop_until]: OK (rc = 1) --- stdout --- --- stderr --- Error from server (NotFound): error when deleting "/tmp/tmp2kq272r9": secrets "sslcert" not found [loop_until]: kubectl --namespace=xlou apply --filename /tmp/tmp2kq272r9 [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- secret/sslcert created --- stderr --- The following components will be deployed: - ds-cts (DS) - ds-idrepo (DS) - am (AM) - amster (Amster) - idm (IDM) - end-user-ui (EndUserUi) - login-ui (LoginUi) - admin-ui (AdminUi) [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/forgeops build all --config-profile=cdk --push-to gcr.io/engineeringpit/lodestar-images --tag=xlou [run_command]: env={'HOME': '/home/xslou', 'PATH': '/mnt/disks/data/xslou/lodestar-fork/ext/bin:/home/xslou/.local/bin:/home/xslou/bin:~/bin:~/workshop/lodestar-fork/ext/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin', 'DOCKER_SCAN_SUGGEST': 'false'} The push refers to repository [gcr.io/engineeringpit/lodestar-images/am] 5f70bf18a086: Preparing c09101d9181a: Preparing cb324cf89811: Preparing 5f70bf18a086: Preparing 11ac03f3a24b: Preparing 67c330ec8773: Preparing 563f13631928: Preparing 5cc2fef50bf6: Preparing b805253edaf9: Preparing 2a05609f9eb7: Preparing ee38209e0e6f: Preparing e0030b440866: Preparing fb30fdb7c058: Preparing f47b6d67a712: Preparing e7ba581cd813: Preparing dcd02f5423e8: Preparing 8fdb126a4104: Preparing 526280e97149: Preparing 3eade0ed9ec1: Preparing e36ed6debb2b: Preparing 0fd5b5653d27: Preparing 01fa552d83b7: Preparing 6075904963c7: Preparing 404f28493dc2: Preparing 0a5e104fee35: Preparing c98d2b1bbbf6: Preparing 8d02698ec561: Preparing a808c8d1e663: Preparing ab9c297d25c9: Preparing 160d155f20aa: Preparing 1c2837c8f9f6: Preparing 15e4b3240a56: Preparing 12c7083a16b4: Preparing a3c06f196aa3: Preparing d20ece5534b3: Preparing 5049b77a7211: Preparing c45c46033aa8: Preparing b8a36d10656a: Preparing 0fd5b5653d27: Waiting 01fa552d83b7: Waiting 6075904963c7: Waiting 404f28493dc2: Waiting 0a5e104fee35: Waiting c98d2b1bbbf6: Waiting 8d02698ec561: Waiting a808c8d1e663: Waiting ab9c297d25c9: Waiting 160d155f20aa: Waiting 1c2837c8f9f6: Waiting 15e4b3240a56: Waiting 12c7083a16b4: Waiting a3c06f196aa3: Waiting d20ece5534b3: Waiting 5049b77a7211: Waiting c45c46033aa8: Waiting b8a36d10656a: Waiting ee38209e0e6f: Waiting e0030b440866: Waiting fb30fdb7c058: Waiting f47b6d67a712: Waiting e7ba581cd813: Waiting dcd02f5423e8: Waiting 8fdb126a4104: Waiting 526280e97149: Waiting 3eade0ed9ec1: Waiting e36ed6debb2b: Waiting 563f13631928: Waiting 5cc2fef50bf6: Waiting b805253edaf9: Waiting 2a05609f9eb7: Waiting cb324cf89811: Layer already exists 11ac03f3a24b: Layer already exists 5f70bf18a086: Layer already exists c09101d9181a: Layer already exists 67c330ec8773: Layer already exists 563f13631928: Layer already exists ee38209e0e6f: Layer already exists 5cc2fef50bf6: Layer already exists b805253edaf9: Layer already exists 2a05609f9eb7: Layer already exists e0030b440866: Layer already exists fb30fdb7c058: Layer already exists f47b6d67a712: Layer already exists dcd02f5423e8: Layer already exists e7ba581cd813: Layer already exists 526280e97149: Layer already exists 8fdb126a4104: Layer already exists 3eade0ed9ec1: Layer already exists e36ed6debb2b: Layer already exists 0fd5b5653d27: Layer already exists 01fa552d83b7: Layer already exists 404f28493dc2: Layer already exists 0a5e104fee35: Layer already exists c98d2b1bbbf6: Layer already exists 6075904963c7: Layer already exists 8d02698ec561: Layer already exists a808c8d1e663: Layer already exists 160d155f20aa: Layer already exists 1c2837c8f9f6: Layer already exists ab9c297d25c9: Layer already exists 15e4b3240a56: Layer already exists 12c7083a16b4: Layer already exists a3c06f196aa3: Layer already exists 5049b77a7211: Layer already exists c45c46033aa8: Layer already exists d20ece5534b3: Layer already exists b8a36d10656a: Layer already exists xlou: digest: sha256:66bfac8bd4bc1a71e39f313710078aac93681ca576b22af234556b0b4589b577 size: 8256 The push refers to repository [gcr.io/engineeringpit/lodestar-images/idm] 97a547b3f6e9: Preparing 3eb9771a009f: Preparing 59684c2b39f2: Preparing 42d28cb3a9e5: Preparing 2050fec9c124: Preparing 5f70bf18a086: Preparing 76b893fb3893: Preparing e0d22da03bc0: Preparing e95bc9916c31: Preparing b691ba8b88f7: Preparing 4cc825b0b052: Preparing 64b54d37c134: Preparing d85b356ec3b5: Preparing e95bc9916c31: Waiting b691ba8b88f7: Waiting 4cc825b0b052: Waiting 64b54d37c134: Waiting d85b356ec3b5: Waiting 76b893fb3893: Waiting e0d22da03bc0: Waiting 5f70bf18a086: Waiting 59684c2b39f2: Layer already exists 2050fec9c124: Layer already exists 42d28cb3a9e5: Layer already exists 3eb9771a009f: Layer already exists 97a547b3f6e9: Layer already exists 5f70bf18a086: Layer already exists e95bc9916c31: Layer already exists e0d22da03bc0: Layer already exists b691ba8b88f7: Layer already exists 76b893fb3893: Layer already exists 64b54d37c134: Layer already exists d85b356ec3b5: Layer already exists 4cc825b0b052: Layer already exists xlou: digest: sha256:664a0519a4c844f22781ae04ab861ff0821fa305cdc70d8498fab40f191ef9a9 size: 3034 The push refers to repository [gcr.io/engineeringpit/lodestar-images/ds] 53abd0139f0b: Preparing 893d87c39840: Preparing d7debbb825d9: Preparing 32348f044aec: Preparing 6c001e2b6996: Preparing 1b8b833cd565: Preparing 5f70bf18a086: Preparing f7fd43ef4c02: Preparing 1b8b833cd565: Waiting 5f70bf18a086: Waiting f7fd43ef4c02: Waiting 7539a00bf17a: Preparing 6f2a14e5cf43: Preparing 7539a00bf17a: Waiting eb4da744c8d0: Preparing 82f581cdd867: Preparing 8553b91047da: Preparing 6f2a14e5cf43: Waiting eb4da744c8d0: Waiting 82f581cdd867: Waiting 8553b91047da: Waiting d7debbb825d9: Layer already exists 893d87c39840: Layer already exists 6c001e2b6996: Layer already exists 32348f044aec: Layer already exists 53abd0139f0b: Layer already exists 1b8b833cd565: Layer already exists f7fd43ef4c02: Layer already exists 5f70bf18a086: Layer already exists 7539a00bf17a: Layer already exists 6f2a14e5cf43: Layer already exists eb4da744c8d0: Layer already exists 82f581cdd867: Layer already exists 8553b91047da: Layer already exists xlou: digest: sha256:7890b221cbd1b4c28f5d247e2290addf29203788fa0b8829048ed247279bd9b5 size: 3046 The push refers to repository [gcr.io/engineeringpit/lodestar-images/ig] f32ddc8d674f: Preparing e1e64a249e0a: Preparing 74eeec92b392: Preparing 10a523df8618: Preparing 4fb17506c7d6: Preparing 5696f243e6cc: Preparing 964c1eecc7f5: Preparing ab8038891451: Preparing c6f8bfcecf05: Preparing 315cd8c5da97: Preparing d456513ae67c: Preparing 67a4178b7d47: Preparing 5696f243e6cc: Waiting 964c1eecc7f5: Waiting ab8038891451: Waiting c6f8bfcecf05: Waiting 315cd8c5da97: Waiting d456513ae67c: Waiting 67a4178b7d47: Waiting 10a523df8618: Layer already exists f32ddc8d674f: Layer already exists e1e64a249e0a: Layer already exists 4fb17506c7d6: Layer already exists 74eeec92b392: Layer already exists 964c1eecc7f5: Layer already exists ab8038891451: Layer already exists 5696f243e6cc: Layer already exists c6f8bfcecf05: Layer already exists 315cd8c5da97: Layer already exists 67a4178b7d47: Layer already exists d456513ae67c: Layer already exists xlou: digest: sha256:c7c6bcae94d0bd089c6ed0efc947db82c1c2cee379d1539af9ec805dea9bb1d8 size: 2827 The push refers to repository [gcr.io/engineeringpit/lodestar-images/ds-idrepo] e3d10570bd65: Preparing 70c4a85e0765: Preparing de08045b9036: Preparing 0df38de64d88: Preparing 9494b5f1b11f: Preparing c49083b347d3: Preparing f9fb53b8dba9: Preparing 5f70bf18a086: Preparing 8b6c9fbaceb2: Preparing 1b8b833cd565: Preparing 5f70bf18a086: Preparing f7fd43ef4c02: Preparing 7539a00bf17a: Preparing 6f2a14e5cf43: Preparing eb4da744c8d0: Preparing 82f581cdd867: Preparing 8553b91047da: Preparing c49083b347d3: Waiting f9fb53b8dba9: Waiting 5f70bf18a086: Waiting 8b6c9fbaceb2: Waiting 1b8b833cd565: Waiting f7fd43ef4c02: Waiting 7539a00bf17a: Waiting 6f2a14e5cf43: Waiting eb4da744c8d0: Waiting 82f581cdd867: Waiting 8553b91047da: Waiting 9494b5f1b11f: Layer already exists 0df38de64d88: Layer already exists e3d10570bd65: Layer already exists 70c4a85e0765: Layer already exists de08045b9036: Layer already exists c49083b347d3: Layer already exists 5f70bf18a086: Layer already exists 8b6c9fbaceb2: Layer already exists 1b8b833cd565: Layer already exists f9fb53b8dba9: Layer already exists f7fd43ef4c02: Layer already exists 7539a00bf17a: Layer already exists 6f2a14e5cf43: Layer already exists eb4da744c8d0: Layer already exists 82f581cdd867: Layer already exists 8553b91047da: Layer already exists xlou: digest: sha256:dbd4655eb8b317a3ae03d3a9b57fbdee5672d41a33669d53f0ac2bb9f282c3d6 size: 3868 The push refers to repository [gcr.io/engineeringpit/lodestar-images/ds-cts] 7ef85619dc20: Preparing ff4cbc809bf2: Preparing a8ca0f25023b: Preparing f9fb53b8dba9: Preparing ca5f602b0d9f: Preparing 8b6c9fbaceb2: Preparing 1b8b833cd565: Preparing 5f70bf18a086: Preparing f7fd43ef4c02: Preparing 7539a00bf17a: Preparing 6f2a14e5cf43: Preparing eb4da744c8d0: Preparing 82f581cdd867: Preparing 8553b91047da: Preparing 8b6c9fbaceb2: Waiting 1b8b833cd565: Waiting 5f70bf18a086: Waiting f7fd43ef4c02: Waiting 7539a00bf17a: Waiting 6f2a14e5cf43: Waiting eb4da744c8d0: Waiting 82f581cdd867: Waiting 8553b91047da: Waiting ff4cbc809bf2: Layer already exists a8ca0f25023b: Layer already exists ca5f602b0d9f: Layer already exists f9fb53b8dba9: Layer already exists 7ef85619dc20: Layer already exists 8b6c9fbaceb2: Layer already exists 1b8b833cd565: Layer already exists 5f70bf18a086: Layer already exists f7fd43ef4c02: Layer already exists 7539a00bf17a: Layer already exists 6f2a14e5cf43: Layer already exists eb4da744c8d0: Layer already exists 82f581cdd867: Layer already exists 8553b91047da: Layer already exists xlou: digest: sha256:420412053a18973a6c198a46f97278ffa0d980b969cb60f80bec88d77b8a140b size: 3251 Updated the image_defaulter with your new image for am: "gcr.io/engineeringpit/lodestar-images/am:xlou". Updated the image_defaulter with your new image for idm: "gcr.io/engineeringpit/lodestar-images/idm:xlou". Updated the image_defaulter with your new image for ds: "gcr.io/engineeringpit/lodestar-images/ds:xlou". Updated the image_defaulter with your new image for ig: "gcr.io/engineeringpit/lodestar-images/ig:xlou". Updated the image_defaulter with your new image for ds-idrepo: "gcr.io/engineeringpit/lodestar-images/ds-idrepo:xlou". Updated the image_defaulter with your new image for ds-cts: "gcr.io/engineeringpit/lodestar-images/ds-cts:xlou". #1 [internal] load .dockerignore #1 transferring context: 2B done #1 DONE 0.0s #2 [internal] load build definition from Dockerfile #2 transferring dockerfile: 489B done #2 DONE 0.0s #3 [internal] load metadata for gcr.io/forgerock-io/am-cdk/pit1:7.4.0-043c211b139ffe96bb9eb3a723a409253d1abf1c #3 DONE 0.6s #4 [1/5] FROM gcr.io/forgerock-io/am-cdk/pit1:7.4.0-043c211b139ffe96bb9eb3a723a409253d1abf1c@sha256:7a5ace02dbc45b98f2da784f5dc34fa83bea196d916411c4fb8728ec01b3294e #4 DONE 0.0s #5 [internal] load build context #5 transferring context: 2.94kB done #5 DONE 0.0s #6 [2/5] RUN echo "\033[0;36m*** Building 'cdk' profile ***\033[0m" #6 CACHED #7 [3/5] COPY --chown=forgerock:root config-profiles/cdk/ /home/forgerock/openam/ #7 CACHED #8 [4/5] COPY --chown=forgerock:root *.sh /home/forgerock/ #8 CACHED #9 [5/5] WORKDIR /home/forgerock #9 CACHED #10 exporting to image #10 exporting layers done #10 writing image sha256:b8613da4d6f4b1c18cbfe22be567ac7630b3ff702b625186ea1ffc715aee3284 done #10 naming to gcr.io/engineeringpit/lodestar-images/am:xlou done #10 DONE 0.0s #1 [internal] load .dockerignore #1 transferring context: 2B done #1 DONE 0.0s #2 [internal] load build definition from Dockerfile #2 transferring dockerfile: 1.07kB done #2 DONE 0.0s #3 [internal] load metadata for gcr.io/forgerock-io/idm-cdk/pit1:7.4.0-292ff61a9416bbbef0d5118acbe9ed7d498d0f1a #3 DONE 0.5s #4 [1/6] FROM gcr.io/forgerock-io/idm-cdk/pit1:7.4.0-292ff61a9416bbbef0d5118acbe9ed7d498d0f1a@sha256:9a2ffe2ae9dcf9d828ce5a47594108c9740ff71012258c5ee4803fb313912061 #4 DONE 0.0s #5 [internal] load build context #5 transferring context: 279.82kB 0.0s done #5 DONE 0.0s #6 [3/6] RUN rm -f bundle/org.apache.felix.webconsole*.jar && rm -f bundle/openidm-felix-webconsole-*.jar #6 CACHED #7 [4/6] RUN echo "\033[0;36m*** Building 'cdk' profile ***\033[0m" #7 CACHED #8 [5/6] COPY --chown=forgerock:root config-profiles/cdk/ /opt/openidm #8 CACHED #9 [2/6] COPY debian-buster-sources.list /etc/apt/sources.list #9 CACHED #10 [6/6] COPY --chown=forgerock:root . /opt/openidm #10 CACHED #11 exporting to image #11 exporting layers done #11 writing image sha256:4bc70e4ce9488eafbb7207c049ed857fdc90320e91b7aa552e5420daecfaaf65 done #11 naming to gcr.io/engineeringpit/lodestar-images/idm:xlou done #11 DONE 0.0s #1 [internal] load build definition from Dockerfile #1 transferring dockerfile: 1.44kB done #1 DONE 0.0s #2 [internal] load .dockerignore #2 transferring context: 2B done #2 DONE 0.0s #3 [internal] load metadata for gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef #3 DONE 0.5s #4 [1/6] FROM gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef@sha256:bf08d264ca39f629598b651b09670878b2f80df77baf211ba42b7a5915c13b56 #4 DONE 0.0s #5 [internal] load build context #5 transferring context: 106.39kB 0.0s done #5 DONE 0.0s #6 [4/6] COPY --chown=forgerock:root ldif-ext /opt/opendj/ldif-ext #6 CACHED #7 [2/6] RUN apt-get update && apt-get install -y --no-install-recommends vim ncat dnsutils #7 CACHED #8 [3/6] COPY --chown=forgerock:root default-scripts /opt/opendj/default-scripts #8 CACHED #9 [5/6] COPY --chown=forgerock:root *.sh /opt/opendj/ #9 CACHED #10 [6/6] RUN ./ds-setup.sh && rm ./ds-setup.sh && rm -fr ldif-ext #10 CACHED #11 exporting to image #11 exporting layers done #11 writing image sha256:69c2b770da8341ed1000a6e0238372c80f40e5c6f73420b07933536122b82510 done #11 naming to gcr.io/engineeringpit/lodestar-images/ds:xlou done #11 DONE 0.0s #1 [internal] load .dockerignore #1 transferring context: 2B done #1 DONE 0.0s #2 [internal] load build definition from Dockerfile #2 transferring dockerfile: 504B done #2 DONE 0.0s #3 [internal] load metadata for gcr.io/forgerock-io/ig/pit1:7.3.0-latest-postcommit #3 DONE 0.5s #4 [1/5] FROM gcr.io/forgerock-io/ig/pit1:7.3.0-latest-postcommit@sha256:6cecde6c0a83637b7fe45bad89e44d9a1d1c7f1465e4fb2b47e5628209f80806 #4 DONE 0.0s #5 [internal] load build context #5 transferring context: 18.23kB done #5 DONE 0.0s #6 [2/5] COPY debian-buster-sources.list /etc/apt/sources.list #6 CACHED #7 [3/5] RUN echo "\033[0;36m*** Building 'cdk' profile ***\033[0m" #7 CACHED #8 [4/5] COPY --chown=forgerock:root config-profiles/cdk/ /var/ig #8 CACHED #9 [5/5] COPY --chown=forgerock:root . /var/ig #9 CACHED #10 exporting to image #10 exporting layers done #10 writing image sha256:a41b151894fea0001ac2c73d68d87a320b3c67c67661fbed2662a986bd076f2f done #10 naming to gcr.io/engineeringpit/lodestar-images/ig:xlou done #10 DONE 0.0s #1 [internal] load .dockerignore #1 transferring context: 2B done #1 DONE 0.0s #2 [internal] load build definition from Dockerfile #2 transferring dockerfile: 2.45kB done #2 DONE 0.0s #3 [internal] load metadata for gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef #3 DONE 0.5s #4 [ 1/10] FROM gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef@sha256:bf08d264ca39f629598b651b09670878b2f80df77baf211ba42b7a5915c13b56 #4 DONE 0.0s #5 [internal] load build context #5 transferring context: 103.39kB 0.0s done #5 DONE 0.0s #6 [ 8/10] COPY --chown=forgerock:root idrepo/*.ldif /var/tmp/ #6 CACHED #7 [ 5/10] COPY --chown=forgerock:root idrepo /opt/opendj/ #7 CACHED #8 [ 4/10] COPY --chown=forgerock:root common /opt/opendj/ #8 CACHED #9 [ 6/10] COPY --chown=forgerock:root scripts /opt/opendj/scripts #9 CACHED #10 [ 7/10] COPY --chown=forgerock:root uma /opt/opendj/uma #10 CACHED #11 [ 9/10] RUN chmod +w template/setup-profiles/AM/config/6.5/base-entries.ldif && cat scripts/external-am-datastore.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat uma/opendj_uma_audit.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat uma/opendj_uma_pending_requests.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat uma/opendj_uma_resource_set_labels.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat uma/opendj_uma_resource_sets.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat /var/tmp/alpha_bravo.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && chmod +w template/setup-profiles/AM/identity-store/7.0/base-entries.ldif && cat /var/tmp/orgs.ldif >> template/setup-profiles/AM/identity-store/7.0/base-entries.ldif && rm /var/tmp/*ldif #11 CACHED #12 [ 2/10] COPY debian-buster-sources.list /etc/apt/sources.list #12 CACHED #13 [ 3/10] WORKDIR /opt/opendj #13 CACHED #14 [10/10] RUN bin/setup.sh && bin/relax-security-settings.sh && rm bin/setup.sh bin/relax-security-settings.sh #14 CACHED #15 exporting to image #15 exporting layers done #15 writing image sha256:f14f024ab851d96361bf27ea255694c585fa1bda55159ea71eb54d6ad7bbf670 done #15 naming to gcr.io/engineeringpit/lodestar-images/ds-idrepo:xlou done #15 DONE 0.0s #1 [internal] load .dockerignore #1 transferring context: 2B done #1 DONE 0.0s #2 [internal] load build definition from Dockerfile #2 transferring dockerfile: 529B done #2 DONE 0.0s #3 [internal] load metadata for gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef #3 DONE 0.4s #4 [1/7] FROM gcr.io/forgerock-io/ds/pit1:7.4.0-5addc3263b048fa1029966f912c3280fe5b9c3ef@sha256:bf08d264ca39f629598b651b09670878b2f80df77baf211ba42b7a5915c13b56 #4 DONE 0.0s #5 [internal] load build context #5 transferring context: 4.77kB done #5 DONE 0.0s #6 [5/7] COPY --chown=forgerock:root cts /opt/opendj/ #6 CACHED #7 [4/7] COPY --chown=forgerock:root common /opt/opendj/ #7 CACHED #8 [6/7] COPY --chown=forgerock:root scripts /opt/opendj/scripts #8 CACHED #9 [2/7] COPY debian-buster-sources.list /etc/apt/sources.list #9 CACHED #10 [3/7] RUN chown -R forgerock:root /opt/opendj #10 CACHED #11 [7/7] RUN bin/setup.sh && bin/relax-security-settings.sh && rm bin/setup.sh bin/relax-security-settings.sh #11 CACHED #12 exporting to image #12 exporting layers done #12 writing image sha256:ca47500eed69060bbc446fffe43c24bcedc3261be2df0eccc7a6129374d7bccf done #12 naming to gcr.io/engineeringpit/lodestar-images/ds-cts:xlou done #12 DONE 0.0s [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/forgeops install --namespace=xlou --fqdn xlou.iam.xlou-cdm.engineeringpit.com --custom /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/overlay/internal-profiles/medium-old --legacy all [run_command]: env={'HOME': '/home/xslou', 'PATH': '/mnt/disks/data/xslou/lodestar-fork/ext/bin:/home/xslou/.local/bin:/home/xslou/bin:~/bin:~/workshop/lodestar-fork/ext/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin', 'DOCKER_SCAN_SUGGEST': 'false'} customresourcedefinition.apiextensions.k8s.io/secretagentconfigurations.secret-agent.secrets.forgerock.io condition met deployment.apps/secret-agent-controller-manager condition met NAME READY STATUS RESTARTS AGE secret-agent-controller-manager-85df555854-xg66k 2/2 Running 0 27h configmap/dev-utils created configmap/platform-config created ingress.networking.k8s.io/forgerock created ingress.networking.k8s.io/ig created certificate.cert-manager.io/ds-master-cert created certificate.cert-manager.io/ds-ssl-cert created issuer.cert-manager.io/selfsigned-issuer created secretagentconfiguration.secret-agent.secrets.forgerock.io/forgerock-sac created Checking cert-manager and related CRDs: cert-manager CRD found in cluster. Checking secret-agent operator and related CRDs: secret-agent CRD found in cluster.  Checking secret-agent operator is running... secret-agent operator is running Installing component(s): ['all'] platform: "custom-old" in namespace: "xlou" from deployment manifests in "None".  Deploying base.yaml. This is a one time activity.  Waiting for K8s secrets. Waiting for secret "am-env-secrets" to exist in the cluster: ..done Waiting for secret "idm-env-secrets" to exist in the cluster: ...done Waiting for secret "ds-passwords" to exist in the cluster: done Waiting for secret "ds-env-secrets" to exist in the cluster: secret/cloud-storage-credentials-cts created secret/cloud-storage-credentials-idrepo created service/ds-cts created service/ds-idrepo created statefulset.apps/ds-cts created statefulset.apps/ds-idrepo created job.batch/ldif-importer created done  Deploying ds.yaml. This includes all directory resources.  Waiting for DS deployment. This can take a few minutes. First installation takes longer. Waiting for statefulset "ds-idrepo" to exist in the cluster: Waiting for 3 pods to be ready... Waiting for 2 pods to be ready... Waiting for 1 pods to be ready... statefulset rolling update complete 3 pods at revision ds-idrepo-77b74cbf96... done Waiting for Service Account Password Update: done Waiting for statefulset "ds-cts" to exist in the cluster: statefulset rolling update complete 3 pods at revision ds-cts-7c476b6f6d... done Waiting for Service Account Password Update: configmap/amster-files created configmap/idm created configmap/idm-logging-properties created service/am created service/idm created deployment.apps/am created deployment.apps/idm created job.batch/amster created done Cleaning up amster components.  Deploying apps.  Waiting for AM deployment. This can take a few minutes. First installation takes longer. Waiting for deployment "am" to exist in the cluster: deployment.apps/am condition met configmap/amster-retain created done  Waiting for amster job to complete. This can take several minutes. Waiting for job "amster" to exist in the cluster: job.batch/amster condition met done  Waiting for IDM deployment. This can take a few minutes. First installation takes longer. Waiting for deployment "idm" to exist in the cluster: pod/idm-58dc667486-gclr8 condition met pod/idm-58dc667486-h86fb condition met service/admin-ui created service/end-user-ui created service/login-ui created deployment.apps/admin-ui created deployment.apps/end-user-ui created deployment.apps/login-ui created done  Deploying UI.  Waiting for K8s secrets. Waiting for secret "am-env-secrets" to exist in the cluster: done Waiting for secret "idm-env-secrets" to exist in the cluster: done Waiting for secret "ds-passwords" to exist in the cluster: done Waiting for secret "ds-env-secrets" to exist in the cluster: done  Relevant passwords: qAVhjrH2ahSOqYrBZgEcSXZS (amadmin user) wEcQVfDwdrG8TaZMLfSX2nfQQCYInCpr (uid=admin user) 2AjtiUroja5rEF3tKqkiPV8rQalYv9TM (App str svc acct (uid=am-config,ou=admins,ou=am-config)) QtrmaAQiekIcOvPkr0dZMddst3Heb4ZO (CTS svc acct (uid=openam_cts,ou=admins,ou=famrecords,ou=openam-session,ou=tokens)) ksOgXiM4JeStGHnNcOJCRLcxuH0mn3uo (ID repo svc acct (uid=am-identity-bind-account,ou=admins,ou=identities))  Relevant URLs: https://xlou.iam.xlou-cdm.engineeringpit.com/platform https://xlou.iam.xlou-cdm.engineeringpit.com/admin https://xlou.iam.xlou-cdm.engineeringpit.com/am https://xlou.iam.xlou-cdm.engineeringpit.com/enduser  Enjoy your deployment! **************************** Initializing component pods for DS-CTS **************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets --selector app=ds-cts --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods --selector app=ds-cts --output jsonpath={.items[*].metadata.name}` | grep 3 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- ds-cts-0 ds-cts-1 ds-cts-2 --- stderr --- -------------------- Check pod ds-cts-0 is running -------------------- [loop_until]: kubectl --namespace=xlou get pods ds-cts-0 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-cts-0 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-cts-0 --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:53:52Z --- stderr --- ------------- Check pod ds-cts-0 filesystem is accessible ------------- [loop_until]: kubectl --namespace=xlou exec ds-cts-0 --container ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------------ Check pod ds-cts-0 restart count ------------------ [loop_until]: kubectl --namespace=xlou get pod ds-cts-0 --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-cts-0 has been restarted 0 times. -------------------- Check pod ds-cts-1 is running -------------------- [loop_until]: kubectl --namespace=xlou get pods ds-cts-1 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-cts-1 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-cts-1 --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:54:12Z --- stderr --- ------------- Check pod ds-cts-1 filesystem is accessible ------------- [loop_until]: kubectl --namespace=xlou exec ds-cts-1 --container ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------------ Check pod ds-cts-1 restart count ------------------ [loop_until]: kubectl --namespace=xlou get pod ds-cts-1 --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-cts-1 has been restarted 0 times. -------------------- Check pod ds-cts-2 is running -------------------- [loop_until]: kubectl --namespace=xlou get pods ds-cts-2 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-cts-2 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-cts-2 --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:54:39Z --- stderr --- ------------- Check pod ds-cts-2 filesystem is accessible ------------- [loop_until]: kubectl --namespace=xlou exec ds-cts-2 --container ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------------ Check pod ds-cts-2 restart count ------------------ [loop_until]: kubectl --namespace=xlou get pod ds-cts-2 --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-cts-2 has been restarted 0 times. ************************** Initializing component pods for DS-IDREPO ************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets --selector app=ds-idrepo --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods --selector app=ds-idrepo --output jsonpath={.items[*].metadata.name}` | grep 3 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- ds-idrepo-0 ds-idrepo-1 ds-idrepo-2 --- stderr --- ------------------ Check pod ds-idrepo-0 is running ------------------ [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-0 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-0 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-0 --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:53:52Z --- stderr --- ----------- Check pod ds-idrepo-0 filesystem is accessible ----------- [loop_until]: kubectl --namespace=xlou exec ds-idrepo-0 --container ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ----------------- Check pod ds-idrepo-0 restart count ----------------- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-0 --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-idrepo-0 has been restarted 0 times. ------------------ Check pod ds-idrepo-1 is running ------------------ [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-1 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-1 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-1 --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:54:29Z --- stderr --- ----------- Check pod ds-idrepo-1 filesystem is accessible ----------- [loop_until]: kubectl --namespace=xlou exec ds-idrepo-1 --container ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ----------------- Check pod ds-idrepo-1 restart count ----------------- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-1 --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-idrepo-1 has been restarted 0 times. ------------------ Check pod ds-idrepo-2 is running ------------------ [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-2 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-2 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-2 --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:55:00Z --- stderr --- ----------- Check pod ds-idrepo-2 filesystem is accessible ----------- [loop_until]: kubectl --namespace=xlou exec ds-idrepo-2 --container ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ----------------- Check pod ds-idrepo-2 restart count ----------------- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-2 --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-idrepo-2 has been restarted 0 times. ****************************** Initializing component pods for AM ****************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app=am --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods --selector app=am --output jsonpath={.items[*].metadata.name}` | grep 3 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- am-7849cf7bdb-72wqt am-7849cf7bdb-cwb58 am-7849cf7bdb-vcl2f --- stderr --- -------------- Check pod am-7849cf7bdb-72wqt is running -------------- [loop_until]: kubectl --namespace=xlou get pods am-7849cf7bdb-72wqt -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods am-7849cf7bdb-72wqt -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod am-7849cf7bdb-72wqt --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:55:33Z --- stderr --- ------- Check pod am-7849cf7bdb-72wqt filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec am-7849cf7bdb-72wqt --container openam -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev etc home lib lib32 lib64 libx32 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------- Check pod am-7849cf7bdb-72wqt restart count ------------- [loop_until]: kubectl --namespace=xlou get pod am-7849cf7bdb-72wqt --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod am-7849cf7bdb-72wqt has been restarted 0 times. -------------- Check pod am-7849cf7bdb-cwb58 is running -------------- [loop_until]: kubectl --namespace=xlou get pods am-7849cf7bdb-cwb58 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods am-7849cf7bdb-cwb58 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod am-7849cf7bdb-cwb58 --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:55:32Z --- stderr --- ------- Check pod am-7849cf7bdb-cwb58 filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec am-7849cf7bdb-cwb58 --container openam -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev etc home lib lib32 lib64 libx32 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------- Check pod am-7849cf7bdb-cwb58 restart count ------------- [loop_until]: kubectl --namespace=xlou get pod am-7849cf7bdb-cwb58 --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod am-7849cf7bdb-cwb58 has been restarted 0 times. -------------- Check pod am-7849cf7bdb-vcl2f is running -------------- [loop_until]: kubectl --namespace=xlou get pods am-7849cf7bdb-vcl2f -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods am-7849cf7bdb-vcl2f -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod am-7849cf7bdb-vcl2f --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:55:33Z --- stderr --- ------- Check pod am-7849cf7bdb-vcl2f filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec am-7849cf7bdb-vcl2f --container openam -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev etc home lib lib32 lib64 libx32 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------- Check pod am-7849cf7bdb-vcl2f restart count ------------- [loop_until]: kubectl --namespace=xlou get pod am-7849cf7bdb-vcl2f --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod am-7849cf7bdb-vcl2f has been restarted 0 times. **************************** Initializing component pods for AMSTER **************************** ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou --field-selector status.phase!=Failed get pods --selector app=amster --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- amster-2mrjd --- stderr --- ***************************** Initializing component pods for IDM ***************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployment --selector app=idm --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods --selector app=idm --output jsonpath={.items[*].metadata.name}` | grep 2 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- idm-58dc667486-gclr8 idm-58dc667486-h86fb --- stderr --- -------------- Check pod idm-58dc667486-gclr8 is running -------------- [loop_until]: kubectl --namespace=xlou get pods idm-58dc667486-gclr8 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods idm-58dc667486-gclr8 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod idm-58dc667486-gclr8 --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:55:33Z --- stderr --- ------- Check pod idm-58dc667486-gclr8 filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec idm-58dc667486-gclr8 --container openidm -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-11 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------ Check pod idm-58dc667486-gclr8 restart count ------------ [loop_until]: kubectl --namespace=xlou get pod idm-58dc667486-gclr8 --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod idm-58dc667486-gclr8 has been restarted 0 times. -------------- Check pod idm-58dc667486-h86fb is running -------------- [loop_until]: kubectl --namespace=xlou get pods idm-58dc667486-h86fb -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods idm-58dc667486-h86fb -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod idm-58dc667486-h86fb --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:55:33Z --- stderr --- ------- Check pod idm-58dc667486-h86fb filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec idm-58dc667486-h86fb --container openidm -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-11 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------ Check pod idm-58dc667486-h86fb restart count ------------ [loop_until]: kubectl --namespace=xlou get pod idm-58dc667486-h86fb --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod idm-58dc667486-h86fb has been restarted 0 times. ************************* Initializing component pods for END-USER-UI ************************* --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app.kubernetes.io/name=end-user-ui --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods --selector app.kubernetes.io/name=end-user-ui --output jsonpath={.items[*].metadata.name}` | grep 1 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- end-user-ui-fdf99b7d7-4gw9n --- stderr --- ---------- Check pod end-user-ui-fdf99b7d7-4gw9n is running ---------- [loop_until]: kubectl --namespace=xlou get pods end-user-ui-fdf99b7d7-4gw9n -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods end-user-ui-fdf99b7d7-4gw9n -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod end-user-ui-fdf99b7d7-4gw9n --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:56:35Z --- stderr --- --- Check pod end-user-ui-fdf99b7d7-4gw9n filesystem is accessible --- [loop_until]: kubectl --namespace=xlou exec end-user-ui-fdf99b7d7-4gw9n --container end-user-ui -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev docker-entrypoint.d docker-entrypoint.sh entrypoint.sh etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var variable_replacement.sh --- stderr --- --------- Check pod end-user-ui-fdf99b7d7-4gw9n restart count --------- [loop_until]: kubectl --namespace=xlou get pod end-user-ui-fdf99b7d7-4gw9n --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod end-user-ui-fdf99b7d7-4gw9n has been restarted 0 times. *************************** Initializing component pods for LOGIN-UI *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app.kubernetes.io/name=login-ui --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods --selector app.kubernetes.io/name=login-ui --output jsonpath={.items[*].metadata.name}` | grep 1 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- login-ui-7fdfb4d89f-szrsp --- stderr --- ----------- Check pod login-ui-7fdfb4d89f-szrsp is running ----------- [loop_until]: kubectl --namespace=xlou get pods login-ui-7fdfb4d89f-szrsp -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods login-ui-7fdfb4d89f-szrsp -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod login-ui-7fdfb4d89f-szrsp --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:56:36Z --- stderr --- ---- Check pod login-ui-7fdfb4d89f-szrsp filesystem is accessible ---- [loop_until]: kubectl --namespace=xlou exec login-ui-7fdfb4d89f-szrsp --container login-ui -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev docker-entrypoint.d docker-entrypoint.sh entrypoint.sh etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var variable_replacement.sh --- stderr --- ---------- Check pod login-ui-7fdfb4d89f-szrsp restart count ---------- [loop_until]: kubectl --namespace=xlou get pod login-ui-7fdfb4d89f-szrsp --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod login-ui-7fdfb4d89f-szrsp has been restarted 0 times. *************************** Initializing component pods for ADMIN-UI *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app.kubernetes.io/name=admin-ui --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods --selector app.kubernetes.io/name=admin-ui --output jsonpath={.items[*].metadata.name}` | grep 1 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- admin-ui-7c7859f797-fzrqk --- stderr --- ----------- Check pod admin-ui-7c7859f797-fzrqk is running ----------- [loop_until]: kubectl --namespace=xlou get pods admin-ui-7c7859f797-fzrqk -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods admin-ui-7c7859f797-fzrqk -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod admin-ui-7c7859f797-fzrqk --output jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-05-28T05:56:35Z --- stderr --- ---- Check pod admin-ui-7c7859f797-fzrqk filesystem is accessible ---- [loop_until]: kubectl --namespace=xlou exec admin-ui-7c7859f797-fzrqk --container admin-ui -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev docker-entrypoint.d docker-entrypoint.sh entrypoint.sh etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var variable_replacement.sh --- stderr --- ---------- Check pod admin-ui-7c7859f797-fzrqk restart count ---------- [loop_until]: kubectl --namespace=xlou get pod admin-ui-7c7859f797-fzrqk --output jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod admin-ui-7c7859f797-fzrqk has been restarted 0 times. ***************************** Checking DS-CTS component is running ***************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets --selector app=ds-cts --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- -------------- Waiting for 3 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get statefulsets ds-cts --output jsonpath="current:{.status.currentReplicas} ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "current:3 ready:3 replicas:3" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- current:3 ready:3 replicas:3 --- stderr --- *************************** Checking DS-IDREPO component is running *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets --selector app=ds-idrepo --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- -------------- Waiting for 3 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get statefulsets ds-idrepo --output jsonpath="current:{.status.currentReplicas} ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "current:3 ready:3 replicas:3" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- current:3 ready:3 replicas:3 --- stderr --- ******************************* Checking AM component is running ******************************* --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app=am --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- -------------- Waiting for 3 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployments am --output jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:3 replicas:3" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:3 replicas:3 --- stderr --- ***************************** Checking AMSTER component is running ***************************** --------------------- Get expected number of pods --------------------- -------------- Waiting for 1 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get jobs amster --output jsonpath="{.status.succeeded}" | grep "1" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ****************************** Checking IDM component is running ****************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployment --selector app=idm --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2 --- stderr --- -------------- Waiting for 2 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployment idm --output jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:2 replicas:2" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:2 replicas:2 --- stderr --- ************************** Checking END-USER-UI component is running ************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app.kubernetes.io/name=end-user-ui --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- -------------- Waiting for 1 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployments end-user-ui --output jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:1 replicas:1" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:1 replicas:1 --- stderr --- **************************** Checking LOGIN-UI component is running **************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app.kubernetes.io/name=login-ui --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- -------------- Waiting for 1 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployments login-ui --output jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:1 replicas:1" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:1 replicas:1 --- stderr --- **************************** Checking ADMIN-UI component is running **************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app.kubernetes.io/name=admin-ui --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- -------------- Waiting for 1 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployments admin-ui --output jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:1 replicas:1" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:1 replicas:1 --- stderr --- ****************************** Livecheck stage: After deployment ****************************** ---------------------- Running DS-CTS livecheck ---------------------- Livecheck to ds-cts-0 [loop_until]: kubectl --namespace=xlou get secret ds-passwords --output jsonpath="{.data.dirmanager\.pw}" [loop_until]: (max_time=60, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- d0VjUVZmRHdkckc4VGFaTUxmU1gybmZRUUNZSW5DcHI= --- stderr --- [run_command]: kubectl --namespace=xlou exec ds-cts-0 --container ds -- ldapsearch --noPropertiesFile --port 1389 --useStartTls --trustAll --bindDn "uid=admin" --bindPassword "wEcQVfDwdrG8TaZMLfSX2nfQQCYInCpr" --baseDn "" --searchScope base "(&)" alive [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- dn: alive: true --- stderr --- Livecheck to ds-cts-1 [run_command]: kubectl --namespace=xlou exec ds-cts-1 --container ds -- ldapsearch --noPropertiesFile --port 1389 --useStartTls --trustAll --bindDn "uid=admin" --bindPassword "wEcQVfDwdrG8TaZMLfSX2nfQQCYInCpr" --baseDn "" --searchScope base "(&)" alive [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- dn: alive: true --- stderr --- Livecheck to ds-cts-2 [run_command]: kubectl --namespace=xlou exec ds-cts-2 --container ds -- ldapsearch --noPropertiesFile --port 1389 --useStartTls --trustAll --bindDn "uid=admin" --bindPassword "wEcQVfDwdrG8TaZMLfSX2nfQQCYInCpr" --baseDn "" --searchScope base "(&)" alive [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- dn: alive: true --- stderr --- --------------------- Running DS-IDREPO livecheck --------------------- Livecheck to ds-idrepo-0 [loop_until]: kubectl --namespace=xlou get secret ds-passwords --output jsonpath="{.data.dirmanager\.pw}" [loop_until]: (max_time=60, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- d0VjUVZmRHdkckc4VGFaTUxmU1gybmZRUUNZSW5DcHI= --- stderr --- [run_command]: kubectl --namespace=xlou exec ds-idrepo-0 --container ds -- ldapsearch --noPropertiesFile --port 1389 --useStartTls --trustAll --bindDn "uid=admin" --bindPassword "wEcQVfDwdrG8TaZMLfSX2nfQQCYInCpr" --baseDn "" --searchScope base "(&)" alive [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- dn: alive: true --- stderr --- Livecheck to ds-idrepo-1 [run_command]: kubectl --namespace=xlou exec ds-idrepo-1 --container ds -- ldapsearch --noPropertiesFile --port 1389 --useStartTls --trustAll --bindDn "uid=admin" --bindPassword "wEcQVfDwdrG8TaZMLfSX2nfQQCYInCpr" --baseDn "" --searchScope base "(&)" alive [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- dn: alive: true --- stderr --- Livecheck to ds-idrepo-2 [run_command]: kubectl --namespace=xlou exec ds-idrepo-2 --container ds -- ldapsearch --noPropertiesFile --port 1389 --useStartTls --trustAll --bindDn "uid=admin" --bindPassword "wEcQVfDwdrG8TaZMLfSX2nfQQCYInCpr" --baseDn "" --searchScope base "(&)" alive [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- dn: alive: true --- stderr --- ------------------------ Running AM livecheck ------------------------ Livecheck to https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/health/ready [http_cmd]: curl -L --request GET "https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/health/ready" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- [loop_until]: kubectl --namespace=xlou get secret am-env-secrets --output jsonpath="{.data.AM_PASSWORDS_AMADMIN_CLEAR}" [loop_until]: (max_time=60, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- cUFWaGpySDJhaFNPcVlyQlpnRWNTWFpT --- stderr --- Authenticate user amadmin via REST [http_cmd]: curl --header "X-OpenAM-Username: amadmin" --header "X-OpenAM-Password: qAVhjrH2ahSOqYrBZgEcSXZS" --header "Content-Type: application/json" --header "Accept-API-Version: resource=2.0, protocol=1.0" -L --request POST "https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/authenticate?realm=/" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- { "tokenId": "beJ--9VOhD57Tg-BVAIzlj4RdOc.*AAJTSQACMDIAAlNLABxuVWJlay9neEdiMFQyTzB2a3RQMHhaVFp3aVU9AAR0eXBlAANDVFMAAlMxAAIwMQ..*", "successUrl": "/am/console", "realm": "/" } ---------------------- Running AMSTER livecheck ---------------------- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou --field-selector status.phase!=Failed get pods --selector app=amster --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- amster-2mrjd --- stderr --- Amster livecheck is passed ------------------------ Running IDM livecheck ------------------------ Livecheck to https://xlou.iam.xlou-cdm.engineeringpit.com/openidm/info/ping [loop_until]: kubectl --namespace=xlou get secret idm-env-secrets --output jsonpath="{.data.OPENIDM_ADMIN_PASSWORD}" [loop_until]: (max_time=60, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- aDM5ZXpZaWNxcWtWcW1BNmp3WDFDQ3JR --- stderr --- Set admin password: h39ezYicqqkVqmA6jwX1CCrQ [http_cmd]: curl --header "X-OpenIDM-Username: anonymous" --header "X-OpenIDM-Password: anonymous" -L --request GET "https://xlou.iam.xlou-cdm.engineeringpit.com/openidm/info/ping" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- { "_id": "", "_rev": "", "shortDesc": "OpenIDM ready", "state": "ACTIVE_READY" } -------------------- Running END-USER-UI livecheck -------------------- Livecheck to https://xlou.iam.xlou-cdm.engineeringpit.com/enduser [http_cmd]: curl -L --request GET "https://xlou.iam.xlou-cdm.engineeringpit.com/enduser" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- Identity Management
[] --------------------- Running LOGIN-UI livecheck --------------------- Livecheck to https://xlou.iam.xlou-cdm.engineeringpit.com/am/XUI [http_cmd]: curl -L --request GET "https://xlou.iam.xlou-cdm.engineeringpit.com/am/XUI" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- Login
[] --------------------- Running ADMIN-UI livecheck --------------------- Livecheck to https://xlou.iam.xlou-cdm.engineeringpit.com/platform [http_cmd]: curl -L --request GET "https://xlou.iam.xlou-cdm.engineeringpit.com/platform" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- Platform Admin
[] LIVECHECK SUCCEEDED **************************** Initializing component pods for DS-CTS **************************** --------------------- Get DS-CTS software version --------------------- [loop_until]: kubectl --namespace=xlou exec ds-cts-0 --container ds -- find /opt/opendj -name opendj-core.jar [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /opt/opendj/lib/opendj-core.jar --- stderr --- [loop_until]: kubectl --namespace=xlou cp ds-cts-0:/opt/opendj/lib/opendj-core.jar /tmp/ds-cts_info/opendj-core.jar --container ds [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- ************************** Initializing component pods for DS-IDREPO ************************** ------------------- Get DS-IDREPO software version ------------------- [loop_until]: kubectl --namespace=xlou exec ds-idrepo-0 --container ds -- find /opt/opendj -name opendj-core.jar [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /opt/opendj/lib/opendj-core.jar --- stderr --- [loop_until]: kubectl --namespace=xlou cp ds-idrepo-0:/opt/opendj/lib/opendj-core.jar /tmp/ds-idrepo_info/opendj-core.jar --container ds [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- ****************************** Initializing component pods for AM ****************************** ----------------------- Get AM software version ----------------------- Getting product version from https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/serverinfo/version Create new LoginSession for user "amadmin" [LoginSession] Get "token_id" token for user "amadmin" [LoginSession] Obtaining new "token_id" token from server because it was not obtained yet for the user Authenticate user amadmin via REST [http_cmd]: curl --header "X-OpenAM-Username: amadmin" --header "X-OpenAM-Password: qAVhjrH2ahSOqYrBZgEcSXZS" --header "Content-Type: application/json" --header "Accept-API-Version: resource=2.0, protocol=1.0" -L --request POST "https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/authenticate?realm=/" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- { "tokenId": "yOSK-K6NjQxiMD8pSW56nopiLwY.*AAJTSQACMDIAAlNLABxyQ2pKNENzVTR0OGtaM2V0SGQvWm11OVp1ZEU9AAR0eXBlAANDVFMAAlMxAAIwMQ..*", "successUrl": "/am/console", "realm": "/" } [LoginSession] Obtaining session info [http_cmd]: curl --header "iPlanetDirectoryPro: yOSK-K6NjQxiMD8pSW56nopiLwY.*AAJTSQACMDIAAlNLABxyQ2pKNENzVTR0OGtaM2V0SGQvWm11OVp1ZEU9AAR0eXBlAANDVFMAAlMxAAIwMQ..*" --header "Content-Type: application/json" --header "Accept-API-Version: resource=3.0, protocol=2.1" --header "filters_cookie: yOSK-K6NjQxiMD8pSW56nopiLwY.*AAJTSQACMDIAAlNLABxyQ2pKNENzVTR0OGtaM2V0SGQvWm11OVp1ZEU9AAR0eXBlAANDVFMAAlMxAAIwMQ..*" -L --request POST "https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/realms/root/sessions?_action=getSessionInfo" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- { "username": "amadmin", "universalId": "id=amadmin,ou=user,ou=am-config", "realm": "/", "latestAccessTime": "2023-05-28T05:57:40Z", "maxIdleExpirationTime": "2023-05-28T06:27:40Z", "maxSessionExpirationTime": "2023-05-28T07:57:39Z", "properties": { "AMCtxId": "86a64c6a-b02d-4f61-b398-c3bf536d79e6-113" } } [http_cmd]: curl -L --request GET --cookie "amlbcookie=01" --cookie "iPlanetDirectoryPro=yOSK-K6NjQxiMD8pSW56nopiLwY.*AAJTSQACMDIAAlNLABxyQ2pKNENzVTR0OGtaM2V0SGQvWm11OVp1ZEU9AAR0eXBlAANDVFMAAlMxAAIwMQ..*" --cookie "route=1685253461.648.27870.223985|f60edb382037eb2df1e800d563ad78a7" "https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/serverinfo/version" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- { "_id": "version", "_rev": "-580918788", "version": "7.4.0-SNAPSHOT", "fullVersion": "ForgeRock Access Management 7.4.0-SNAPSHOT Build 043c211b139ffe96bb9eb3a723a409253d1abf1c (2023-May-24 15:01)", "revision": "043c211b139ffe96bb9eb3a723a409253d1abf1c", "date": "2023-May-24 15:01" } **************************** Initializing component pods for AMSTER **************************** ***************************** Initializing component pods for IDM ***************************** ---------------------- Get IDM software version ---------------------- Getting product version from https://xlou.iam.xlou-cdm.engineeringpit.com/openidm/info/version [http_cmd]: curl --header "X-OpenIDM-Username: anonymous" --header "X-OpenIDM-Password: anonymous" -L --request GET "https://xlou.iam.xlou-cdm.engineeringpit.com/openidm/info/version" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- { "_id": "version", "productVersion": "7.4.0-SNAPSHOT", "productBuildDate": "20230525155619", "productRevision": "292ff61a94" } ************************* Initializing component pods for END-USER-UI ************************* ------------------ Get END-USER-UI software version ------------------ [loop_until]: kubectl --namespace=xlou exec end-user-ui-fdf99b7d7-4gw9n --container end-user-ui -- find /usr/share/nginx/html -name chunk-vendors.*.js [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /usr/share/nginx/html/js/chunk-vendors.a51e8e31.js --- stderr --- [loop_until]: kubectl --namespace=xlou cp end-user-ui-fdf99b7d7-4gw9n:/usr/share/nginx/html/js/chunk-vendors.a51e8e31.js /tmp/end-user-ui_info/chunk-vendors.a51e8e31.js --container end-user-ui [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- *************************** Initializing component pods for LOGIN-UI *************************** -------------------- Get LOGIN-UI software version -------------------- [loop_until]: kubectl --namespace=xlou exec login-ui-7fdfb4d89f-szrsp --container login-ui -- find /usr/share/nginx/html -name chunk-vendors.*.js [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /usr/share/nginx/html/js/chunk-vendors.74d0e1f7.js --- stderr --- [loop_until]: kubectl --namespace=xlou cp login-ui-7fdfb4d89f-szrsp:/usr/share/nginx/html/js/chunk-vendors.74d0e1f7.js /tmp/login-ui_info/chunk-vendors.74d0e1f7.js --container login-ui [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- *************************** Initializing component pods for ADMIN-UI *************************** -------------------- Get ADMIN-UI software version -------------------- [loop_until]: kubectl --namespace=xlou exec admin-ui-7c7859f797-fzrqk --container admin-ui -- find /usr/share/nginx/html -name chunk-vendors.*.js [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /usr/share/nginx/html/js/chunk-vendors.f7439746.js --- stderr --- [loop_until]: kubectl --namespace=xlou cp admin-ui-7c7859f797-fzrqk:/usr/share/nginx/html/js/chunk-vendors.f7439746.js /tmp/admin-ui_info/chunk-vendors.f7439746.js --container admin-ui [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- ==================================================================================================== ================ Admin password for DS-CTS is: wEcQVfDwdrG8TaZMLfSX2nfQQCYInCpr ================ ==================================================================================================== ==================================================================================================== ============== Admin password for DS-IDREPO is: wEcQVfDwdrG8TaZMLfSX2nfQQCYInCpr ============== ==================================================================================================== ==================================================================================================== ====================== Admin password for AM is: qAVhjrH2ahSOqYrBZgEcSXZS ====================== ==================================================================================================== ==================================================================================================== ===================== Admin password for IDM is: h39ezYicqqkVqmA6jwX1CCrQ ===================== ==================================================================================================== *************************************** Dumping pod list *************************************** Dumping pod list to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/_pod-list.txt **************************** Initializing component pods for DS-CTS **************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets --selector app=ds-cts --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods --selector app=ds-cts --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- ds-cts-0 ds-cts-1 ds-cts-2 --- stderr --- ************************** Initializing component pods for DS-IDREPO ************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets --selector app=ds-idrepo --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods --selector app=ds-idrepo --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- ds-idrepo-0 ds-idrepo-1 ds-idrepo-2 --- stderr --- ****************************** Initializing component pods for AM ****************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app=am --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods --selector app=am --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- am-7849cf7bdb-72wqt am-7849cf7bdb-cwb58 am-7849cf7bdb-vcl2f --- stderr --- **************************** Initializing component pods for AMSTER **************************** ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods --selector app=amster --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- amster-2mrjd --- stderr --- ***************************** Initializing component pods for IDM ***************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployment --selector app=idm --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods --selector app=idm --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- idm-58dc667486-gclr8 idm-58dc667486-h86fb --- stderr --- ************************* Initializing component pods for END-USER-UI ************************* --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app.kubernetes.io/name=end-user-ui --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods --selector app.kubernetes.io/name=end-user-ui --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- end-user-ui-fdf99b7d7-4gw9n --- stderr --- *************************** Initializing component pods for LOGIN-UI *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app.kubernetes.io/name=login-ui --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods --selector app.kubernetes.io/name=login-ui --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- login-ui-7fdfb4d89f-szrsp --- stderr --- *************************** Initializing component pods for ADMIN-UI *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments --selector app.kubernetes.io/name=admin-ui --output jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods --selector app.kubernetes.io/name=admin-ui --output jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- admin-ui-7c7859f797-fzrqk --- stderr --- *********************************** Dumping components logs *********************************** ----------------------- Dumping logs for DS-CTS ----------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/ds-cts-0.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/ds-cts-1.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/ds-cts-2.txt Check pod logs for errors --------------------- Dumping logs for DS-IDREPO --------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/ds-idrepo-0.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/ds-idrepo-1.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/ds-idrepo-2.txt Check pod logs for errors ------------------------- Dumping logs for AM ------------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/am-7849cf7bdb-72wqt.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/am-7849cf7bdb-cwb58.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/am-7849cf7bdb-vcl2f.txt Check pod logs for errors ----------------------- Dumping logs for AMSTER ----------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/amster-2mrjd.txt Check pod logs for errors ------------------------ Dumping logs for IDM ------------------------ Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/idm-58dc667486-gclr8.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/idm-58dc667486-h86fb.txt Check pod logs for errors -------------------- Dumping logs for END-USER-UI -------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/end-user-ui-fdf99b7d7-4gw9n.txt Check pod logs for errors ---------------------- Dumping logs for LOGIN-UI ---------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/login-ui-7fdfb4d89f-szrsp.txt Check pod logs for errors ---------------------- Dumping logs for ADMIN-UI ---------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/simple_managed_users/pod-logs/stack/20230528-055747-after-deployment/admin-ui-7c7859f797-fzrqk.txt Check pod logs for errors [28/May/2023 05:58:09] - INFO: Deployment successful ________________________________________________________________________________ [28/May/2023 05:58:09] Deploy_all_forgerock_components post : Post method ________________________________________________________________________________ Setting result to PASS Task has been successfully stopped