--Task-- name: Deploy_all_forgerock_components enabled: True class_name: DeployComponentsTask source_name: controller source_namespace: >default< target_name: controller target_namespace: >default< start: 0 stop: None timeout: no timeout loop: False interval: None dependencies: ['Enable_prometheus_admin_api'] wait_for: [] options: {} group_name: None Current dir: /mnt/disks/data/xslou/lodestar-fork/pyrock ________________________________________________________________________________ [04/Apr/2023 21:56:46] Deploy_all_forgerock_components pre : Initialising task parameters ________________________________________________________________________________ task will be executed on controller (localhost) ________________________________________________________________________________ [04/Apr/2023 21:56:46] Deploy_all_forgerock_components step1 : Deploy components ________________________________________________________________________________ ******************************** Cleaning up existing namespace ******************************** ------------------ Deleting secret agent controller ------------------ [loop_until]: kubectl --namespace=xlou delete sac --all [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- No resources found --- stderr --- ----------------------- Deleting all resources ----------------------- [loop_until]: kubectl --namespace=xlou delete all --all --grace-period=0 --force [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- No resources found --- stderr --- warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. [loop_until]: kubectl -n xlou get pods | grep "No resources found" [loop_until]: (max_time=360, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- No resources found in xlou namespace. ------------------------- Deleting configmap ------------------------- [loop_until]: kubectl --namespace=xlou get configmap -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- --------------------------- Deleting secret --------------------------- [loop_until]: kubectl --namespace=xlou get secret -o jsonpath='{.items[?(@.type=="Opaque")].metadata.name}' [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- -------------------------- Deleting ingress -------------------------- [loop_until]: kubectl --namespace=xlou get ingress -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- ---------------------------- Deleting pvc ---------------------------- [loop_until]: kubectl --namespace=xlou get pvc -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- [loop_until]: kubectl --namespace=xlou delete pv ds-backup-xlou --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- warning: deleting cluster-scoped resources, not scoped to the provided namespace ----------------- Deleting admin clusterrolebindings ----------------- [loop_until]: kubectl get clusterrolebinding -o jsonpath="{range .items[?(@.subjects[0].namespace=='xlou')]}{.metadata.name} {end}" [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- ------------------------- Deleting namespace ------------------------- [loop_until]: kubectl delete namespaces xlou --ignore-not-found --grace-period=0 --force [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. [loop_until]: awk -F" " "{print NF}" <<< `kubectl get namespace xlou --ignore-not-found` | grep 0 [loop_until]: (max_time=600, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- ************************************* Creating deployment ************************************* Creating normal (forgeops) type deployment for deployment: stack ------- Custom component configuration present. Loading values ------- ------------------ Deleting secret agent controller ------------------ [loop_until]: kubectl --namespace=xlou delete sac --all [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- No resources found --- stderr --- ----------------------- Deleting all resources ----------------------- [loop_until]: kubectl --namespace=xlou delete all --all --grace-period=0 --force [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- No resources found --- stderr --- warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. [loop_until]: kubectl -n xlou get pods | grep "No resources found" [loop_until]: (max_time=360, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- No resources found in xlou namespace. ------------------------- Deleting configmap ------------------------- [loop_until]: kubectl --namespace=xlou get configmap -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- --------------------------- Deleting secret --------------------------- [loop_until]: kubectl --namespace=xlou get secret -o jsonpath='{.items[?(@.type=="Opaque")].metadata.name}' [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- -------------------------- Deleting ingress -------------------------- [loop_until]: kubectl --namespace=xlou get ingress -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- ---------------------------- Deleting pvc ---------------------------- [loop_until]: kubectl --namespace=xlou get pvc -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- [loop_until]: kubectl --namespace=xlou delete pv ds-backup-xlou --ignore-not-found [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- warning: deleting cluster-scoped resources, not scoped to the provided namespace ----------------- Deleting admin clusterrolebindings ----------------- [loop_until]: kubectl get clusterrolebinding -o jsonpath="{range .items[?(@.subjects[0].namespace=='xlou')]}{.metadata.name} {end}" [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- ------------------------- Deleting namespace ------------------------- [loop_until]: kubectl delete namespaces xlou --ignore-not-found --grace-period=0 --force [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. [loop_until]: awk -F" " "{print NF}" <<< `kubectl get namespace xlou --ignore-not-found` | grep 0 [loop_until]: (max_time=600, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- --- stderr --- [loop_until]: kubectl create namespace xlou [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- namespace/xlou created --- stderr --- [loop_until]: kubectl label namespace xlou self-service=false timeout=48 [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- namespace/xlou labeled --- stderr --- ************************************ Configuring components ************************************ No custom config provided. Nothing to do. No custom features provided. Nothing to do. ---- Updating components image tag/repo from platform-images repo ---- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --clean [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Cleaning up. [WARNING] Found nothing to clean. --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref sustaining/7.3.x --products ds [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Repo is at 9777e59ff49e0bfc417da81c0ee2b528cbae330c on branch sustaining/7.3.x [INFO] Updating products ds [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/dsutil/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/ds-new/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/proxy/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo/Dockerfile to use gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref sustaining/7.3.x --products am [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Found existing files, attempting to not clone. [INFO] Repo is at 9777e59ff49e0bfc417da81c0ee2b528cbae330c on branch sustaining/7.3.x [INFO] Updating products am [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am/Dockerfile to use gcr.io/forgerock-io/am-cdk/pit1:7.3.0-d0e0f0f08d45c97e1331f11b6b2ae9fc31d5e28d --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref sustaining/7.3.x --products amster [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Found existing files, attempting to not clone. [INFO] Repo is at 9777e59ff49e0bfc417da81c0ee2b528cbae330c on branch sustaining/7.3.x [INFO] Updating products amster [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster/Dockerfile to use gcr.io/forgerock-io/amster/pit1:7.3.0-d0e0f0f08d45c97e1331f11b6b2ae9fc31d5e28d --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref sustaining/7.3.x --products idm [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Found existing files, attempting to not clone. [INFO] Repo is at 9777e59ff49e0bfc417da81c0ee2b528cbae330c on branch sustaining/7.3.x [INFO] Updating products idm [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm/Dockerfile to use gcr.io/forgerock-io/idm-cdk/pit1:7.3.0-ed278902ceea12882b9d14775050c5120defecb9 --- stderr --- [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/set-images --ref sustaining/7.3.x --products ui [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- [INFO] Setting repo up in /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/build/platform-images. [INFO] Found existing files, attempting to not clone. [INFO] Repo is at 9777e59ff49e0bfc417da81c0ee2b528cbae330c on branch sustaining/7.3.x [INFO] Updating products ui [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui to use gcr.io/forgerock-io/platform-admin-ui/docker-build:7.2.0-1dec1ed7c0e3192e8afe46c4c8239c7108414857 [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui to use gcr.io/forgerock-io/platform-enduser-ui/docker-build:7.2.0-1dec1ed7c0e3192e8afe46c4c8239c7108414857 [INFO] Updating /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui to use gcr.io/forgerock-io/platform-login-ui/docker-build:7.2.0-1dec1ed7c0e3192e8afe46c4c8239c7108414857 --- stderr --- - Checking if component Dockerfile/kustomize needs additional update - [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker ds cts [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts --- stderr --- Checking if ds-cts dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts/Dockerfile: gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/cts/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker ds idrepo [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo --- stderr --- Checking if ds-idrepo dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo/Dockerfile: gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/ds/idrepo/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker am [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am --- stderr --- Checking if am dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am/Dockerfile: gcr.io/forgerock-io/am-cdk/pit1:7.3.0-d0e0f0f08d45c97e1331f11b6b2ae9fc31d5e28d No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/am/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker amster [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster --- stderr --- Checking if amster dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster/Dockerfile: gcr.io/forgerock-io/amster/pit1:7.3.0-d0e0f0f08d45c97e1331f11b6b2ae9fc31d5e28d No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/amster/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path docker idm [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm --- stderr --- Checking if idm dockerfile /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm/Dockerfile needs additional update (custom tag/repo from config.yaml or image name resolution) Read FROM line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm/Dockerfile: gcr.io/forgerock-io/idm-cdk/pit1:7.3.0-ed278902ceea12882b9d14775050c5120defecb9 No need to update FROM line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/docker/idm/Dockerfile [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path kustomize base end-user-ui [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui --- stderr --- Checking if end-user-ui kustomize file /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui/kustomization.yaml needs additional update (custom tag/repo from config.yaml or image name resolution) Read image line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui/kustomization.yaml: gcr.io/forgerock-io/platform-enduser-ui/docker-build:7.2.0-1dec1ed7c0e3192e8afe46c4c8239c7108414857 No need to update the image line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/end-user-ui/kustomization.yaml [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path kustomize base login-ui [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui --- stderr --- Checking if login-ui kustomize file /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui/kustomization.yaml needs additional update (custom tag/repo from config.yaml or image name resolution) Read image line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui/kustomization.yaml: gcr.io/forgerock-io/platform-login-ui/docker-build:7.2.0-1dec1ed7c0e3192e8afe46c4c8239c7108414857 No need to update the image line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/login-ui/kustomization.yaml [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path kustomize base admin-ui [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui --- stderr --- Checking if admin-ui kustomize file /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui/kustomization.yaml needs additional update (custom tag/repo from config.yaml or image name resolution) Read image line from /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui/kustomization.yaml: gcr.io/forgerock-io/platform-admin-ui/docker-build:7.2.0-1dec1ed7c0e3192e8afe46c4c8239c7108414857 No need to update the image line for /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/base/admin-ui/kustomization.yaml [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/config path kustomize overlay medium [run_command]: OK (rc = 0 - expected to be in [0]) --- stdout --- /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/overlay/medium --- stderr --- [loop_until]: kubectl --namespace=xlou delete -f /tmp/tmpmnwv83sv [loop_until]: (max_time=180, interval=5, expected_rc=[0, 1] [loop_until]: OK (rc = 1) --- stdout --- --- stderr --- Error from server (NotFound): error when deleting "/tmp/tmpmnwv83sv": secrets "sslcert" not found [loop_until]: kubectl --namespace=xlou apply -f /tmp/tmpmnwv83sv [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- secret/sslcert created --- stderr --- The following components will be deployed: - ds-cts (DS) - ds-idrepo (DS) - am (AM) - amster (Amster) - idm (IDM) - end-user-ui (EndUserUi) - login-ui (LoginUi) - admin-ui (AdminUi) [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/forgeops build all --config-profile=cdk --push-to gcr.io/engineeringpit/lodestar-images --tag=xlou [run_command]: env={'HOME': '/home/xslou', 'PATH': '/mnt/disks/data/xslou/lodestar-fork/ext/bin:/home/xslou/.local/bin:/home/xslou/bin:~/bin:~/workshop/lodestar-fork/ext/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin', 'DOCKER_SCAN_SUGGEST': 'false'} Sending build context to Docker daemon 10.24kB Step 1/6 : FROM gcr.io/forgerock-io/am-cdk/pit1:7.3.0-d0e0f0f08d45c97e1331f11b6b2ae9fc31d5e28d 7.3.0-d0e0f0f08d45c97e1331f11b6b2ae9fc31d5e28d: Pulling from forgerock-io/am-cdk/pit1 74ac377868f8: Pulling fs layer b9cabe75b440: Pulling fs layer cff6df0cca35: Pulling fs layer 342f22ed35dc: Pulling fs layer 5337ce44e879: Pulling fs layer f87847a136f0: Pulling fs layer 0dc924ac702d: Pulling fs layer 2eaa8c835f5b: Pulling fs layer 4067e81c4a28: Pulling fs layer 40c9a643934d: Pulling fs layer 315476dc1f68: Pulling fs layer bb1872ff5dcd: Pulling fs layer afc37b3ccef9: Pulling fs layer 56802119ef29: Pulling fs layer e1bea1b8f7d6: Pulling fs layer 035fd063e2d4: Pulling fs layer a63d5117412c: Pulling fs layer 5ec97bbd2749: Pulling fs layer f8014230425a: Pulling fs layer 5466137d3132: Pulling fs layer a5bc11d60f74: Pulling fs layer 1945fa1736b5: Pulling fs layer 6ca22b83a41e: Pulling fs layer 17a704cdb8b9: Pulling fs layer 51e25bbfd7fa: Pulling fs layer c5b1230e6204: Pulling fs layer f014a46844a5: Pulling fs layer 7df5530c55d8: Pulling fs layer a4c1ff1581b9: Pulling fs layer c6eda0d72ed5: Pulling fs layer d278dc3813bb: Pulling fs layer 93a39aa039c6: Pulling fs layer c21b034ff2d9: Pulling fs layer 1782132001ca: Pulling fs layer 2eaa8c835f5b: Waiting 4067e81c4a28: Waiting 40c9a643934d: Waiting 315476dc1f68: Waiting bb1872ff5dcd: Waiting afc37b3ccef9: Waiting 56802119ef29: Waiting e1bea1b8f7d6: Waiting 035fd063e2d4: Waiting a63d5117412c: Waiting 5ec97bbd2749: Waiting f8014230425a: Waiting 5466137d3132: Waiting a5bc11d60f74: Waiting 1945fa1736b5: Waiting 6ca22b83a41e: Waiting 17a704cdb8b9: Waiting 51e25bbfd7fa: Waiting c5b1230e6204: Waiting f014a46844a5: Waiting 7df5530c55d8: Waiting a4c1ff1581b9: Waiting c6eda0d72ed5: Waiting d278dc3813bb: Waiting 93a39aa039c6: Waiting c21b034ff2d9: Waiting 1782132001ca: Waiting 342f22ed35dc: Waiting 5337ce44e879: Waiting f87847a136f0: Waiting 0dc924ac702d: Waiting b9cabe75b440: Verifying Checksum b9cabe75b440: Download complete 342f22ed35dc: Verifying Checksum 342f22ed35dc: Download complete 74ac377868f8: Verifying Checksum 74ac377868f8: Download complete 5337ce44e879: Verifying Checksum 5337ce44e879: Download complete 0dc924ac702d: Verifying Checksum 0dc924ac702d: Download complete 2eaa8c835f5b: Verifying Checksum 2eaa8c835f5b: Download complete f87847a136f0: Verifying Checksum f87847a136f0: Download complete cff6df0cca35: Verifying Checksum cff6df0cca35: Download complete 315476dc1f68: Verifying Checksum 315476dc1f68: Download complete 40c9a643934d: Verifying Checksum 40c9a643934d: Download complete bb1872ff5dcd: Verifying Checksum bb1872ff5dcd: Download complete afc37b3ccef9: Verifying Checksum afc37b3ccef9: Download complete 56802119ef29: Verifying Checksum 56802119ef29: Download complete e1bea1b8f7d6: Verifying Checksum e1bea1b8f7d6: Download complete 4067e81c4a28: Verifying Checksum 4067e81c4a28: Download complete 74ac377868f8: Pull complete a63d5117412c: Verifying Checksum a63d5117412c: Download complete 5ec97bbd2749: Verifying Checksum 5ec97bbd2749: Download complete f8014230425a: Verifying Checksum f8014230425a: Download complete 5466137d3132: Verifying Checksum 5466137d3132: Download complete a5bc11d60f74: Verifying Checksum a5bc11d60f74: Download complete 1945fa1736b5: Verifying Checksum 1945fa1736b5: Download complete 6ca22b83a41e: Verifying Checksum 6ca22b83a41e: Download complete 51e25bbfd7fa: Verifying Checksum 51e25bbfd7fa: Download complete b9cabe75b440: Pull complete c5b1230e6204: Verifying Checksum c5b1230e6204: Download complete f014a46844a5: Verifying Checksum f014a46844a5: Download complete 17a704cdb8b9: Verifying Checksum 17a704cdb8b9: Download complete 7df5530c55d8: Verifying Checksum 7df5530c55d8: Download complete a4c1ff1581b9: Verifying Checksum a4c1ff1581b9: Download complete d278dc3813bb: Verifying Checksum d278dc3813bb: Download complete 93a39aa039c6: Verifying Checksum 93a39aa039c6: Download complete c21b034ff2d9: Verifying Checksum c21b034ff2d9: Download complete 1782132001ca: Verifying Checksum 1782132001ca: Download complete c6eda0d72ed5: Verifying Checksum c6eda0d72ed5: Download complete 035fd063e2d4: Verifying Checksum 035fd063e2d4: Download complete cff6df0cca35: Pull complete 342f22ed35dc: Pull complete 5337ce44e879: Pull complete f87847a136f0: Pull complete 0dc924ac702d: Pull complete 2eaa8c835f5b: Pull complete 4067e81c4a28: Pull complete 40c9a643934d: Pull complete 315476dc1f68: Pull complete bb1872ff5dcd: Pull complete afc37b3ccef9: Pull complete 56802119ef29: Pull complete e1bea1b8f7d6: Pull complete 035fd063e2d4: Pull complete a63d5117412c: Pull complete 5ec97bbd2749: Pull complete f8014230425a: Pull complete 5466137d3132: Pull complete a5bc11d60f74: Pull complete 1945fa1736b5: Pull complete 6ca22b83a41e: Pull complete 17a704cdb8b9: Pull complete 51e25bbfd7fa: Pull complete c5b1230e6204: Pull complete f014a46844a5: Pull complete 7df5530c55d8: Pull complete a4c1ff1581b9: Pull complete c6eda0d72ed5: Pull complete d278dc3813bb: Pull complete 93a39aa039c6: Pull complete c21b034ff2d9: Pull complete 1782132001ca: Pull complete Digest: sha256:1aaa269131c07568bfc7589807b9fe1c158208f98402f923cfed5765890ed15d Status: Downloaded newer image for gcr.io/forgerock-io/am-cdk/pit1:7.3.0-d0e0f0f08d45c97e1331f11b6b2ae9fc31d5e28d ---> 7fde2db0c1a7 Step 2/6 : ARG CONFIG_PROFILE=cdk ---> Running in 664979fac0c7 Removing intermediate container 664979fac0c7 ---> 3b6bafb9dd96 Step 3/6 : RUN echo "\033[0;36m*** Building '${CONFIG_PROFILE}' profile ***\033[0m" ---> Running in 6a258195c529 *** Building 'cdk' profile *** Removing intermediate container 6a258195c529 ---> e061dbf289d4 Step 4/6 : COPY --chown=forgerock:root config-profiles/${CONFIG_PROFILE}/ /home/forgerock/openam/ ---> a15461410049 Step 5/6 : COPY --chown=forgerock:root *.sh /home/forgerock/ ---> bdceaa346236 Step 6/6 : WORKDIR /home/forgerock ---> Running in 5bfe11b7c665 Removing intermediate container 5bfe11b7c665 ---> 4616b0c0f5b5 Successfully built 4616b0c0f5b5 Successfully tagged gcr.io/engineeringpit/lodestar-images/am:xlou The push refers to repository [gcr.io/engineeringpit/lodestar-images/am] 0960b22180ad: Preparing b887d792e734: Preparing c57088fdcee3: Preparing d4ad75fa2fdc: Preparing a0c5569fc01c: Preparing a944cfc93389: Preparing 0d41684213f1: Preparing a6bf82e5f30e: Preparing f4e8ea536b52: Preparing 0c36748cae3b: Preparing a58435e83fb9: Preparing dd3634e8d393: Preparing 2d12481c2ea9: Preparing 87be3a900d8b: Preparing 7e80f3ea9489: Preparing b72448ecede0: Preparing 978097307a8b: Preparing 433373b25f31: Preparing 3af12dc1cc36: Preparing f09f443715a2: Preparing 2a26d579527d: Preparing 5477d6949583: Preparing a292c25c2f9a: Preparing 80818bbf4b5d: Preparing 0fa2950fff43: Preparing 691ea1d866ef: Preparing 82f4367b1510: Preparing 1ddb90d48e77: Preparing 16979ca3c8a2: Preparing df948e73a878: Preparing 9af81ef986f7: Preparing 4d02a34d0051: Preparing 749d6cfaa5eb: Preparing 8ac55e7a3c46: Preparing 04d1dcab20cb: Preparing b93c1bd012ab: Preparing a944cfc93389: Waiting 0d41684213f1: Waiting a6bf82e5f30e: Waiting f4e8ea536b52: Waiting 0c36748cae3b: Waiting a58435e83fb9: Waiting dd3634e8d393: Waiting 2d12481c2ea9: Waiting 87be3a900d8b: Waiting 7e80f3ea9489: Waiting b72448ecede0: Waiting 978097307a8b: Waiting 433373b25f31: Waiting 3af12dc1cc36: Waiting f09f443715a2: Waiting 2a26d579527d: Waiting 5477d6949583: Waiting a292c25c2f9a: Waiting 80818bbf4b5d: Waiting 0fa2950fff43: Waiting 691ea1d866ef: Waiting 82f4367b1510: Waiting 1ddb90d48e77: Waiting 16979ca3c8a2: Waiting df948e73a878: Waiting 9af81ef986f7: Waiting 4d02a34d0051: Waiting 749d6cfaa5eb: Waiting 8ac55e7a3c46: Waiting 04d1dcab20cb: Waiting b93c1bd012ab: Waiting c57088fdcee3: Layer already exists a0c5569fc01c: Layer already exists d4ad75fa2fdc: Layer already exists a944cfc93389: Layer already exists 0d41684213f1: Layer already exists a6bf82e5f30e: Layer already exists f4e8ea536b52: Layer already exists 0c36748cae3b: Layer already exists a58435e83fb9: Layer already exists dd3634e8d393: Layer already exists 2d12481c2ea9: Layer already exists 87be3a900d8b: Layer already exists 7e80f3ea9489: Layer already exists b72448ecede0: Layer already exists 978097307a8b: Layer already exists 433373b25f31: Layer already exists 3af12dc1cc36: Layer already exists f09f443715a2: Layer already exists 2a26d579527d: Layer already exists 5477d6949583: Layer already exists a292c25c2f9a: Layer already exists 80818bbf4b5d: Layer already exists 0fa2950fff43: Layer already exists 691ea1d866ef: Layer already exists 82f4367b1510: Layer already exists 1ddb90d48e77: Layer already exists 16979ca3c8a2: Layer already exists df948e73a878: Layer already exists 9af81ef986f7: Layer already exists 4d02a34d0051: Layer already exists 749d6cfaa5eb: Layer already exists 8ac55e7a3c46: Layer already exists 04d1dcab20cb: Layer already exists b93c1bd012ab: Layer already exists b887d792e734: Pushed 0960b22180ad: Pushed xlou: digest: sha256:fae86102dd8374e78ac5d98db56845fe98742e7aebebd5445a81ba84bfa5eed5 size: 7843 Sending build context to Docker daemon 316.4kB Step 1/8 : FROM gcr.io/forgerock-io/idm-cdk/pit1:7.3.0-ed278902ceea12882b9d14775050c5120defecb9 7.3.0-ed278902ceea12882b9d14775050c5120defecb9: Pulling from forgerock-io/idm-cdk/pit1 3689b8de819b: Pulling fs layer fa5b059e0b99: Pulling fs layer 617d05562c0c: Pulling fs layer c241d9cc0a4c: Pulling fs layer 757aaba62692: Pulling fs layer 66332df398d2: Pulling fs layer 5e7c1e6e5e34: Pulling fs layer 4f4fb700ef54: Pulling fs layer 757aaba62692: Waiting 66332df398d2: Waiting 5e7c1e6e5e34: Waiting 4f4fb700ef54: Waiting c241d9cc0a4c: Waiting 617d05562c0c: Verifying Checksum 617d05562c0c: Download complete 3689b8de819b: Verifying Checksum 3689b8de819b: Download complete c241d9cc0a4c: Verifying Checksum c241d9cc0a4c: Download complete 757aaba62692: Verifying Checksum 757aaba62692: Download complete 5e7c1e6e5e34: Verifying Checksum 5e7c1e6e5e34: Download complete 4f4fb700ef54: Verifying Checksum 4f4fb700ef54: Download complete fa5b059e0b99: Verifying Checksum fa5b059e0b99: Download complete 3689b8de819b: Pull complete fa5b059e0b99: Pull complete 617d05562c0c: Pull complete c241d9cc0a4c: Pull complete 757aaba62692: Pull complete 66332df398d2: Verifying Checksum 66332df398d2: Download complete 66332df398d2: Pull complete 5e7c1e6e5e34: Pull complete 4f4fb700ef54: Pull complete Digest: sha256:cdeaa966c6d8a59cd7da18e762f4d34e3715d8f8f461e3c6dadebde247bdf6c7 Status: Downloaded newer image for gcr.io/forgerock-io/idm-cdk/pit1:7.3.0-ed278902ceea12882b9d14775050c5120defecb9 ---> a65af4c6b1fa Step 2/8 : COPY debian-buster-sources.list /etc/apt/sources.list ---> f35fd9042fbb Step 3/8 : RUN rm -f bundle/org.apache.felix.webconsole*.jar && rm -f bundle/openidm-felix-webconsole-*.jar ---> Running in 3e7ae6c8a553 Removing intermediate container 3e7ae6c8a553 ---> b3351455c0aa Step 4/8 : ENV JAVA_OPTS -XX:MaxRAMPercentage=65 -XX:InitialRAMPercentage=65 -XX:MaxTenuringThreshold=1 -Djava.security.egd=file:/dev/urandom -XshowSettings:vm -XX:+PrintFlagsFinal ---> Running in 9bcd43ce542b Removing intermediate container 9bcd43ce542b ---> 80244a39bd26 Step 5/8 : ARG CONFIG_PROFILE=cdk ---> Running in c32aa6db77d4 Removing intermediate container c32aa6db77d4 ---> be6b8a08193f Step 6/8 : RUN echo "\033[0;36m*** Building '${CONFIG_PROFILE}' profile ***\033[0m" ---> Running in c427a05b8517 *** Building 'cdk' profile *** Removing intermediate container c427a05b8517 ---> 45b799b85c39 Step 7/8 : COPY --chown=forgerock:root config-profiles/${CONFIG_PROFILE}/ /opt/openidm ---> 70676c2c2db5 Step 8/8 : COPY --chown=forgerock:root . /opt/openidm ---> d94f706dc9fe Successfully built d94f706dc9fe Successfully tagged gcr.io/engineeringpit/lodestar-images/idm:xlou The push refers to repository [gcr.io/engineeringpit/lodestar-images/idm] f0f47c91bb0f: Preparing ed3fa3b9ee05: Preparing 8db55b71b5f9: Preparing 5f70bf18a086: Preparing 38fe551f2463: Preparing 68884240a8b2: Preparing 1ae67d4978a1: Preparing 8d20f850e8e0: Preparing 09308431d152: Preparing fd535567db7f: Preparing c9182c130984: Preparing 1ae67d4978a1: Waiting 8d20f850e8e0: Waiting 09308431d152: Waiting fd535567db7f: Waiting c9182c130984: Waiting 68884240a8b2: Waiting 38fe551f2463: Layer already exists 5f70bf18a086: Layer already exists 68884240a8b2: Layer already exists 1ae67d4978a1: Layer already exists 8d20f850e8e0: Layer already exists 09308431d152: Layer already exists fd535567db7f: Layer already exists c9182c130984: Layer already exists ed3fa3b9ee05: Pushed f0f47c91bb0f: Pushed 8db55b71b5f9: Pushed xlou: digest: sha256:65ef8394bbe3e47615f5a8f3854cb7d3e012702151c41a59dd937fa232344ec5 size: 2621 Sending build context to Docker daemon 128.5kB Step 1/11 : FROM gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309: Pulling from forgerock-io/ds/pit1 f1f26f570256: Pulling fs layer 27e9bff78739: Pulling fs layer b159bdbf9a6d: Pulling fs layer ac5463df2dd9: Pulling fs layer ace0dbce7cf9: Pulling fs layer 7e6c5b0fc828: Pulling fs layer 4f4fb700ef54: Pulling fs layer a349b087d326: Pulling fs layer ace0dbce7cf9: Waiting 7e6c5b0fc828: Waiting 4f4fb700ef54: Waiting a349b087d326: Waiting ac5463df2dd9: Waiting b159bdbf9a6d: Verifying Checksum b159bdbf9a6d: Download complete f1f26f570256: Verifying Checksum f1f26f570256: Download complete ac5463df2dd9: Verifying Checksum ac5463df2dd9: Download complete ace0dbce7cf9: Verifying Checksum ace0dbce7cf9: Download complete 4f4fb700ef54: Verifying Checksum 4f4fb700ef54: Download complete 27e9bff78739: Verifying Checksum 27e9bff78739: Download complete a349b087d326: Verifying Checksum a349b087d326: Download complete 7e6c5b0fc828: Verifying Checksum 7e6c5b0fc828: Download complete f1f26f570256: Pull complete 27e9bff78739: Pull complete b159bdbf9a6d: Pull complete ac5463df2dd9: Pull complete ace0dbce7cf9: Pull complete 7e6c5b0fc828: Pull complete 4f4fb700ef54: Pull complete a349b087d326: Pull complete Digest: sha256:7ccc10d8dd41b2615dd0fea7dbd5658f9ef06ca8ec929960646e226e7503d559 Status: Downloaded newer image for gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 ---> 9deba3244b26 Step 2/11 : USER root ---> Running in b226cfa31e52 Removing intermediate container b226cfa31e52 ---> 940f8554f1f4 Step 3/11 : RUN apt-get update && apt-get install -y --no-install-recommends vim ncat dnsutils ---> Running in f0f133d12a16 Get:1 http://deb.debian.org/debian bullseye InRelease [116 kB] Get:2 http://deb.debian.org/debian-security bullseye-security InRelease [48.4 kB] Get:3 http://deb.debian.org/debian bullseye-updates InRelease [44.1 kB] Get:4 http://deb.debian.org/debian bullseye/main amd64 Packages [8183 kB] Get:5 http://deb.debian.org/debian-security bullseye-security/main amd64 Packages [236 kB] Get:6 http://deb.debian.org/debian bullseye-updates/main amd64 Packages [14.6 kB] Fetched 8642 kB in 1s (6175 kB/s) Reading package lists... Reading package lists... Building dependency tree... Reading state information... The following additional packages will be installed: bind9-dnsutils bind9-host bind9-libs libbsd0 libdbus-1-3 libedit2 libfstrm0 libicu67 libjson-c5 liblmdb0 liblua5.3-0 libmaxminddb0 libmd0 libpcap0.8 libprotobuf-c1 libuv1 libxml2 vim-common vim-runtime xxd Suggested packages: mmdb-bin ctags vim-doc vim-scripts Recommended packages: dbus The following NEW packages will be installed: bind9-dnsutils bind9-host bind9-libs dnsutils libbsd0 libdbus-1-3 libedit2 libfstrm0 libicu67 libjson-c5 liblmdb0 liblua5.3-0 libmaxminddb0 libmd0 libpcap0.8 libprotobuf-c1 libuv1 libxml2 ncat vim vim-common vim-runtime xxd 0 upgraded, 23 newly installed, 0 to remove and 1 not upgraded. Need to get 21.3 MB of archives. After this operation, 81.2 MB of additional disk space will be used. Get:1 http://deb.debian.org/debian bullseye/main amd64 xxd amd64 2:8.2.2434-3+deb11u1 [192 kB] Get:2 http://deb.debian.org/debian bullseye/main amd64 vim-common all 2:8.2.2434-3+deb11u1 [226 kB] Get:3 http://deb.debian.org/debian bullseye/main amd64 libuv1 amd64 1.40.0-2 [132 kB] Get:4 http://deb.debian.org/debian bullseye/main amd64 libfstrm0 amd64 0.6.0-1+b1 [21.5 kB] Get:5 http://deb.debian.org/debian bullseye/main amd64 libjson-c5 amd64 0.15-2 [42.8 kB] Get:6 http://deb.debian.org/debian bullseye/main amd64 liblmdb0 amd64 0.9.24-1 [45.0 kB] Get:7 http://deb.debian.org/debian bullseye/main amd64 libmaxminddb0 amd64 1.5.2-1 [29.8 kB] Get:8 http://deb.debian.org/debian bullseye/main amd64 libprotobuf-c1 amd64 1.3.3-1+b2 [27.0 kB] Get:9 http://deb.debian.org/debian bullseye/main amd64 libicu67 amd64 67.1-7 [8622 kB] Get:10 http://deb.debian.org/debian bullseye/main amd64 libxml2 amd64 2.9.10+dfsg-6.7+deb11u3 [693 kB] Get:11 http://deb.debian.org/debian-security bullseye-security/main amd64 bind9-libs amd64 1:9.16.37-1~deb11u1 [1424 kB] Get:12 http://deb.debian.org/debian-security bullseye-security/main amd64 bind9-host amd64 1:9.16.37-1~deb11u1 [308 kB] Get:13 http://deb.debian.org/debian bullseye/main amd64 libmd0 amd64 1.0.3-3 [28.0 kB] Get:14 http://deb.debian.org/debian bullseye/main amd64 libbsd0 amd64 0.11.3-1 [108 kB] Get:15 http://deb.debian.org/debian bullseye/main amd64 libedit2 amd64 3.1-20191231-2+b1 [96.7 kB] Get:16 http://deb.debian.org/debian-security bullseye-security/main amd64 bind9-dnsutils amd64 1:9.16.37-1~deb11u1 [404 kB] Get:17 http://deb.debian.org/debian-security bullseye-security/main amd64 dnsutils all 1:9.16.37-1~deb11u1 [267 kB] Get:18 http://deb.debian.org/debian bullseye/main amd64 libdbus-1-3 amd64 1.12.24-0+deb11u1 [222 kB] Get:19 http://deb.debian.org/debian bullseye/main amd64 liblua5.3-0 amd64 5.3.3-1.1+b1 [120 kB] Get:20 http://deb.debian.org/debian bullseye/main amd64 libpcap0.8 amd64 1.10.0-2 [159 kB] Get:21 http://deb.debian.org/debian bullseye/main amd64 ncat amd64 7.91+dfsg1+really7.80+dfsg1-2 [383 kB] Get:22 http://deb.debian.org/debian bullseye/main amd64 vim-runtime all 2:8.2.2434-3+deb11u1 [6226 kB] Get:23 http://deb.debian.org/debian bullseye/main amd64 vim amd64 2:8.2.2434-3+deb11u1 [1494 kB] debconf: delaying package configuration, since apt-utils is not installed Fetched 21.3 MB in 0s (99.9 MB/s) Selecting previously unselected package xxd. (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 7977 files and directories currently installed.) Preparing to unpack .../00-xxd_2%3a8.2.2434-3+deb11u1_amd64.deb ... Unpacking xxd (2:8.2.2434-3+deb11u1) ... Selecting previously unselected package vim-common. Preparing to unpack .../01-vim-common_2%3a8.2.2434-3+deb11u1_all.deb ... Unpacking vim-common (2:8.2.2434-3+deb11u1) ... Selecting previously unselected package libuv1:amd64. Preparing to unpack .../02-libuv1_1.40.0-2_amd64.deb ... Unpacking libuv1:amd64 (1.40.0-2) ... Selecting previously unselected package libfstrm0:amd64. Preparing to unpack .../03-libfstrm0_0.6.0-1+b1_amd64.deb ... Unpacking libfstrm0:amd64 (0.6.0-1+b1) ... Selecting previously unselected package libjson-c5:amd64. Preparing to unpack .../04-libjson-c5_0.15-2_amd64.deb ... Unpacking libjson-c5:amd64 (0.15-2) ... Selecting previously unselected package liblmdb0:amd64. Preparing to unpack .../05-liblmdb0_0.9.24-1_amd64.deb ... Unpacking liblmdb0:amd64 (0.9.24-1) ... Selecting previously unselected package libmaxminddb0:amd64. Preparing to unpack .../06-libmaxminddb0_1.5.2-1_amd64.deb ... Unpacking libmaxminddb0:amd64 (1.5.2-1) ... Selecting previously unselected package libprotobuf-c1:amd64. Preparing to unpack .../07-libprotobuf-c1_1.3.3-1+b2_amd64.deb ... Unpacking libprotobuf-c1:amd64 (1.3.3-1+b2) ... Selecting previously unselected package libicu67:amd64. Preparing to unpack .../08-libicu67_67.1-7_amd64.deb ... Unpacking libicu67:amd64 (67.1-7) ... Selecting previously unselected package libxml2:amd64. Preparing to unpack .../09-libxml2_2.9.10+dfsg-6.7+deb11u3_amd64.deb ... Unpacking libxml2:amd64 (2.9.10+dfsg-6.7+deb11u3) ... Selecting previously unselected package bind9-libs:amd64. Preparing to unpack .../10-bind9-libs_1%3a9.16.37-1~deb11u1_amd64.deb ... Unpacking bind9-libs:amd64 (1:9.16.37-1~deb11u1) ... Selecting previously unselected package bind9-host. Preparing to unpack .../11-bind9-host_1%3a9.16.37-1~deb11u1_amd64.deb ... Unpacking bind9-host (1:9.16.37-1~deb11u1) ... Selecting previously unselected package libmd0:amd64. Preparing to unpack .../12-libmd0_1.0.3-3_amd64.deb ... Unpacking libmd0:amd64 (1.0.3-3) ... Selecting previously unselected package libbsd0:amd64. Preparing to unpack .../13-libbsd0_0.11.3-1_amd64.deb ... Unpacking libbsd0:amd64 (0.11.3-1) ... Selecting previously unselected package libedit2:amd64. Preparing to unpack .../14-libedit2_3.1-20191231-2+b1_amd64.deb ... Unpacking libedit2:amd64 (3.1-20191231-2+b1) ... Selecting previously unselected package bind9-dnsutils. Preparing to unpack .../15-bind9-dnsutils_1%3a9.16.37-1~deb11u1_amd64.deb ... Unpacking bind9-dnsutils (1:9.16.37-1~deb11u1) ... Selecting previously unselected package dnsutils. Preparing to unpack .../16-dnsutils_1%3a9.16.37-1~deb11u1_all.deb ... Unpacking dnsutils (1:9.16.37-1~deb11u1) ... Selecting previously unselected package libdbus-1-3:amd64. Preparing to unpack .../17-libdbus-1-3_1.12.24-0+deb11u1_amd64.deb ... Unpacking libdbus-1-3:amd64 (1.12.24-0+deb11u1) ... Selecting previously unselected package liblua5.3-0:amd64. Preparing to unpack .../18-liblua5.3-0_5.3.3-1.1+b1_amd64.deb ... Unpacking liblua5.3-0:amd64 (5.3.3-1.1+b1) ... Selecting previously unselected package libpcap0.8:amd64. Preparing to unpack .../19-libpcap0.8_1.10.0-2_amd64.deb ... Unpacking libpcap0.8:amd64 (1.10.0-2) ... Selecting previously unselected package ncat. Preparing to unpack .../20-ncat_7.91+dfsg1+really7.80+dfsg1-2_amd64.deb ... Unpacking ncat (7.91+dfsg1+really7.80+dfsg1-2) ... Selecting previously unselected package vim-runtime. Preparing to unpack .../21-vim-runtime_2%3a8.2.2434-3+deb11u1_all.deb ... Adding 'diversion of /usr/share/vim/vim82/doc/help.txt to /usr/share/vim/vim82/doc/help.txt.vim-tiny by vim-runtime' Adding 'diversion of /usr/share/vim/vim82/doc/tags to /usr/share/vim/vim82/doc/tags.vim-tiny by vim-runtime' Unpacking vim-runtime (2:8.2.2434-3+deb11u1) ... Selecting previously unselected package vim. Preparing to unpack .../22-vim_2%3a8.2.2434-3+deb11u1_amd64.deb ... Unpacking vim (2:8.2.2434-3+deb11u1) ... Setting up liblmdb0:amd64 (0.9.24-1) ... Setting up libicu67:amd64 (67.1-7) ... Setting up libmaxminddb0:amd64 (1.5.2-1) ... Setting up libfstrm0:amd64 (0.6.0-1+b1) ... Setting up libprotobuf-c1:amd64 (1.3.3-1+b2) ... Setting up xxd (2:8.2.2434-3+deb11u1) ... Setting up libuv1:amd64 (1.40.0-2) ... Setting up vim-common (2:8.2.2434-3+deb11u1) ... Setting up libdbus-1-3:amd64 (1.12.24-0+deb11u1) ... Setting up libmd0:amd64 (1.0.3-3) ... Setting up liblua5.3-0:amd64 (5.3.3-1.1+b1) ... Setting up vim-runtime (2:8.2.2434-3+deb11u1) ... Setting up libbsd0:amd64 (0.11.3-1) ... Setting up libjson-c5:amd64 (0.15-2) ... Setting up libxml2:amd64 (2.9.10+dfsg-6.7+deb11u3) ... Setting up vim (2:8.2.2434-3+deb11u1) ... update-alternatives: using /usr/bin/vim.basic to provide /usr/bin/vim (vim) in auto mode update-alternatives: using /usr/bin/vim.basic to provide /usr/bin/vimdiff (vimdiff) in auto mode update-alternatives: using /usr/bin/vim.basic to provide /usr/bin/rvim (rvim) in auto mode update-alternatives: using /usr/bin/vim.basic to provide /usr/bin/rview (rview) in auto mode update-alternatives: using /usr/bin/vim.basic to provide /usr/bin/vi (vi) in auto mode update-alternatives: warning: skip creation of /usr/share/man/da/man1/vi.1.gz because associated file /usr/share/man/da/man1/vim.1.gz (of link group vi) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/de/man1/vi.1.gz because associated file /usr/share/man/de/man1/vim.1.gz (of link group vi) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/fr/man1/vi.1.gz because associated file /usr/share/man/fr/man1/vim.1.gz (of link group vi) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/it/man1/vi.1.gz because associated file /usr/share/man/it/man1/vim.1.gz (of link group vi) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/ja/man1/vi.1.gz because associated file /usr/share/man/ja/man1/vim.1.gz (of link group vi) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/pl/man1/vi.1.gz because associated file /usr/share/man/pl/man1/vim.1.gz (of link group vi) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/ru/man1/vi.1.gz because associated file /usr/share/man/ru/man1/vim.1.gz (of link group vi) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/man1/vi.1.gz because associated file /usr/share/man/man1/vim.1.gz (of link group vi) doesn't exist update-alternatives: using /usr/bin/vim.basic to provide /usr/bin/view (view) in auto mode update-alternatives: warning: skip creation of /usr/share/man/da/man1/view.1.gz because associated file /usr/share/man/da/man1/vim.1.gz (of link group view) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/de/man1/view.1.gz because associated file /usr/share/man/de/man1/vim.1.gz (of link group view) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/fr/man1/view.1.gz because associated file /usr/share/man/fr/man1/vim.1.gz (of link group view) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/it/man1/view.1.gz because associated file /usr/share/man/it/man1/vim.1.gz (of link group view) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/ja/man1/view.1.gz because associated file /usr/share/man/ja/man1/vim.1.gz (of link group view) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/pl/man1/view.1.gz because associated file /usr/share/man/pl/man1/vim.1.gz (of link group view) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/ru/man1/view.1.gz because associated file /usr/share/man/ru/man1/vim.1.gz (of link group view) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/man1/view.1.gz because associated file /usr/share/man/man1/vim.1.gz (of link group view) doesn't exist update-alternatives: using /usr/bin/vim.basic to provide /usr/bin/ex (ex) in auto mode update-alternatives: warning: skip creation of /usr/share/man/da/man1/ex.1.gz because associated file /usr/share/man/da/man1/vim.1.gz (of link group ex) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/de/man1/ex.1.gz because associated file /usr/share/man/de/man1/vim.1.gz (of link group ex) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/fr/man1/ex.1.gz because associated file /usr/share/man/fr/man1/vim.1.gz (of link group ex) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/it/man1/ex.1.gz because associated file /usr/share/man/it/man1/vim.1.gz (of link group ex) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/ja/man1/ex.1.gz because associated file /usr/share/man/ja/man1/vim.1.gz (of link group ex) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/pl/man1/ex.1.gz because associated file /usr/share/man/pl/man1/vim.1.gz (of link group ex) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/ru/man1/ex.1.gz because associated file /usr/share/man/ru/man1/vim.1.gz (of link group ex) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/man1/ex.1.gz because associated file /usr/share/man/man1/vim.1.gz (of link group ex) doesn't exist update-alternatives: using /usr/bin/vim.basic to provide /usr/bin/editor (editor) in auto mode update-alternatives: warning: skip creation of /usr/share/man/da/man1/editor.1.gz because associated file /usr/share/man/da/man1/vim.1.gz (of link group editor) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/de/man1/editor.1.gz because associated file /usr/share/man/de/man1/vim.1.gz (of link group editor) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/fr/man1/editor.1.gz because associated file /usr/share/man/fr/man1/vim.1.gz (of link group editor) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/it/man1/editor.1.gz because associated file /usr/share/man/it/man1/vim.1.gz (of link group editor) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/ja/man1/editor.1.gz because associated file /usr/share/man/ja/man1/vim.1.gz (of link group editor) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/pl/man1/editor.1.gz because associated file /usr/share/man/pl/man1/vim.1.gz (of link group editor) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/ru/man1/editor.1.gz because associated file /usr/share/man/ru/man1/vim.1.gz (of link group editor) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/man1/editor.1.gz because associated file /usr/share/man/man1/vim.1.gz (of link group editor) doesn't exist Setting up bind9-libs:amd64 (1:9.16.37-1~deb11u1) ... Setting up libedit2:amd64 (3.1-20191231-2+b1) ... Setting up libpcap0.8:amd64 (1.10.0-2) ... Setting up ncat (7.91+dfsg1+really7.80+dfsg1-2) ... update-alternatives: using /usr/bin/ncat to provide /bin/nc (nc) in auto mode update-alternatives: warning: skip creation of /usr/share/man/man1/nc.1.gz because associated file /usr/share/man/man1/ncat.1.gz (of link group nc) doesn't exist update-alternatives: warning: skip creation of /usr/share/man/man1/netcat.1.gz because associated file /usr/share/man/man1/ncat.1.gz (of link group nc) doesn't exist Setting up bind9-host (1:9.16.37-1~deb11u1) ... Setting up bind9-dnsutils (1:9.16.37-1~deb11u1) ... Setting up dnsutils (1:9.16.37-1~deb11u1) ... Processing triggers for libc-bin (2.31-13+deb11u5) ... Removing intermediate container f0f133d12a16 ---> cf9d5ab64a80 Step 4/11 : USER forgerock ---> Running in ee488c8e812c Removing intermediate container ee488c8e812c ---> 809bbce7913d Step 5/11 : ENV DS_DATA_DIR /opt/opendj/data ---> Running in c0afe7eee4dc Removing intermediate container c0afe7eee4dc ---> 2ee60fd844e4 Step 6/11 : ENV PEM_KEYS_DIRECTORY "/var/run/secrets/keys/ds" ---> Running in 35f1f7d4246d Removing intermediate container 35f1f7d4246d ---> 5144146c5a06 Step 7/11 : ENV PEM_TRUSTSTORE_DIRECTORY "/var/run/secrets/keys/truststore" ---> Running in bdb5a64872db Removing intermediate container bdb5a64872db ---> 94a62f25db63 Step 8/11 : COPY --chown=forgerock:root default-scripts /opt/opendj/default-scripts ---> d3058378d774 Step 9/11 : COPY --chown=forgerock:root ldif-ext /opt/opendj/ldif-ext ---> 8325e4c48557 Step 10/11 : COPY --chown=forgerock:root *.sh /opt/opendj/ ---> 409b1dea0e01 Step 11/11 : RUN ./ds-setup.sh && rm ./ds-setup.sh && rm -fr ldif-ext ---> Running in 9a475aca5989 + rm -f template/config/tools.properties + rm -rf -- README* bat *.zip *.png *.bat setup.sh + ./bin/dskeymgr create-deployment-key --deploymentKeyPassword password + deploymentKey=AESAFCMBYnN5e1aBNpUaB5ThB4tJJQ5CBVN1bkVDAPBtdkWLDT4Q9j8 + ./setup --instancePath /opt/opendj/data --serverId docker --hostname localhost --deploymentKey AESAFCMBYnN5e1aBNpUaB5ThB4tJJQ5CBVN1bkVDAPBtdkWLDT4Q9j8 --deploymentKeyPassword password --rootUserPassword password --adminConnectorPort 4444 --ldapPort 1389 --enableStartTls --ldapsPort 1636 --httpPort 8080 --httpsPort 8443 --replicationPort 8989 --rootUserDn uid=admin --monitorUserDn uid=monitor --monitorUserPassword password --acceptLicense READ THIS SOFTWARE LICENSE AGREEMENT CAREFULLY. BY DOWNLOADING OR INSTALLING THE FORGEROCK SOFTWARE, YOU, ON BEHALF OF YOURSELF AND YOUR COMPANY, AGREE TO BE BOUND BY THIS SOFTWARE LICENSE AGREEMENT. IF YOU DO NOT AGREE TO THESE TERMS, DO NOT DOWNLOAD OR INSTALL THE FORGEROCK SOFTWARE. 1. Software License. 1.1. Development Right to Use. If Company intends to or does use the ForgeRock Software only for the purpose(s) of developing, testing, prototyping and demonstrating its application software, then ForgeRock hereby grants Company a nonexclusive, nontransferable, limited license to use the ForgeRock Software only for those purposes, solely at Company's facilities and only in a non-production environment. ForgeRock may audit Company's use of the ForgeRock Software to confirm that a production license is not required upon reasonable written notice to Company. If Company intends to use the ForgeRock Software in a live environment, Company must purchase a production license and may only use the ForgeRock Software licensed thereunder in accordance with the terms and conditions of that subscription agreement. 1.2. Restrictions. Except as expressly set forth in this ForgeRock Software License Agreement (the "Agreement"), Company shall not, directly or indirectly: (a) sublicense, resell, rent, lease, distribute or otherwise transfer rights or usage in the ForgeRock Software, including without limitation to Company subsidiaries and affiliates; (b) remove or alter any copyright, trademark or proprietary notices in the ForgeRock Software; or (c) use the ForgeRock Software in any way that would subject the ForgeRock Software, in whole in or in part, to a Copyleft License. As used herein, "Copyleft License" means a software license that requires that information necessary for reproducing and modifying such software must be made available publicly to recipients of executable versions of such software (see, e.g., GNU General Public License and http://www.gnu.org/copyleft/). 2. Proprietary Rights. 2.1. ForgeRock Intellectual Property. Title to and ownership of all copies of the ForgeRock Software whether in machine-readable (source, object code or other format) or printed form, and all related technical know-how and all rights therein (including without limitation all intellectual property rights applicable thereto), belong to ForgeRock and its licensors and shall remain the exclusive property thereof. ForgeRock's name, logo, trade names and trademarks are owned exclusively by ForgeRock and no right is granted to Company to use any of the foregoing except as expressly permitted herein. All rights not expressly granted to Company are reserved by ForgeRock and its licensors. 2.2. Suggestions. Company hereby grants to ForgeRock a royalty-free, worldwide, transferable, sublicensable and irrevocable right and license to use, copy, modify and distribute, including by incorporating into any product or service owned by ForgeRock, any suggestions, enhancements, recommendations or other feedback provided by Company relating to any product or service owned or offered by ForgeRock. 2.3. Source Code. The source code underlying the ForgeRock Software is available at www.forgerock.org. 3. Term and Termination. The terms of this Agreement shall commence on the Effective Date and shall continue in force unless earlier terminated in accordance this Section. This Agreement shall terminate without notice to Company in the event Company is in material breach of any of the terms and conditions of this Agreement. As used herein, "Effective Date" means the date on which Company first accepted this Agreement and downloads the ForgeRock Software. 4. Disclaimer of Warranties. THE FORGEROCK SOFTWARE LICENSED HEREUNDER IS LICENSED "AS IS" AND WITHOUT WARRANTY OF ANY KIND. FORGEROCK AND IT'S LICENSORS EXPRESSLY DISCLAIM ALL WARRANTIES, WHETHER EXPRESS, IMPLIED OR STATUTORY, INCLUDING, WITHOUT LIMITATION, THE IMPLIED WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND ANY WARRANTY OF NON-INFRINGEMENT. 5. General Indemnification. Company shall defend, indemnify and hold ForgeRock harmless from and against any and all liabilities, damages, losses, costs and expenses (including but not limited to reasonable fees of attorneys and other professionals) payable to third parties based upon any claim arising out of or related to the use of Company's products, provided that ForgeRock: (a) promptly notifies Company of the claim; (b) provides Company with all reasonable information and assistance, at Company's expense, to defend or settle such a claim; and (c) grants Company authority and control of the defense or settlement of such claim. Company shall not settle any such claim, without ForgeRock's prior written consent, if such settlement would in any manner effect ForgeRock's rights in the ForgeRock Software or otherwise. ForgeRock reserves the right to retain counsel, at ForgeRock's expense, to participate in the defense and settlement of any such claim. 6. Limitation of Liability. IN NO EVENT SHALL FORGEROCK BE LIABLE FOR THE COST OF PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES, ANY LOST PROFITS, REVENUE, OR DATA, INTERRUPTION OF BUSINESS OR FOR ANY INCIDENTAL, SPECIAL, CONSEQUENTIAL OR INDIRECT DAMAGES OF ANY KIND, AND WHETHER ARISING OUT OF BREACH OF WARRANTY, BREACH OF CONTRACT, TORT (INCLUDING NEGLIGENCE), STRICT LIABILITY OR OTHERWISE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE OR IF SUCH DAMAGE COULD HAVE BEEN REASONABLY FORESEEN. IN NO EVENT SHALL FORGEROCK'S LIABILITY ARISING OUT OF OR RELATED TO THIS AGREEMENT WHETHER IN CONTRACT, TORT OR UNDER ANY OTHER THEORY OF LIABILITY, EXCEED IN THE AGGREGATE $1,000 USD. 7. General. 7.1. Governing Law. This Agreement shall be governed by and interpreted in accordance with the laws of the State of California without reference to its conflicts of law provisions. 7.2. Assignment. Company may not assign any of its rights or obligations under this Agreement without the prior written consent of ForgeRock, which consent shall not be unreasonably withheld. Any assignment not in conformity with this Section shall be null and void. 7.3. Waiver. A waiver on one occasion shall not be construed as a waiver of any right on any future occasion. No delay or omission by a party in exercising any of its rights hereunder shall operate as a waiver of such rights. 7.4. Compliance with Law. The ForgeRock Software is subject to U.S. export control laws, including the U.S. Export Administration Act and its associated regulations, and may be subject to export or import regulations in other countries. Company agrees to comply with all laws and regulations of the United States and other countries ("Export Laws") to assure that neither the ForgeRock Software, nor any direct products thereof are; (a) exported, directly or indirectly, in violation of Export Laws, either to any countries that are subject to U.S. export restrictions or to any end user who has been prohibited from participating in the U.S. export transactions by any federal agency of the U.S. government or (b) intended to be used for any purpose prohibited by Export Laws, including, without limitation, nuclear, chemical, or biological weapons proliferation. 7.5. US Government Restrictions. Company acknowledges that the ForgeRock Software consists of "commercial computer software" and "commercial computer software documentation" as such terms are defined in the Code of Federal Regulations. No Government procurement regulations or contract clauses or provisions shall be deemed a part of any transaction between the parties unless its inclusion is required by law, or mutually agreed in writing by the parties in connection with a specific transaction. Use, duplication, reproduction, release, modification, disclosure or transfer of the ForgeRock Software is restricted in accordance with the terms of this Agreement. 7.6. Provision Severability. In the event that it is determined by a court of competent jurisdiction that any provision of this Agreement is invalid, illegal, or otherwise unenforceable, such provision shall be enforced as nearly as possible in accordance with the stated intention of the parties, while the remainder of this Agreement shall remain in full force and effect and bind the parties according to its terms. To the extent any provision cannot be enforced in accordance with the stated intentions of the parties, such terms and conditions shall be deemed not to be a part of this Agreement. 7.7. Entire Agreement. This Agreement constitutes the entire and exclusive agreement between the parties with respect to the subject matter hereof and supersede any prior agreements between the parties with respect to such subject matter Validating parameters..... Done Configuring certificates..... Done Configuring server..... Done To see basic server status and configuration, you can launch /opt/opendj/bin/status + ./bin/dsconfig --offline --no-prompt --batch set-global-configuration-prop --set "server-id:&{ds.server.id|docker}" set-global-configuration-prop --set "group-id:&{ds.group.id|default}" set-global-configuration-prop --set "advertised-listen-address:&{ds.advertised.listen.address|localhost}" set-global-configuration-prop --advanced --set "trust-transaction-ids:&{platform.trust.transaction.header|false}" delete-log-publisher --publisher-name "File-Based Error Logger" The Log Publisher was deleted successfully delete-log-publisher --publisher-name "File-Based Access Logger" The Log Publisher was deleted successfully delete-log-publisher --publisher-name "File-Based Audit Logger " The Log Publisher was deleted successfully delete-log-publisher --publisher-name "File-Based HTTP Access Logger" The Log Publisher was deleted successfully delete-log-publisher --publisher-name "Json File-Based Access Logger" The Log Publisher was deleted successfully delete-log-publisher --publisher-name "Json File-Based HTTP Access Logger" The Log Publisher was deleted successfully create-log-publisher --type console-error --publisher-name "Console Error Logger" --set enabled:true --set default-severity:error --set default-severity:warning --set default-severity:notice --set override-severity:SYNC=INFO,ERROR,WARNING,NOTICE The Console Error Log Publisher was created successfully create-log-publisher --type external-access --publisher-name "Console LDAP Access Logger" --set enabled:true --set config-file:config/audit-handlers/ldap-access-stdout.json --set "filtering-policy:&{ds.log.filtering.policy|inclusive}" The External Access Log Publisher was created successfully create-log-publisher --type external-http-access --publisher-name "Console HTTP Access Logger" --set enabled:true --set config-file:config/audit-handlers/http-access-stdout.json The External HTTP Access Log Publisher was created successfully delete-sasl-mechanism-handler --handler-name "GSSAPI" The SASL Mechanism Handler was deleted successfully set-synchronization-provider-prop --provider-name "Multimaster synchronization" --set "bootstrap-replication-server:&{ds.bootstrap.replication.servers|localhost:8989}" set-synchronization-provider-prop --provider-name "Multimaster synchronization" --set "replication-purge-delay:86400 s" + dsconfig --offline --no-prompt --batch set-global-configuration-prop --set "unauthenticated-requests-policy:allow" set-password-policy-prop --policy-name "Default Password Policy" --set "require-secure-authentication:false" --set "require-secure-password-changes:false" --reset "password-validator" set-password-policy-prop --policy-name "Root Password Policy" --set "require-secure-authentication:false" --set "require-secure-password-changes:false" --reset "password-validator" + dsconfig --offline --no-prompt --batch create-trust-manager-provider --provider-name "PEM Trust Manager" --type pem --set enabled:true --set pem-directory:/var/run/secrets/keys/truststore The Pem Trust Manager Provider was created successfully set-connection-handler-prop --handler-name https --set trust-manager-provider:"PEM Trust Manager" set-connection-handler-prop --handler-name ldap --set trust-manager-provider:"PEM Trust Manager" set-connection-handler-prop --handler-name ldaps --set trust-manager-provider:"PEM Trust Manager" set-synchronization-provider-prop --provider-name "Multimaster Synchronization" --set trust-manager-provider:"PEM Trust Manager" set-administration-connector-prop --set trust-manager-provider:"PEM Trust Manager" delete-trust-manager-provider --provider-name "PKCS12" The Trust Manager Provider was deleted successfully create-key-manager-provider --provider-name "PEM Key Manager" --type pem --set enabled:true --set pem-directory:/var/run/secrets/keys/ds The Pem Key Manager Provider was created successfully set-connection-handler-prop --handler-name https --set key-manager-provider:"PEM Key Manager" set-connection-handler-prop --handler-name ldap --set key-manager-provider:"PEM Key Manager" set-connection-handler-prop --handler-name ldaps --set key-manager-provider:"PEM Key Manager" set-synchronization-provider-prop --provider-name "Multimaster Synchronization" --set key-manager-provider:"PEM Key Manager" set-crypto-manager-prop --set key-manager-provider:"PEM Key Manager" set-administration-connector-prop --set key-manager-provider:"PEM Key Manager" delete-key-manager-provider --provider-name "PKCS12" The Key Manager Provider was deleted successfully + cd /opt/opendj/data + rm -fr legal-notices + rm -fr lib/extensions + ln -s /opt/opendj/lib/extensions lib/extensions + ldifmodify config/config.ldif + mv config/config.ldif.tmp config/config.ldif + removeUserPassword db/rootUser/rootUser.ldif uid=admin + file=db/rootUser/rootUser.ldif + dn=uid=admin + ../bin/ldifmodify db/rootUser/rootUser.ldif + rm db/rootUser/rootUser.ldif + mv db/rootUser/rootUser.ldif.tmp db/rootUser/rootUser.ldif + removeUserPassword db/monitorUser/monitorUser.ldif uid=monitor + file=db/monitorUser/monitorUser.ldif + dn=uid=monitor + ../bin/ldifmodify db/monitorUser/monitorUser.ldif + rm db/monitorUser/monitorUser.ldif + mv db/monitorUser/monitorUser.ldif.tmp db/monitorUser/monitorUser.ldif + echo source <(/opt/opendj/bin/bash-completion) + tar cvfz /opt/opendj/data.tar.gz bak changelogDb classes config db extlib import-tmp ldif lib locks logs var bak/ changelogDb/ classes/ config/ config/config.ldif config/snmp/ config/snmp/security/ config/snmp/security/opendj-snmp.security config/messages/ config/messages/account-enabled.template config/messages/password-changed.template config/messages/account-unlocked.template config/messages/account-idle-locked.template config/messages/password-expiring.template config/messages/account-temporarily-locked.template config/messages/account-expired.template config/messages/account-reset-locked.template config/messages/account-permanently-locked.template config/messages/password-expired.template config/messages/password-reset.template config/messages/account-disabled.template config/wordlist.txt config/audit-handlers/ config/audit-handlers/ldap-access-stdout.json config/audit-handlers/oracle_tables-example.sql config/audit-handlers/elasticsearch-config.json-example config/audit-handlers/jms-config.json-example config/audit-handlers/postgres_tables-example.sql config/audit-handlers/elasticsearch-index-setup-example.json config/audit-handlers/http-access-stdout.json config/audit-handlers/splunk-config.json-example config/audit-handlers/syslog-config.json-example config/audit-handlers/mysql_tables-example.sql config/audit-handlers/json-stdout-config.json-example config/audit-handlers/jdbc-config.json-example config/keystore.pin config/rest2ldap/ config/rest2ldap/endpoints/ config/rest2ldap/endpoints/api/ config/rest2ldap/endpoints/api/example-v1.json config/MakeLDIF/ config/MakeLDIF/example.template config/MakeLDIF/first.names config/MakeLDIF/cities config/MakeLDIF/last.names config/MakeLDIF/addrate.template config/MakeLDIF/people_and_groups.template config/MakeLDIF/states config/MakeLDIF/streets config/keystore config/java.properties config/common-passwords.txt db/ db/schema/ db/schema/03-keystore.ldif db/schema/00-core.ldif db/schema/03-rfc2926.ldif db/schema/03-changelog.ldif db/schema/03-rfc2714.ldif db/schema/02-config.ldif db/schema/03-uddiv3.ldif db/schema/01-pwpolicy.ldif db/schema/03-rfc3112.ldif db/schema/05-samba.ldif db/schema/05-rfc4876.ldif db/schema/03-rfc3712.ldif db/schema/05-solaris.ldif db/schema/03-pwpolicyextension.ldif db/schema/04-rfc2307bis.ldif db/schema/03-rfc2713.ldif db/schema/06-compat.ldif db/schema/03-rfc2739.ldif db/rootUser/ db/rootUser/rootUser.ldif db/tasks/ db/monitorUser/ db/monitorUser/monitorUser.ldif db/adminRoot/ db/adminRoot/admin-backend.ldif extlib/ import-tmp/ ldif/ lib/ lib/extensions locks/ locks/server.lock logs/ var/ var/upgrade/ var/upgrade/schema.ldif.current var/data.version + cd /opt/opendj + chmod -R a+rw template/setup-profiles/AM + cat ldif-ext/external-am-datastore.ldif ldif-ext/uma/opendj_uma_audit.ldif ldif-ext/uma/opendj_uma_pending_requests.ldif ldif-ext/uma/opendj_uma_resource_set_labels.ldif ldif-ext/uma/opendj_uma_resource_sets.ldif ldif-ext/alpha_bravo.ldif + cat ldif-ext/orgs.ldif Removing intermediate container 9a475aca5989 ---> a0fc9ea9dde5 [Warning] One or more build-args [CONFIG_PROFILE] were not consumed Successfully built a0fc9ea9dde5 Successfully tagged gcr.io/engineeringpit/lodestar-images/ds:xlou The push refers to repository [gcr.io/engineeringpit/lodestar-images/ds] 18e98dbd20ef: Preparing a336b11b2782: Preparing 4a0d86b33513: Preparing d67ba3f62f1a: Preparing f91128dd9303: Preparing 7e178ceb030b: Preparing 5f70bf18a086: Preparing ebc5f6f12c2b: Preparing 7e178ceb030b: Waiting 5f70bf18a086: Waiting 6a900ee656e0: Preparing 2a85b1b0509d: Preparing 5df41b110532: Preparing acc7038f329f: Preparing 3af14c9a24c9: Preparing ebc5f6f12c2b: Waiting 6a900ee656e0: Waiting 2a85b1b0509d: Waiting 5df41b110532: Waiting acc7038f329f: Waiting 3af14c9a24c9: Waiting a336b11b2782: Pushed 4a0d86b33513: Pushed d67ba3f62f1a: Pushed 7e178ceb030b: Layer already exists 5f70bf18a086: Layer already exists 6a900ee656e0: Layer already exists 2a85b1b0509d: Layer already exists ebc5f6f12c2b: Layer already exists 5df41b110532: Layer already exists 3af14c9a24c9: Layer already exists acc7038f329f: Layer already exists 18e98dbd20ef: Pushed f91128dd9303: Pushed xlou: digest: sha256:c681bf08f147ff85097975e25bc4573d10ce1252d197bed26796dcda7f95db6e size: 3046 Sending build context to Docker daemon 292.9kB Step 1/10 : FROM gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 ---> 9deba3244b26 Step 2/10 : COPY debian-buster-sources.list /etc/apt/sources.list ---> 242b4b870710 Step 3/10 : WORKDIR /opt/opendj ---> Running in 576abe038337 Removing intermediate container 576abe038337 ---> 59c382afd681 Step 4/10 : COPY --chown=forgerock:root common /opt/opendj/ ---> 66596d7c68fa Step 5/10 : COPY --chown=forgerock:root idrepo /opt/opendj/ ---> 0b2d90998258 Step 6/10 : COPY --chown=forgerock:root scripts /opt/opendj/scripts ---> 0f23026554a0 Step 7/10 : COPY --chown=forgerock:root uma /opt/opendj/uma ---> 474d385d011b Step 8/10 : COPY --chown=forgerock:root idrepo/*.ldif /var/tmp/ ---> 9796698ed556 Step 9/10 : RUN chmod +w template/setup-profiles/AM/config/6.5/base-entries.ldif && cat scripts/external-am-datastore.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat uma/opendj_uma_audit.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat uma/opendj_uma_pending_requests.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat uma/opendj_uma_resource_set_labels.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat uma/opendj_uma_resource_sets.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && cat /var/tmp/alpha_bravo.ldif >> template/setup-profiles/AM/config/6.5/base-entries.ldif && chmod +w template/setup-profiles/AM/identity-store/7.0/base-entries.ldif && cat /var/tmp/orgs.ldif >> template/setup-profiles/AM/identity-store/7.0/base-entries.ldif && rm /var/tmp/*ldif ---> Running in d9001be28105 Removing intermediate container d9001be28105 ---> dedb019c50f8 Step 10/10 : RUN bin/setup.sh && bin/relax-security-settings.sh && rm bin/setup.sh bin/relax-security-settings.sh ---> Running in 519d11db15f0 Configuring profile AM configuration data store....... Done Configuring profile AM identity data store...... Done Configuring profile IDM external repository........ Done Configuring profile AM CTS data store....... Done Configuring profile DS proxied server...... Done create-trust-manager-provider --provider-name "PEM Trust Manager" --type pem --set enabled:true --set pem-directory:pem-trust-directory The Pem Trust Manager Provider was created successfully set-connection-handler-prop --handler-name https --set trust-manager-provider:"PEM Trust Manager" set-connection-handler-prop --handler-name ldap --set trust-manager-provider:"PEM Trust Manager" set-connection-handler-prop --handler-name ldaps --set trust-manager-provider:"PEM Trust Manager" set-synchronization-provider-prop --provider-name "Multimaster Synchronization" --set trust-manager-provider:"PEM Trust Manager" set-administration-connector-prop --set trust-manager-provider:"PEM Trust Manager" delete-trust-manager-provider --provider-name "PKCS12" The Trust Manager Provider was deleted successfully create-key-manager-provider --provider-name "PEM Key Manager" --type pem --set enabled:true --set pem-directory:pem-keys-directory The Pem Key Manager Provider was created successfully set-connection-handler-prop --handler-name https --set key-manager-provider:"PEM Key Manager" set-connection-handler-prop --handler-name ldap --set key-manager-provider:"PEM Key Manager" set-connection-handler-prop --handler-name ldaps --set key-manager-provider:"PEM Key Manager" set-synchronization-provider-prop --provider-name "Multimaster Synchronization" --set key-manager-provider:"PEM Key Manager" set-crypto-manager-prop --set key-manager-provider:"PEM Key Manager" set-administration-connector-prop --set key-manager-provider:"PEM Key Manager" delete-key-manager-provider --provider-name "PKCS12" The Key Manager Provider was deleted successfully create-backend-index --backend-name amIdentityStore --set index-type:equality --type generic --index-name fr-idm-uuid The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:equality --index-name fr-idm-effectiveApplications The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:equality --index-name fr-idm-effectiveGroup The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:presence --index-name fr-idm-lastSync The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-user-manager --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-user-meta --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-user-notifications --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-user-roles --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-user-authzroles-internal-role --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-user-authzroles-managed-role --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-organization-owner --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-organization-admin --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-organization-member --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:ordering --type generic --index-name fr-idm-managed-user-active-date The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:ordering --type generic --index-name fr-idm-managed-user-inactive-date The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-user-groups --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-assignment-member --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-application-member --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully create-backend-index --backend-name amIdentityStore --set index-type:extensible --index-name fr-idm-managed-application-owner --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.7 --set index-extensible-matching-rule:1.3.6.1.4.1.36733.2.1.4.9 The Backend Index was created successfully + dsconfig --offline --no-prompt --batch set-global-configuration-prop --set "unauthenticated-requests-policy:allow" set-password-policy-prop --policy-name "Default Password Policy" --set "require-secure-authentication:false" --set "require-secure-password-changes:false" --reset "password-validator" set-password-policy-prop --policy-name "Root Password Policy" --set "require-secure-authentication:false" --set "require-secure-password-changes:false" --reset "password-validator" Removing intermediate container 519d11db15f0 ---> d38a1d7517ab [Warning] One or more build-args [CONFIG_PROFILE] were not consumed Successfully built d38a1d7517ab Successfully tagged gcr.io/engineeringpit/lodestar-images/ds-idrepo:xlou The push refers to repository [gcr.io/engineeringpit/lodestar-images/ds-idrepo] dc9f0ee16c52: Preparing 914325fc3061: Preparing ccdd15aa0213: Preparing fbc23e69881e: Preparing 5b0c208c31b9: Preparing 11178a752757: Preparing 1d7513b2afe6: Preparing d30a64a3168b: Preparing 7e178ceb030b: Preparing 5f70bf18a086: Preparing ebc5f6f12c2b: Preparing 6a900ee656e0: Preparing 2a85b1b0509d: Preparing 5df41b110532: Preparing acc7038f329f: Preparing 3af14c9a24c9: Preparing 11178a752757: Waiting 1d7513b2afe6: Waiting d30a64a3168b: Waiting 7e178ceb030b: Waiting 5f70bf18a086: Waiting ebc5f6f12c2b: Waiting 6a900ee656e0: Waiting 2a85b1b0509d: Waiting 5df41b110532: Waiting acc7038f329f: Waiting 3af14c9a24c9: Waiting fbc23e69881e: Pushed 914325fc3061: Pushed ccdd15aa0213: Pushed 5b0c208c31b9: Pushed dc9f0ee16c52: Pushed 5f70bf18a086: Layer already exists 7e178ceb030b: Layer already exists ebc5f6f12c2b: Layer already exists 6a900ee656e0: Layer already exists 2a85b1b0509d: Layer already exists 5df41b110532: Layer already exists acc7038f329f: Layer already exists 3af14c9a24c9: Layer already exists 1d7513b2afe6: Pushed 11178a752757: Pushed d30a64a3168b: Pushed xlou: digest: sha256:77a575f2f59e3ff456b297a22ae08bd765a0146e0929b0dfb78d3b46ed07d3e1 size: 3662 Sending build context to Docker daemon 292.9kB Step 1/10 : FROM gcr.io/forgerock-io/ds/pit1:7.3.0-47dd3dc1b26e0d8a982cad26d51b3a91ed1e9309 ---> 9deba3244b26 Step 2/10 : USER root ---> Using cache ---> 940f8554f1f4 Step 3/10 : COPY debian-buster-sources.list /etc/apt/sources.list ---> 2902861e037e Step 4/10 : RUN chown -R forgerock:root /opt/opendj ---> Running in b30757a3a9ae Removing intermediate container b30757a3a9ae ---> 6a476b8f89ff Step 5/10 : USER forgerock ---> Running in c5cd50455fd2 Removing intermediate container c5cd50455fd2 ---> b40a6645c251 Step 6/10 : COPY --chown=forgerock:root common /opt/opendj/ ---> 0f56c74b8ed2 Step 7/10 : COPY --chown=forgerock:root cts /opt/opendj/ ---> 6dbb457fc345 Step 8/10 : COPY --chown=forgerock:root scripts /opt/opendj/scripts ---> c910481b416c Step 9/10 : ARG profile_version ---> Running in c06bd4fd4f26 Removing intermediate container c06bd4fd4f26 ---> a92240d3dbaa Step 10/10 : RUN bin/setup.sh && bin/relax-security-settings.sh && rm bin/setup.sh bin/relax-security-settings.sh ---> Running in 1cc69b4594fe Configuring profile AM CTS data store....... Done Configuring profile DS proxied server...... Done create-trust-manager-provider --provider-name "PEM Trust Manager" --type pem --set enabled:true --set pem-directory:pem-trust-directory The Pem Trust Manager Provider was created successfully set-connection-handler-prop --handler-name https --set trust-manager-provider:"PEM Trust Manager" set-connection-handler-prop --handler-name ldap --set trust-manager-provider:"PEM Trust Manager" set-connection-handler-prop --handler-name ldaps --set trust-manager-provider:"PEM Trust Manager" set-synchronization-provider-prop --provider-name "Multimaster Synchronization" --set trust-manager-provider:"PEM Trust Manager" set-administration-connector-prop --set trust-manager-provider:"PEM Trust Manager" delete-trust-manager-provider --provider-name "PKCS12" The Trust Manager Provider was deleted successfully create-key-manager-provider --provider-name "PEM Key Manager" --type pem --set enabled:true --set pem-directory:pem-keys-directory The Pem Key Manager Provider was created successfully set-connection-handler-prop --handler-name https --set key-manager-provider:"PEM Key Manager" set-connection-handler-prop --handler-name ldap --set key-manager-provider:"PEM Key Manager" set-connection-handler-prop --handler-name ldaps --set key-manager-provider:"PEM Key Manager" set-synchronization-provider-prop --provider-name "Multimaster Synchronization" --set key-manager-provider:"PEM Key Manager" set-crypto-manager-prop --set key-manager-provider:"PEM Key Manager" set-administration-connector-prop --set key-manager-provider:"PEM Key Manager" delete-key-manager-provider --provider-name "PKCS12" The Key Manager Provider was deleted successfully + dsconfig --offline --no-prompt --batch set-global-configuration-prop --set "unauthenticated-requests-policy:allow" set-password-policy-prop --policy-name "Default Password Policy" --set "require-secure-authentication:false" --set "require-secure-password-changes:false" --reset "password-validator" set-password-policy-prop --policy-name "Root Password Policy" --set "require-secure-authentication:false" --set "require-secure-password-changes:false" --reset "password-validator" Removing intermediate container 1cc69b4594fe ---> 17e53e0aefca [Warning] One or more build-args [CONFIG_PROFILE] were not consumed Successfully built 17e53e0aefca Successfully tagged gcr.io/engineeringpit/lodestar-images/ds-cts:xlou The push refers to repository [gcr.io/engineeringpit/lodestar-images/ds-cts] 990692bdb0ac: Preparing 03e5f1c13c2d: Preparing 97cc5908563c: Preparing 1b6ebd141400: Preparing e2da494ad62f: Preparing 0aa196dcfeeb: Preparing 7e178ceb030b: Preparing 5f70bf18a086: Preparing ebc5f6f12c2b: Preparing 6a900ee656e0: Preparing 2a85b1b0509d: Preparing 5df41b110532: Preparing acc7038f329f: Preparing 3af14c9a24c9: Preparing 0aa196dcfeeb: Waiting 7e178ceb030b: Waiting 5f70bf18a086: Waiting ebc5f6f12c2b: Waiting 6a900ee656e0: Waiting 2a85b1b0509d: Waiting 5df41b110532: Waiting acc7038f329f: Waiting 3af14c9a24c9: Waiting 03e5f1c13c2d: Pushed 1b6ebd141400: Pushed 97cc5908563c: Pushed 990692bdb0ac: Pushed 7e178ceb030b: Layer already exists 5f70bf18a086: Layer already exists ebc5f6f12c2b: Layer already exists 2a85b1b0509d: Layer already exists 6a900ee656e0: Layer already exists 5df41b110532: Layer already exists 3af14c9a24c9: Layer already exists acc7038f329f: Layer already exists 0aa196dcfeeb: Pushed e2da494ad62f: Pushed xlou: digest: sha256:0885ccb41d442b815fe0ec1443124023961234db5b285a663bed37dff7207864 size: 3251 Sending build context to Docker daemon 34.3kB Step 1/6 : FROM gcr.io/forgerock-io/ig/pit1:7.3.0-latest-postcommit ---> b97c2a3010cf Step 2/6 : COPY debian-buster-sources.list /etc/apt/sources.list ---> Using cache ---> 6f1f8dfb827f Step 3/6 : ARG CONFIG_PROFILE=cdk ---> Using cache ---> 1ea3d9405c1c Step 4/6 : RUN echo "\033[0;36m*** Building '${CONFIG_PROFILE}' profile ***\033[0m" ---> Using cache ---> 2c46bb490dfd Step 5/6 : COPY --chown=forgerock:root config-profiles/${CONFIG_PROFILE}/ /var/ig ---> Using cache ---> 1fa17c930470 Step 6/6 : COPY --chown=forgerock:root . /var/ig ---> Using cache ---> f91a15d0eb36 Successfully built f91a15d0eb36 Successfully tagged gcr.io/engineeringpit/lodestar-images/ig:xlou The push refers to repository [gcr.io/engineeringpit/lodestar-images/ig] 902bdba5abb8: Preparing 9de06495f434: Preparing bf96000f7582: Preparing 4fb17506c7d6: Preparing 5696f243e6cc: Preparing 964c1eecc7f5: Preparing ab8038891451: Preparing c6f8bfcecf05: Preparing 315cd8c5da97: Preparing d456513ae67c: Preparing 67a4178b7d47: Preparing ab8038891451: Waiting c6f8bfcecf05: Waiting 315cd8c5da97: Waiting d456513ae67c: Waiting 67a4178b7d47: Waiting 964c1eecc7f5: Waiting 902bdba5abb8: Layer already exists bf96000f7582: Layer already exists 4fb17506c7d6: Layer already exists 5696f243e6cc: Layer already exists 9de06495f434: Layer already exists ab8038891451: Layer already exists 964c1eecc7f5: Layer already exists 315cd8c5da97: Layer already exists d456513ae67c: Layer already exists c6f8bfcecf05: Layer already exists 67a4178b7d47: Layer already exists xlou: digest: sha256:91c570f2b277b1f7f2f6dee6d07e94121a7f9703dc7964abd12ed996abf848b5 size: 2621 Updated the image_defaulter with your new image for am: "gcr.io/engineeringpit/lodestar-images/am:xlou". Updated the image_defaulter with your new image for idm: "gcr.io/engineeringpit/lodestar-images/idm:xlou". Updated the image_defaulter with your new image for ds: "gcr.io/engineeringpit/lodestar-images/ds:xlou". Updated the image_defaulter with your new image for ds-idrepo: "gcr.io/engineeringpit/lodestar-images/ds-idrepo:xlou". Updated the image_defaulter with your new image for ds-cts: "gcr.io/engineeringpit/lodestar-images/ds-cts:xlou". Updated the image_defaulter with your new image for ig: "gcr.io/engineeringpit/lodestar-images/ig:xlou". [run_command]: /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/bin/forgeops install --namespace=xlou --fqdn xlou.iam.xlou-cdm.engineeringpit.com --custom /mnt/disks/data/xslou/lodestar-fork/ext/forgeops_stack/kustomize/overlay/internal-profiles/medium-old --legacy all [run_command]: env={'HOME': '/home/xslou', 'PATH': '/mnt/disks/data/xslou/lodestar-fork/ext/bin:/home/xslou/.local/bin:/home/xslou/bin:~/bin:~/workshop/lodestar-fork/ext/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin', 'DOCKER_SCAN_SUGGEST': 'false'} customresourcedefinition.apiextensions.k8s.io/secretagentconfigurations.secret-agent.secrets.forgerock.io condition met deployment.apps/secret-agent-controller-manager condition met NAME READY STATUS RESTARTS AGE secret-agent-controller-manager-75c755487b-xlkk5 2/2 Running 0 6d15h configmap/dev-utils created configmap/platform-config created ingress.networking.k8s.io/forgerock created ingress.networking.k8s.io/ig created certificate.cert-manager.io/ds-master-cert created certificate.cert-manager.io/ds-ssl-cert created issuer.cert-manager.io/selfsigned-issuer created secretagentconfiguration.secret-agent.secrets.forgerock.io/forgerock-sac created Checking cert-manager and related CRDs: cert-manager CRD found in cluster. Checking secret-agent operator and related CRDs: secret-agent CRD found in cluster.  Checking secret-agent operator is running... secret-agent operator is running Installing component(s): ['all'] platform: "custom-old" in namespace: "xlou".  Deploying base.yaml. This is a one time activity.  Waiting for K8s secrets. Waiting for secret "am-env-secrets" to exist in the cluster: ..done Waiting for secret "idm-env-secrets" to exist in the cluster: ...done Waiting for secret "ds-passwords" to exist in the cluster: done Waiting for secret "ds-env-secrets" to exist in the cluster: secret/cloud-storage-credentials-cts created secret/cloud-storage-credentials-idrepo created service/ds-cts created service/ds-idrepo created statefulset.apps/ds-cts created statefulset.apps/ds-idrepo created job.batch/ldif-importer created done  Deploying ds.yaml. This includes all directory resources.  Waiting for DS deployment. This can take a few minutes. First installation takes longer. Waiting for statefulset "ds-idrepo" to exist in the cluster: Waiting for 3 pods to be ready... Waiting for 2 pods to be ready... Waiting for 1 pods to be ready... statefulset rolling update complete 3 pods at revision ds-idrepo-7b446fff4d... done Waiting for Service Account Password Update: done Waiting for statefulset "ds-cts" to exist in the cluster: statefulset rolling update complete 3 pods at revision ds-cts-87b85b6bd... done Waiting for Service Account Password Update: configmap/amster-files created configmap/idm created configmap/idm-logging-properties created service/am created service/idm created deployment.apps/am created deployment.apps/idm created job.batch/amster created done Cleaning up amster components.  Deploying apps.  Waiting for AM deployment. This can take a few minutes. First installation takes longer. Waiting for deployment "am" to exist in the cluster: deployment.apps/am condition met configmap/amster-retain created done  Waiting for amster job to complete. This can take several minutes. Waiting for job "amster" to exist in the cluster: job.batch/amster condition met done  Waiting for IDM deployment. This can take a few minutes. First installation takes longer. Waiting for deployment "idm" to exist in the cluster: pod/idm-6b4845dbc5-27kmw condition met pod/idm-6b4845dbc5-scjz5 condition met service/admin-ui created service/end-user-ui created service/login-ui created deployment.apps/admin-ui created deployment.apps/end-user-ui created deployment.apps/login-ui created done  Deploying UI.  Waiting for K8s secrets. Waiting for secret "am-env-secrets" to exist in the cluster: done Waiting for secret "idm-env-secrets" to exist in the cluster: done Waiting for secret "ds-passwords" to exist in the cluster: done Waiting for secret "ds-env-secrets" to exist in the cluster: done  Relevant passwords: mgT0vEdQh58FTmJ2xwE4DVIb (amadmin user) xglItyOwvGHTJ8bEH7MuIXfBKurdnyLT (uid=admin user) YplzJklbmbt3LOKSxP5JcJ9ZxH8FUVZV (App str svc acct (uid=am-config,ou=admins,ou=am-config)) uZu8ApajKD1lQ5f2j8RIghBR18Xz4Dr7 (CTS svc acct (uid=openam_cts,ou=admins,ou=famrecords,ou=openam-session,ou=tokens)) zy3UgWUEsmI37nvlYfTLdV3xhGd4boTS (ID repo svc acct (uid=am-identity-bind-account,ou=admins,ou=identities))  Relevant URLs: https://xlou.iam.xlou-cdm.engineeringpit.com/platform https://xlou.iam.xlou-cdm.engineeringpit.com/admin https://xlou.iam.xlou-cdm.engineeringpit.com/am https://xlou.iam.xlou-cdm.engineeringpit.com/enduser  Enjoy your deployment! **************************** Initializing component pods for DS-CTS **************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets -l app=ds-cts -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods -l app=ds-cts -o jsonpath={.items[*].metadata.name}` | grep 3 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- ds-cts-0 ds-cts-1 ds-cts-2 --- stderr --- -------------------- Check pod ds-cts-0 is running -------------------- [loop_until]: kubectl --namespace=xlou get pods ds-cts-0 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-cts-0 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-cts-0 -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:02:01Z --- stderr --- ------------- Check pod ds-cts-0 filesystem is accessible ------------- [loop_until]: kubectl --namespace=xlou exec ds-cts-0 -c ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------------ Check pod ds-cts-0 restart count ------------------ [loop_until]: kubectl --namespace=xlou get pod ds-cts-0 -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-cts-0 has been restarted 0 times. -------------------- Check pod ds-cts-1 is running -------------------- [loop_until]: kubectl --namespace=xlou get pods ds-cts-1 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-cts-1 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-cts-1 -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:02:26Z --- stderr --- ------------- Check pod ds-cts-1 filesystem is accessible ------------- [loop_until]: kubectl --namespace=xlou exec ds-cts-1 -c ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------------ Check pod ds-cts-1 restart count ------------------ [loop_until]: kubectl --namespace=xlou get pod ds-cts-1 -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-cts-1 has been restarted 0 times. -------------------- Check pod ds-cts-2 is running -------------------- [loop_until]: kubectl --namespace=xlou get pods ds-cts-2 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-cts-2 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-cts-2 -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:02:52Z --- stderr --- ------------- Check pod ds-cts-2 filesystem is accessible ------------- [loop_until]: kubectl --namespace=xlou exec ds-cts-2 -c ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------------ Check pod ds-cts-2 restart count ------------------ [loop_until]: kubectl --namespace=xlou get pod ds-cts-2 -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-cts-2 has been restarted 0 times. ************************** Initializing component pods for DS-IDREPO ************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets -l app=ds-idrepo -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods -l app=ds-idrepo -o jsonpath={.items[*].metadata.name}` | grep 3 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- ds-idrepo-0 ds-idrepo-1 ds-idrepo-2 --- stderr --- ------------------ Check pod ds-idrepo-0 is running ------------------ [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-0 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-0 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-0 -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:02:01Z --- stderr --- ----------- Check pod ds-idrepo-0 filesystem is accessible ----------- [loop_until]: kubectl --namespace=xlou exec ds-idrepo-0 -c ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ----------------- Check pod ds-idrepo-0 restart count ----------------- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-0 -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-idrepo-0 has been restarted 0 times. ------------------ Check pod ds-idrepo-1 is running ------------------ [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-1 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-1 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-1 -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:02:35Z --- stderr --- ----------- Check pod ds-idrepo-1 filesystem is accessible ----------- [loop_until]: kubectl --namespace=xlou exec ds-idrepo-1 -c ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ----------------- Check pod ds-idrepo-1 restart count ----------------- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-1 -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-idrepo-1 has been restarted 0 times. ------------------ Check pod ds-idrepo-2 is running ------------------ [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-2 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods ds-idrepo-2 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-2 -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:03:12Z --- stderr --- ----------- Check pod ds-idrepo-2 filesystem is accessible ----------- [loop_until]: kubectl --namespace=xlou exec ds-idrepo-2 -c ds -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-17 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ----------------- Check pod ds-idrepo-2 restart count ----------------- [loop_until]: kubectl --namespace=xlou get pod ds-idrepo-2 -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod ds-idrepo-2 has been restarted 0 times. ****************************** Initializing component pods for AM ****************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app=am -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods -l app=am -o jsonpath={.items[*].metadata.name}` | grep 3 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- am-685d4f4864-28lqb am-685d4f4864-72vv6 am-685d4f4864-qk4tj --- stderr --- -------------- Check pod am-685d4f4864-28lqb is running -------------- [loop_until]: kubectl --namespace=xlou get pods am-685d4f4864-28lqb -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods am-685d4f4864-28lqb -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod am-685d4f4864-28lqb -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:03:49Z --- stderr --- ------- Check pod am-685d4f4864-28lqb filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec am-685d4f4864-28lqb -c openam -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev etc home lib lib32 lib64 libx32 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------- Check pod am-685d4f4864-28lqb restart count ------------- [loop_until]: kubectl --namespace=xlou get pod am-685d4f4864-28lqb -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod am-685d4f4864-28lqb has been restarted 0 times. -------------- Check pod am-685d4f4864-72vv6 is running -------------- [loop_until]: kubectl --namespace=xlou get pods am-685d4f4864-72vv6 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods am-685d4f4864-72vv6 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod am-685d4f4864-72vv6 -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:03:49Z --- stderr --- ------- Check pod am-685d4f4864-72vv6 filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec am-685d4f4864-72vv6 -c openam -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev etc home lib lib32 lib64 libx32 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------- Check pod am-685d4f4864-72vv6 restart count ------------- [loop_until]: kubectl --namespace=xlou get pod am-685d4f4864-72vv6 -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod am-685d4f4864-72vv6 has been restarted 0 times. -------------- Check pod am-685d4f4864-qk4tj is running -------------- [loop_until]: kubectl --namespace=xlou get pods am-685d4f4864-qk4tj -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods am-685d4f4864-qk4tj -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod am-685d4f4864-qk4tj -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:03:49Z --- stderr --- ------- Check pod am-685d4f4864-qk4tj filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec am-685d4f4864-qk4tj -c openam -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev etc home lib lib32 lib64 libx32 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------- Check pod am-685d4f4864-qk4tj restart count ------------- [loop_until]: kubectl --namespace=xlou get pod am-685d4f4864-qk4tj -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod am-685d4f4864-qk4tj has been restarted 0 times. **************************** Initializing component pods for AMSTER **************************** ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou --field-selector status.phase!=Failed get pods -l app=amster -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- amster-2qmgz --- stderr --- ***************************** Initializing component pods for IDM ***************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployment -l app=idm -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods -l app=idm -o jsonpath={.items[*].metadata.name}` | grep 2 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- idm-6b4845dbc5-27kmw idm-6b4845dbc5-scjz5 --- stderr --- -------------- Check pod idm-6b4845dbc5-27kmw is running -------------- [loop_until]: kubectl --namespace=xlou get pods idm-6b4845dbc5-27kmw -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods idm-6b4845dbc5-27kmw -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod idm-6b4845dbc5-27kmw -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:03:49Z --- stderr --- ------- Check pod idm-6b4845dbc5-27kmw filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec idm-6b4845dbc5-27kmw -c openidm -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-11 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------ Check pod idm-6b4845dbc5-27kmw restart count ------------ [loop_until]: kubectl --namespace=xlou get pod idm-6b4845dbc5-27kmw -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod idm-6b4845dbc5-27kmw has been restarted 0 times. -------------- Check pod idm-6b4845dbc5-scjz5 is running -------------- [loop_until]: kubectl --namespace=xlou get pods idm-6b4845dbc5-scjz5 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods idm-6b4845dbc5-scjz5 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod idm-6b4845dbc5-scjz5 -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:03:49Z --- stderr --- ------- Check pod idm-6b4845dbc5-scjz5 filesystem is accessible ------- [loop_until]: kubectl --namespace=xlou exec idm-6b4845dbc5-scjz5 -c openidm -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Dockerfile.java-11 bin boot dev etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var --- stderr --- ------------ Check pod idm-6b4845dbc5-scjz5 restart count ------------ [loop_until]: kubectl --namespace=xlou get pod idm-6b4845dbc5-scjz5 -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod idm-6b4845dbc5-scjz5 has been restarted 0 times. ************************* Initializing component pods for END-USER-UI ************************* --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app.kubernetes.io/name=end-user-ui -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods -l app.kubernetes.io/name=end-user-ui -o jsonpath={.items[*].metadata.name}` | grep 1 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- end-user-ui-6f5dbc46f8-fwztj --- stderr --- ---------- Check pod end-user-ui-6f5dbc46f8-fwztj is running ---------- [loop_until]: kubectl --namespace=xlou get pods end-user-ui-6f5dbc46f8-fwztj -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods end-user-ui-6f5dbc46f8-fwztj -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod end-user-ui-6f5dbc46f8-fwztj -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:04:57Z --- stderr --- --- Check pod end-user-ui-6f5dbc46f8-fwztj filesystem is accessible --- [loop_until]: kubectl --namespace=xlou exec end-user-ui-6f5dbc46f8-fwztj -c end-user-ui -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev docker-entrypoint.d docker-entrypoint.sh entrypoint.sh etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var variable_replacement.sh --- stderr --- -------- Check pod end-user-ui-6f5dbc46f8-fwztj restart count -------- [loop_until]: kubectl --namespace=xlou get pod end-user-ui-6f5dbc46f8-fwztj -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod end-user-ui-6f5dbc46f8-fwztj has been restarted 0 times. *************************** Initializing component pods for LOGIN-UI *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app.kubernetes.io/name=login-ui -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods -l app.kubernetes.io/name=login-ui -o jsonpath={.items[*].metadata.name}` | grep 1 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- login-ui-857ffdc996-2jzcc --- stderr --- ----------- Check pod login-ui-857ffdc996-2jzcc is running ----------- [loop_until]: kubectl --namespace=xlou get pods login-ui-857ffdc996-2jzcc -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods login-ui-857ffdc996-2jzcc -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod login-ui-857ffdc996-2jzcc -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:04:57Z --- stderr --- ---- Check pod login-ui-857ffdc996-2jzcc filesystem is accessible ---- [loop_until]: kubectl --namespace=xlou exec login-ui-857ffdc996-2jzcc -c login-ui -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev docker-entrypoint.d docker-entrypoint.sh entrypoint.sh etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var variable_replacement.sh --- stderr --- ---------- Check pod login-ui-857ffdc996-2jzcc restart count ---------- [loop_until]: kubectl --namespace=xlou get pod login-ui-857ffdc996-2jzcc -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod login-ui-857ffdc996-2jzcc has been restarted 0 times. *************************** Initializing component pods for ADMIN-UI *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app.kubernetes.io/name=admin-ui -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: awk -F" " "{print NF}" <<< `kubectl --namespace=xlou get pods -l app.kubernetes.io/name=admin-ui -o jsonpath={.items[*].metadata.name}` | grep 1 [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected number of elements found [loop_until]: OK (rc = 0) --- stdout --- admin-ui-56c658bb6-jrg44 --- stderr --- ------------ Check pod admin-ui-56c658bb6-jrg44 is running ------------ [loop_until]: kubectl --namespace=xlou get pods admin-ui-56c658bb6-jrg44 -o=jsonpath={.status.phase} | grep "Running" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- Running --- stderr --- [loop_until]: kubectl --namespace=xlou get pods admin-ui-56c658bb6-jrg44 -o=jsonpath={.status.containerStatuses[*].ready} | grep "true" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- true --- stderr --- [loop_until]: kubectl --namespace=xlou get pod admin-ui-56c658bb6-jrg44 -o jsonpath={.status.startTime} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2023-04-04T22:04:57Z --- stderr --- ----- Check pod admin-ui-56c658bb6-jrg44 filesystem is accessible ----- [loop_until]: kubectl --namespace=xlou exec admin-ui-56c658bb6-jrg44 -c admin-ui -- ls / | grep "bin" [loop_until]: (max_time=360, interval=5, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- bin boot dev docker-entrypoint.d docker-entrypoint.sh entrypoint.sh etc home lib lib64 media mnt opt proc root run sbin srv sys tmp usr var variable_replacement.sh --- stderr --- ---------- Check pod admin-ui-56c658bb6-jrg44 restart count ---------- [loop_until]: kubectl --namespace=xlou get pod admin-ui-56c658bb6-jrg44 -o jsonpath={.status.containerStatuses[*].restartCount} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 0 --- stderr --- Pod admin-ui-56c658bb6-jrg44 has been restarted 0 times. ***************************** Checking DS-CTS component is running ***************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets -l app=ds-cts -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- -------------- Waiting for 3 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get statefulsets ds-cts -o jsonpath="current:{.status.currentReplicas} ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "current:3 ready:3 replicas:3" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- current:3 ready:3 replicas:3 --- stderr --- *************************** Checking DS-IDREPO component is running *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets -l app=ds-idrepo -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- -------------- Waiting for 3 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get statefulsets ds-idrepo -o jsonpath="current:{.status.currentReplicas} ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "current:3 ready:3 replicas:3" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- current:3 ready:3 replicas:3 --- stderr --- ******************************* Checking AM component is running ******************************* --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app=am -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- -------------- Waiting for 3 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployments am -o jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:3 replicas:3" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:3 replicas:3 --- stderr --- ***************************** Checking AMSTER component is running ***************************** --------------------- Get expected number of pods --------------------- -------------- Waiting for 1 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get jobs amster -o jsonpath="{.status.succeeded}" | grep "1" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ****************************** Checking IDM component is running ****************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployment -l app=idm -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2 --- stderr --- -------------- Waiting for 2 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployment idm -o jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:2 replicas:2" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:2 replicas:2 --- stderr --- ************************** Checking END-USER-UI component is running ************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app.kubernetes.io/name=end-user-ui -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- -------------- Waiting for 1 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployments end-user-ui -o jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:1 replicas:1" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:1 replicas:1 --- stderr --- **************************** Checking LOGIN-UI component is running **************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app.kubernetes.io/name=login-ui -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- -------------- Waiting for 1 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployments login-ui -o jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:1 replicas:1" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:1 replicas:1 --- stderr --- **************************** Checking ADMIN-UI component is running **************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app.kubernetes.io/name=admin-ui -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- -------------- Waiting for 1 expected pod(s) to be ready -------------- [loop_until]: kubectl --namespace=xlou get deployments admin-ui -o jsonpath="ready:{.status.readyReplicas} replicas:{.status.replicas}" | grep "ready:1 replicas:1" [loop_until]: (max_time=900, interval=30, expected_rc=[0] [loop_until]: Function succeeded after 0s (rc=0) - expected pattern found [loop_until]: OK (rc = 0) --- stdout --- ready:1 replicas:1 --- stderr --- **************************** Initializing component pods for DS-CTS **************************** --------------------- Get DS-CTS software version --------------------- [loop_until]: kubectl --namespace=xlou exec ds-cts-0 -c ds -- find /opt/opendj -name opendj-core.jar [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /opt/opendj/lib/opendj-core.jar --- stderr --- [loop_until]: kubectl --namespace=xlou cp ds-cts-0:/opt/opendj/lib/opendj-core.jar /tmp/ds-cts_info/opendj-core.jar -c ds [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- ************************** Initializing component pods for DS-IDREPO ************************** ------------------- Get DS-IDREPO software version ------------------- [loop_until]: kubectl --namespace=xlou exec ds-idrepo-0 -c ds -- find /opt/opendj -name opendj-core.jar [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /opt/opendj/lib/opendj-core.jar --- stderr --- [loop_until]: kubectl --namespace=xlou cp ds-idrepo-0:/opt/opendj/lib/opendj-core.jar /tmp/ds-idrepo_info/opendj-core.jar -c ds [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- ****************************** Initializing component pods for AM ****************************** ----------------------- Get AM software version ----------------------- Getting product version from https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/serverinfo/version - Login amadmin to get token [loop_until]: kubectl --namespace=xlou get secret am-env-secrets -o jsonpath="{.data.AM_PASSWORDS_AMADMIN_CLEAR}" [loop_until]: (max_time=60, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- bWdUMHZFZFFoNThGVG1KMnh3RTREVkli --- stderr --- Authenticate user via REST [http_cmd]: curl -H "X-OpenAM-Username: amadmin" -H "X-OpenAM-Password: mgT0vEdQh58FTmJ2xwE4DVIb" -H "Content-Type: application/json" -H "Accept-API-Version: resource=2.0, protocol=1.0" -L -X POST "https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/authenticate?realm=/" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- { "tokenId": "tWxWBGQySz_Kg_D7xXktkv8MUWs.*AAJTSQACMDIAAlNLABxRMlhrTEtGajNKQVZId1hBSnJFcEVZWnk3WjQ9AAR0eXBlAANDVFMAAlMxAAIwMQ..*", "successUrl": "/am/console", "realm": "/" } [http_cmd]: curl -L -X GET --cookie "amlbcookie=01" --cookie "iPlanetDirectoryPro=tWxWBGQySz_Kg_D7xXktkv8MUWs.*AAJTSQACMDIAAlNLABxRMlhrTEtGajNKQVZId1hBSnJFcEVZWnk3WjQ9AAR0eXBlAANDVFMAAlMxAAIwMQ..*" --cookie "route=1680645944.243.3372.662993|f60edb382037eb2df1e800d563ad78a7" "https://xlou.iam.xlou-cdm.engineeringpit.com/am/json/serverinfo/version" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- { "_id": "version", "_rev": "-560502796", "version": "7.3.0-SNAPSHOT", "fullVersion": "ForgeRock Access Management 7.3.0-SNAPSHOT Build d0e0f0f08d45c97e1331f11b6b2ae9fc31d5e28d (2023-March-31 06:52)", "revision": "d0e0f0f08d45c97e1331f11b6b2ae9fc31d5e28d", "date": "2023-March-31 06:52" } **************************** Initializing component pods for AMSTER **************************** ***************************** Initializing component pods for IDM ***************************** ---------------------- Get IDM software version ---------------------- Getting product version from https://xlou.iam.xlou-cdm.engineeringpit.com/openidm/info/version [http_cmd]: curl -H "X-OpenIDM-Username: anonymous" -H "X-OpenIDM-Password: anonymous" -L -X GET "https://xlou.iam.xlou-cdm.engineeringpit.com/openidm/info/version" [http_cmd]: http status code OK --- status code --- http status code is 200 (expected 200) --- http response --- { "_id": "version", "productVersion": "7.3.0-SNAPSHOT", "productBuildDate": "20230330162641", "productRevision": "ed278902ce" } ************************* Initializing component pods for END-USER-UI ************************* ------------------ Get END-USER-UI software version ------------------ [loop_until]: kubectl --namespace=xlou exec end-user-ui-6f5dbc46f8-fwztj -c end-user-ui -- find /usr/share/nginx/html -name chunk-vendors.*.js [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /usr/share/nginx/html/js/chunk-vendors.013bbd1b.js --- stderr --- [loop_until]: kubectl --namespace=xlou cp end-user-ui-6f5dbc46f8-fwztj:/usr/share/nginx/html/js/chunk-vendors.013bbd1b.js /tmp/end-user-ui_info/chunk-vendors.013bbd1b.js -c end-user-ui [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- *************************** Initializing component pods for LOGIN-UI *************************** -------------------- Get LOGIN-UI software version -------------------- [loop_until]: kubectl --namespace=xlou exec login-ui-857ffdc996-2jzcc -c login-ui -- find /usr/share/nginx/html -name chunk-vendors.*.js [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /usr/share/nginx/html/js/chunk-vendors.7c4c4742.js --- stderr --- [loop_until]: kubectl --namespace=xlou cp login-ui-857ffdc996-2jzcc:/usr/share/nginx/html/js/chunk-vendors.7c4c4742.js /tmp/login-ui_info/chunk-vendors.7c4c4742.js -c login-ui [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- *************************** Initializing component pods for ADMIN-UI *************************** -------------------- Get ADMIN-UI software version -------------------- [loop_until]: kubectl --namespace=xlou exec admin-ui-56c658bb6-jrg44 -c admin-ui -- find /usr/share/nginx/html -name chunk-vendors.*.js [loop_until]: (max_time=30, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- /usr/share/nginx/html/js/chunk-vendors.25462d15.js --- stderr --- [loop_until]: kubectl --namespace=xlou cp admin-ui-56c658bb6-jrg44:/usr/share/nginx/html/js/chunk-vendors.25462d15.js /tmp/admin-ui_info/chunk-vendors.25462d15.js -c admin-ui [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- tar: Removing leading `/' from member names --- stderr --- [loop_until]: kubectl --namespace=xlou get secret ds-passwords -o jsonpath="{.data.dirmanager\.pw}" [loop_until]: (max_time=60, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- eGdsSXR5T3d2R0hUSjhiRUg3TXVJWGZCS3VyZG55TFQ= --- stderr --- ==================================================================================================== ================ Admin password for DS-CTS is: xglItyOwvGHTJ8bEH7MuIXfBKurdnyLT ================ ==================================================================================================== [loop_until]: kubectl --namespace=xlou get secret ds-passwords -o jsonpath="{.data.dirmanager\.pw}" [loop_until]: (max_time=60, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- eGdsSXR5T3d2R0hUSjhiRUg3TXVJWGZCS3VyZG55TFQ= --- stderr --- ==================================================================================================== ============== Admin password for DS-IDREPO is: xglItyOwvGHTJ8bEH7MuIXfBKurdnyLT ============== ==================================================================================================== ==================================================================================================== ====================== Admin password for AM is: mgT0vEdQh58FTmJ2xwE4DVIb ====================== ==================================================================================================== [loop_until]: kubectl --namespace=xlou get secret idm-env-secrets -o jsonpath="{.data.OPENIDM_ADMIN_PASSWORD}" [loop_until]: (max_time=60, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- UEl1bFg2bmFOYlY4ZTJZTkJyY0FldmxI --- stderr --- ==================================================================================================== ===================== Admin password for IDM is: PIulX6naNbV8e2YNBrcAevlH ===================== ==================================================================================================== *************************************** Dumping pod list *************************************** Dumping pod list to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/_pod-list.txt **************************** Initializing component pods for DS-CTS **************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets -l app=ds-cts -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods -l app=ds-cts -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- ds-cts-0 ds-cts-1 ds-cts-2 --- stderr --- ************************** Initializing component pods for DS-IDREPO ************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get statefulsets -l app=ds-idrepo -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods -l app=ds-idrepo -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- ds-idrepo-0 ds-idrepo-1 ds-idrepo-2 --- stderr --- ****************************** Initializing component pods for AM ****************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app=am -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 3 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods -l app=am -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- am-685d4f4864-28lqb am-685d4f4864-72vv6 am-685d4f4864-qk4tj --- stderr --- **************************** Initializing component pods for AMSTER **************************** ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods -l app=amster -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- amster-2qmgz --- stderr --- ***************************** Initializing component pods for IDM ***************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployment -l app=idm -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 2 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods -l app=idm -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- idm-6b4845dbc5-27kmw idm-6b4845dbc5-scjz5 --- stderr --- ************************* Initializing component pods for END-USER-UI ************************* --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app.kubernetes.io/name=end-user-ui -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods -l app.kubernetes.io/name=end-user-ui -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- end-user-ui-6f5dbc46f8-fwztj --- stderr --- *************************** Initializing component pods for LOGIN-UI *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app.kubernetes.io/name=login-ui -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods -l app.kubernetes.io/name=login-ui -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- login-ui-857ffdc996-2jzcc --- stderr --- *************************** Initializing component pods for ADMIN-UI *************************** --------------------- Get expected number of pods --------------------- [loop_until]: kubectl --namespace=xlou get deployments -l app.kubernetes.io/name=admin-ui -o jsonpath={.items[*].spec.replicas} [loop_until]: (max_time=180, interval=5, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- 1 --- stderr --- ---------------------------- Get pod list ---------------------------- [loop_until]: kubectl --namespace=xlou get pods -l app.kubernetes.io/name=admin-ui -o jsonpath={.items[*].metadata.name} [loop_until]: (max_time=180, interval=10, expected_rc=[0] [loop_until]: OK (rc = 0) --- stdout --- admin-ui-56c658bb6-jrg44 --- stderr --- *********************************** Dumping components logs *********************************** ----------------------- Dumping logs for DS-CTS ----------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/ds-cts-0.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/ds-cts-1.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/ds-cts-2.txt Check pod logs for errors --------------------- Dumping logs for DS-IDREPO --------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/ds-idrepo-0.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/ds-idrepo-1.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/ds-idrepo-2.txt Check pod logs for errors ------------------------- Dumping logs for AM ------------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/am-685d4f4864-28lqb.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/am-685d4f4864-72vv6.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/am-685d4f4864-qk4tj.txt Check pod logs for errors ----------------------- Dumping logs for AMSTER ----------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/amster-2qmgz.txt Check pod logs for errors ------------------------ Dumping logs for IDM ------------------------ Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/idm-6b4845dbc5-27kmw.txt Check pod logs for errors Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/idm-6b4845dbc5-scjz5.txt Check pod logs for errors -------------------- Dumping logs for END-USER-UI -------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/end-user-ui-6f5dbc46f8-fwztj.txt Check pod logs for errors ---------------------- Dumping logs for LOGIN-UI ---------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/login-ui-857ffdc996-2jzcc.txt Check pod logs for errors ---------------------- Dumping logs for ADMIN-UI ---------------------- Dumping pod description and logs to /mnt/disks/data/xslou/lodestar-fork/results/pyrock/platform/pod-logs/stack/20230404-220550-after-deployment/admin-ui-56c658bb6-jrg44.txt Check pod logs for errors [04/Apr/2023 22:06:10] - INFO: Deployment successful ________________________________________________________________________________ [04/Apr/2023 22:06:10] Deploy_all_forgerock_components post : Post method ________________________________________________________________________________ Setting result to PASS Task has been successfully stopped