Triggered by Gerrit: https://gerrit.onap.org/r/c/policy/docker/+/141338 Running as SYSTEM [EnvInject] - Loading node environment variables. Building remotely on prd-ubuntu1804-docker-8c-8g-22130 (ubuntu1804-docker-8c-8g) in workspace /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp [ssh-agent] Looking for ssh-agent implementation... [ssh-agent] Exec ssh-agent (binary ssh-agent on a remote machine) $ ssh-agent SSH_AUTH_SOCK=/tmp/ssh-JfLPJHF0m5lS/agent.2125 SSH_AGENT_PID=2127 [ssh-agent] Started. Running ssh-add (command line suppressed) Identity added: /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp@tmp/private_key_4334126685366119654.key (/w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp@tmp/private_key_4334126685366119654.key) [ssh-agent] Using credentials onap-jobbuiler (Gerrit user) The recommended git tool is: NONE using credential onap-jenkins-ssh Wiping out workspace first. Cloning the remote Git repository Cloning repository git://cloud.onap.org/mirror/policy/docker.git > git init /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp # timeout=10 Fetching upstream changes from git://cloud.onap.org/mirror/policy/docker.git > git --version # timeout=10 > git --version # 'git version 2.17.1' using GIT_SSH to set credentials Gerrit user Verifying host key using manually-configured host key entries > git fetch --tags --progress -- git://cloud.onap.org/mirror/policy/docker.git +refs/heads/*:refs/remotes/origin/* # timeout=30 > git config remote.origin.url git://cloud.onap.org/mirror/policy/docker.git # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url git://cloud.onap.org/mirror/policy/docker.git # timeout=10 Fetching upstream changes from git://cloud.onap.org/mirror/policy/docker.git using GIT_SSH to set credentials Gerrit user Verifying host key using manually-configured host key entries > git fetch --tags --progress -- git://cloud.onap.org/mirror/policy/docker.git refs/changes/38/141338/2 # timeout=30 > git rev-parse a4383ddb08daf12bc481139efd90352bfa803726^{commit} # timeout=10 JENKINS-19022: warning: possible memory leak due to Git plugin usage; see: https://plugins.jenkins.io/git/#remove-git-plugin-buildsbybranch-builddata-script Checking out Revision a4383ddb08daf12bc481139efd90352bfa803726 (refs/changes/38/141338/2) > git config core.sparsecheckout # timeout=10 > git checkout -f a4383ddb08daf12bc481139efd90352bfa803726 # timeout=30 Commit message: "Fix CSIT Helm kafka installation" > git rev-parse FETCH_HEAD^{commit} # timeout=10 > git rev-list --no-walk ed38a50541249063daf2cfb00b312fb173adeace # timeout=10 provisioning config files... copy managed file [npmrc] to file:/home/jenkins/.npmrc copy managed file [pipconf] to file:/home/jenkins/.config/pip/pip.conf [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash /tmp/jenkins13819560636866581788.sh ---> python-tools-install.sh Setup pyenv: * system (set by /opt/pyenv/version) * 3.8.13 (set by /opt/pyenv/version) * 3.9.13 (set by /opt/pyenv/version) * 3.10.6 (set by /opt/pyenv/version) lf-activate-venv(): INFO: Creating python3 venv at /tmp/venv-WIvM lf-activate-venv(): INFO: Save venv in file: /tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-WIvM/bin to PATH Generating Requirements File Python 3.10.6 pip 25.1.1 from /tmp/venv-WIvM/lib/python3.10/site-packages/pip (python 3.10) appdirs==1.4.4 argcomplete==3.6.2 aspy.yaml==1.3.0 attrs==25.3.0 autopage==0.5.2 beautifulsoup4==4.13.4 boto3==1.38.38 botocore==1.38.38 bs4==0.0.2 cachetools==5.5.2 certifi==2025.6.15 cffi==1.17.1 cfgv==3.4.0 chardet==5.2.0 charset-normalizer==3.4.2 click==8.2.1 cliff==4.10.0 cmd2==2.6.1 cryptography==3.3.2 debtcollector==3.0.0 decorator==5.2.1 defusedxml==0.7.1 Deprecated==1.2.18 distlib==0.3.9 dnspython==2.7.0 docker==7.1.0 dogpile.cache==1.4.0 durationpy==0.10 email_validator==2.2.0 filelock==3.18.0 future==1.0.0 gitdb==4.0.12 GitPython==3.1.44 google-auth==2.40.3 httplib2==0.22.0 identify==2.6.12 idna==3.10 importlib-resources==1.5.0 iso8601==2.1.0 Jinja2==3.1.6 jmespath==1.0.1 jsonpatch==1.33 jsonpointer==3.0.0 jsonschema==4.24.0 jsonschema-specifications==2025.4.1 keystoneauth1==5.11.1 kubernetes==33.1.0 lftools==0.37.13 lxml==5.4.0 MarkupSafe==3.0.2 msgpack==1.1.1 multi_key_dict==2.0.3 munch==4.0.0 netaddr==1.3.0 niet==1.4.2 nodeenv==1.9.1 oauth2client==4.1.3 oauthlib==3.3.0 openstacksdk==4.6.0 os-client-config==2.1.0 os-service-types==1.7.0 osc-lib==4.0.2 oslo.config==9.8.0 oslo.context==6.0.0 oslo.i18n==6.5.1 oslo.log==7.1.0 oslo.serialization==5.7.0 oslo.utils==9.0.0 packaging==25.0 pbr==6.1.1 platformdirs==4.3.8 prettytable==3.16.0 psutil==7.0.0 pyasn1==0.6.1 pyasn1_modules==0.4.2 pycparser==2.22 pygerrit2==2.0.15 PyGithub==2.6.1 PyJWT==2.10.1 PyNaCl==1.5.0 pyparsing==2.4.7 pyperclip==1.9.0 pyrsistent==0.20.0 python-cinderclient==9.7.0 python-dateutil==2.9.0.post0 python-heatclient==4.2.0 python-jenkins==1.8.2 python-keystoneclient==5.6.0 python-magnumclient==4.8.1 python-openstackclient==8.1.0 python-swiftclient==4.8.0 PyYAML==6.0.2 referencing==0.36.2 requests==2.32.4 requests-oauthlib==2.0.0 requestsexceptions==1.4.0 rfc3986==2.0.0 rpds-py==0.25.1 rsa==4.9.1 ruamel.yaml==0.18.14 ruamel.yaml.clib==0.2.12 s3transfer==0.13.0 simplejson==3.20.1 six==1.17.0 smmap==5.0.2 soupsieve==2.7 stevedore==5.4.1 tabulate==0.9.0 toml==0.10.2 tomlkit==0.13.3 tqdm==4.67.1 typing_extensions==4.14.0 tzdata==2025.2 urllib3==1.26.20 virtualenv==20.31.2 wcwidth==0.2.13 websocket-client==1.8.0 wrapt==1.17.2 xdg==6.0.0 xmltodict==0.14.2 yq==3.4.3 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content SET_JDK_VERSION=openjdk17 GIT_URL="git://cloud.onap.org/mirror" [EnvInject] - Variables injected successfully. [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/sh /tmp/jenkins17006610850049742544.sh ---> update-java-alternatives.sh ---> Updating Java version ---> Ubuntu/Debian system detected update-alternatives: using /usr/lib/jvm/java-17-openjdk-amd64/bin/java to provide /usr/bin/java (java) in manual mode update-alternatives: using /usr/lib/jvm/java-17-openjdk-amd64/bin/javac to provide /usr/bin/javac (javac) in manual mode update-alternatives: using /usr/lib/jvm/java-17-openjdk-amd64 to provide /usr/lib/jvm/java-openjdk (java_sdk_openjdk) in manual mode openjdk version "17.0.4" 2022-07-19 OpenJDK Runtime Environment (build 17.0.4+8-Ubuntu-118.04) OpenJDK 64-Bit Server VM (build 17.0.4+8-Ubuntu-118.04, mixed mode, sharing) JAVA_HOME=/usr/lib/jvm/java-17-openjdk-amd64 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path '/tmp/java.env' [EnvInject] - Variables injected successfully. [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/sh -xe /tmp/jenkins8555766164974324724.sh + /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp/csit/run-project-csit.sh xacml-pdp WARNING! Using --password via the CLI is insecure. Use --password-stdin. WARNING! Your password will be stored unencrypted in /home/jenkins/.docker/config.json. Configure a credential helper to remove this warning. See https://docs.docker.com/engine/reference/commandline/login/#credentials-store Login Succeeded docker: 'compose' is not a docker command. See 'docker --help' Docker Compose Plugin not installed. Installing now... % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 60.2M 100 60.2M 0 0 69.6M 0 --:--:-- --:--:-- --:--:-- 69.6M Setting project configuration for: xacml-pdp Configuring docker compose... Starting xacml-pdp using postgres + Grafana/Prometheus postgres Pulling policy-db-migrator Pulling pap Pulling kafka Pulling xacml-pdp Pulling api Pulling zookeeper Pulling prometheus Pulling grafana Pulling da9db072f522 Pulling fs layer 96e38c8865ba Pulling fs layer 795b910b71c0 Pulling fs layer d1bdb495a7aa Pulling fs layer 0444d3911dbb Pulling fs layer b801adf990e2 Pulling fs layer d1bdb495a7aa Waiting 0444d3911dbb Waiting b801adf990e2 Waiting da9db072f522 Downloading [> ] 48.06kB/3.624MB da9db072f522 Pulling fs layer 110a13bd01fb Pulling fs layer 12cf1ed9c784 Pulling fs layer d4108afce2f7 Pulling fs layer 07255172bfd8 Pulling fs layer 22c948928e79 Pulling fs layer e92d65bf8445 Pulling fs layer 7910fddefabc Pulling fs layer d4108afce2f7 Waiting 07255172bfd8 Waiting 22c948928e79 Waiting e92d65bf8445 Waiting 7910fddefabc Waiting 12cf1ed9c784 Waiting da9db072f522 Downloading [> ] 48.06kB/3.624MB 110a13bd01fb Waiting da9db072f522 Pulling fs layer 96e38c8865ba Pulling fs layer 5e06c6bed798 Pulling fs layer 684be6598fc9 Pulling fs layer 0d92cad902ba Pulling fs layer dcc0c3b2850c Pulling fs layer eb7cda286a15 Pulling fs layer eb7cda286a15 Waiting 0d92cad902ba Waiting dcc0c3b2850c Waiting 795b910b71c0 Downloading [> ] 31.67kB/2.323MB 684be6598fc9 Waiting da9db072f522 Downloading [> ] 48.06kB/3.624MB 5e06c6bed798 Waiting da9db072f522 Pulling fs layer 96e38c8865ba Pulling fs layer e5d7009d9e55 Pulling fs layer 1ec5fb03eaee Pulling fs layer d3165a332ae3 Pulling fs layer c124ba1a8b26 Pulling fs layer 6394804c2196 Pulling fs layer d3165a332ae3 Waiting 1ec5fb03eaee Waiting c124ba1a8b26 Waiting 6394804c2196 Waiting da9db072f522 Downloading [> ] 48.06kB/3.624MB e5d7009d9e55 Waiting 96e38c8865ba Downloading [> ] 539.6kB/71.91MB 96e38c8865ba Downloading [> ] 539.6kB/71.91MB 96e38c8865ba Downloading [> ] 539.6kB/71.91MB f18232174bc9 Pulling fs layer 9183b65e90ee Pulling fs layer 3f8d5c908dcc Pulling fs layer 30bb92ff0608 Pulling fs layer 807a2e881ecd Pulling fs layer 4a4d0948b0bf Pulling fs layer 04f6155c873d Pulling fs layer 85dde7dceb0a Pulling fs layer 7009d5001b77 Pulling fs layer f18232174bc9 Waiting 538deb30e80c Pulling fs layer 9183b65e90ee Waiting 85dde7dceb0a Waiting 30bb92ff0608 Waiting 807a2e881ecd Waiting 7009d5001b77 Waiting 4a4d0948b0bf Waiting 04f6155c873d Waiting 538deb30e80c Waiting 3f8d5c908dcc Waiting eca0188f477e Pulling fs layer e444bcd4d577 Pulling fs layer eabd8714fec9 Pulling fs layer 45fd2fec8a19 Pulling fs layer 8f10199ed94b Pulling fs layer f963a77d2726 Pulling fs layer f3a82e9f1761 Pulling fs layer 79161a3f5362 Pulling fs layer e444bcd4d577 Waiting 9c266ba63f51 Pulling fs layer 2e8a7df9c2ee Pulling fs layer eabd8714fec9 Waiting 10f05dd8b1db Pulling fs layer 41dac8b43ba6 Pulling fs layer f963a77d2726 Waiting 71a9f6a9ab4d Pulling fs layer f3a82e9f1761 Waiting 79161a3f5362 Waiting da3ed5db7103 Pulling fs layer c955f6e31a04 Pulling fs layer 9c266ba63f51 Waiting 2e8a7df9c2ee Waiting 10f05dd8b1db Waiting 41dac8b43ba6 Waiting 71a9f6a9ab4d Waiting da3ed5db7103 Waiting c955f6e31a04 Waiting 8f10199ed94b Waiting eca0188f477e Waiting 9fa9226be034 Pulling fs layer 1617e25568b2 Pulling fs layer 6ac0e4adf315 Pulling fs layer f3b09c502777 Pulling fs layer 408012a7b118 Pulling fs layer 44986281b8b9 Pulling fs layer bf70c5107ab5 Pulling fs layer 1ccde423731d Pulling fs layer 408012a7b118 Waiting 44986281b8b9 Waiting 1617e25568b2 Waiting bf70c5107ab5 Waiting 6ac0e4adf315 Waiting 1ccde423731d Waiting 7221d93db8a9 Pulling fs layer 7df673c7455d Pulling fs layer 7221d93db8a9 Waiting 7df673c7455d Waiting 9fa9226be034 Waiting f3b09c502777 Waiting 2d429b9e73a6 Pulling fs layer 46eab5b44a35 Pulling fs layer c4d302cc468d Pulling fs layer 01e0882c90d9 Pulling fs layer 531ee2cf3c0c Pulling fs layer ed54a7dee1d8 Pulling fs layer 12c5c803443f Pulling fs layer e27c75a98748 Pulling fs layer e73cb4a42719 Pulling fs layer a83b68436f09 Pulling fs layer 787d6bee9571 Pulling fs layer 13ff0988aaea Pulling fs layer 4b82842ab819 Pulling fs layer 7e568a0dc8fb Pulling fs layer 2d429b9e73a6 Waiting 46eab5b44a35 Waiting c4d302cc468d Waiting 01e0882c90d9 Waiting a83b68436f09 Waiting 787d6bee9571 Waiting 13ff0988aaea Waiting 4b82842ab819 Waiting 531ee2cf3c0c Waiting ed54a7dee1d8 Waiting 7e568a0dc8fb Waiting e27c75a98748 Waiting 12c5c803443f Waiting e73cb4a42719 Waiting 1e017ebebdbd Pulling fs layer 55f2b468da67 Pulling fs layer 82bfc142787e Pulling fs layer 46baca71a4ef Pulling fs layer b0e0ef7895f4 Pulling fs layer c0c90eeb8aca Pulling fs layer 55f2b468da67 Waiting 82bfc142787e Waiting 1e017ebebdbd Waiting 5cfb27c10ea5 Pulling fs layer 40a5eed61bb0 Pulling fs layer e040ea11fa10 Pulling fs layer 46baca71a4ef Waiting b0e0ef7895f4 Waiting c0c90eeb8aca Waiting 5cfb27c10ea5 Waiting 40a5eed61bb0 Waiting e040ea11fa10 Waiting 09d5a3f70313 Pulling fs layer 356f5c2c843b Pulling fs layer 356f5c2c843b Waiting da9db072f522 Downloading [==============================> ] 2.211MB/3.624MB da9db072f522 Downloading [==============================> ] 2.211MB/3.624MB da9db072f522 Downloading [==============================> ] 2.211MB/3.624MB da9db072f522 Downloading [==============================> ] 2.211MB/3.624MB 795b910b71c0 Downloading [==================================================>] 2.323MB/2.323MB 795b910b71c0 Verifying Checksum 795b910b71c0 Download complete 96e38c8865ba Downloading [==> ] 3.243MB/71.91MB 96e38c8865ba Downloading [==> ] 3.243MB/71.91MB 96e38c8865ba Downloading [==> ] 3.243MB/71.91MB da9db072f522 Downloading [==================================================>] 3.624MB/3.624MB da9db072f522 Verifying Checksum da9db072f522 Verifying Checksum da9db072f522 Verifying Checksum da9db072f522 Downloading [==================================================>] 3.624MB/3.624MB da9db072f522 Download complete da9db072f522 Download complete da9db072f522 Download complete da9db072f522 Download complete da9db072f522 Extracting [> ] 65.54kB/3.624MB da9db072f522 Extracting [> ] 65.54kB/3.624MB da9db072f522 Extracting [> ] 65.54kB/3.624MB da9db072f522 Extracting [> ] 65.54kB/3.624MB d1bdb495a7aa Downloading [> ] 539.6kB/58.78MB 0444d3911dbb Downloading [==================================================>] 1.2kB/1.2kB 0444d3911dbb Verifying Checksum 0444d3911dbb Download complete b801adf990e2 Downloading [==================================================>] 1.17kB/1.17kB b801adf990e2 Downloading [==================================================>] 1.17kB/1.17kB b801adf990e2 Download complete 96e38c8865ba Downloading [=====> ] 7.568MB/71.91MB 96e38c8865ba Downloading [=====> ] 7.568MB/71.91MB 96e38c8865ba Downloading [=====> ] 7.568MB/71.91MB da9db072f522 Extracting [=========> ] 720.9kB/3.624MB da9db072f522 Extracting [=========> ] 720.9kB/3.624MB da9db072f522 Extracting [=========> ] 720.9kB/3.624MB da9db072f522 Extracting [=========> ] 720.9kB/3.624MB d1bdb495a7aa Downloading [==> ] 3.243MB/58.78MB 110a13bd01fb Downloading [> ] 539.6kB/71.86MB da9db072f522 Extracting [==================================================>] 3.624MB/3.624MB da9db072f522 Extracting [==================================================>] 3.624MB/3.624MB da9db072f522 Extracting [==================================================>] 3.624MB/3.624MB da9db072f522 Extracting [==================================================>] 3.624MB/3.624MB 96e38c8865ba Downloading [==========> ] 14.6MB/71.91MB 96e38c8865ba Downloading [==========> ] 14.6MB/71.91MB 96e38c8865ba Downloading [==========> ] 14.6MB/71.91MB d1bdb495a7aa Downloading [======> ] 8.109MB/58.78MB 110a13bd01fb Downloading [=> ] 2.702MB/71.86MB da9db072f522 Pull complete da9db072f522 Pull complete da9db072f522 Pull complete da9db072f522 Pull complete 96e38c8865ba Downloading [===============> ] 22.71MB/71.91MB 96e38c8865ba Downloading [===============> ] 22.71MB/71.91MB 96e38c8865ba Downloading [===============> ] 22.71MB/71.91MB d1bdb495a7aa Downloading [============> ] 15.14MB/58.78MB 110a13bd01fb Downloading [====> ] 6.487MB/71.86MB 96e38c8865ba Downloading [========================> ] 35.14MB/71.91MB 96e38c8865ba Downloading [========================> ] 35.14MB/71.91MB 96e38c8865ba Downloading [========================> ] 35.14MB/71.91MB d1bdb495a7aa Downloading [=====================> ] 24.87MB/58.78MB 110a13bd01fb Downloading [=======> ] 10.81MB/71.86MB 96e38c8865ba Downloading [=================================> ] 47.58MB/71.91MB 96e38c8865ba Downloading [=================================> ] 47.58MB/71.91MB 96e38c8865ba Downloading [=================================> ] 47.58MB/71.91MB d1bdb495a7aa Downloading [==============================> ] 35.68MB/58.78MB 110a13bd01fb Downloading [==========> ] 15.14MB/71.86MB 96e38c8865ba Downloading [=========================================> ] 59.47MB/71.91MB 96e38c8865ba Downloading [=========================================> ] 59.47MB/71.91MB 96e38c8865ba Downloading [=========================================> ] 59.47MB/71.91MB d1bdb495a7aa Downloading [========================================> ] 47.04MB/58.78MB 110a13bd01fb Downloading [=============> ] 19.46MB/71.86MB 96e38c8865ba Downloading [=================================================> ] 70.83MB/71.91MB 96e38c8865ba Downloading [=================================================> ] 70.83MB/71.91MB 96e38c8865ba Downloading [=================================================> ] 70.83MB/71.91MB 96e38c8865ba Verifying Checksum 96e38c8865ba Verifying Checksum 96e38c8865ba Download complete 96e38c8865ba Download complete 96e38c8865ba Download complete d1bdb495a7aa Downloading [=================================================> ] 58.39MB/58.78MB d1bdb495a7aa Verifying Checksum d1bdb495a7aa Download complete d4108afce2f7 Downloading [==================================================>] 1.073kB/1.073kB d4108afce2f7 Download complete 12cf1ed9c784 Downloading [> ] 146.4kB/14.64MB 110a13bd01fb Downloading [=================> ] 25.41MB/71.86MB 07255172bfd8 Downloading [============================> ] 3.003kB/5.24kB 07255172bfd8 Download complete 22c948928e79 Downloading [==================================================>] 1.031kB/1.031kB 22c948928e79 Verifying Checksum 22c948928e79 Download complete e92d65bf8445 Downloading [==================================================>] 1.034kB/1.034kB e92d65bf8445 Download complete 96e38c8865ba Extracting [> ] 557.1kB/71.91MB 96e38c8865ba Extracting [> ] 557.1kB/71.91MB 96e38c8865ba Extracting [> ] 557.1kB/71.91MB 12cf1ed9c784 Downloading [==========================> ] 7.814MB/14.64MB 7910fddefabc Downloading [=======> ] 3.002kB/19.51kB 7910fddefabc Downloading [==================================================>] 19.51kB/19.51kB 7910fddefabc Verifying Checksum 7910fddefabc Download complete 110a13bd01fb Downloading [=======================> ] 34.06MB/71.86MB 5e06c6bed798 Downloading [==================================================>] 296B/296B 5e06c6bed798 Verifying Checksum 5e06c6bed798 Download complete 684be6598fc9 Downloading [=> ] 3.001kB/127.5kB 684be6598fc9 Download complete 12cf1ed9c784 Verifying Checksum 12cf1ed9c784 Download complete 96e38c8865ba Extracting [===> ] 5.014MB/71.91MB 96e38c8865ba Extracting [===> ] 5.014MB/71.91MB 96e38c8865ba Extracting [===> ] 5.014MB/71.91MB 0d92cad902ba Download complete eb7cda286a15 Downloading [==================================================>] 1.119kB/1.119kB eb7cda286a15 Download complete 110a13bd01fb Downloading [=============================> ] 42.17MB/71.86MB e5d7009d9e55 Downloading [==================================================>] 295B/295B e5d7009d9e55 Verifying Checksum e5d7009d9e55 Download complete dcc0c3b2850c Downloading [> ] 539.6kB/76.12MB 1ec5fb03eaee Downloading [=> ] 3.001kB/127kB 1ec5fb03eaee Download complete 96e38c8865ba Extracting [=======> ] 11.14MB/71.91MB 96e38c8865ba Extracting [=======> ] 11.14MB/71.91MB 96e38c8865ba Extracting [=======> ] 11.14MB/71.91MB d3165a332ae3 Downloading [==================================================>] 1.328kB/1.328kB d3165a332ae3 Verifying Checksum d3165a332ae3 Download complete 110a13bd01fb Downloading [==================================> ] 49.74MB/71.86MB c124ba1a8b26 Downloading [> ] 539.6kB/91.87MB dcc0c3b2850c Downloading [=> ] 2.162MB/76.12MB 96e38c8865ba Extracting [===========> ] 16.15MB/71.91MB 96e38c8865ba Extracting [===========> ] 16.15MB/71.91MB 96e38c8865ba Extracting [===========> ] 16.15MB/71.91MB 110a13bd01fb Downloading [=======================================> ] 56.77MB/71.86MB c124ba1a8b26 Downloading [=> ] 3.243MB/91.87MB dcc0c3b2850c Downloading [==> ] 3.784MB/76.12MB 96e38c8865ba Extracting [===============> ] 22.84MB/71.91MB 96e38c8865ba Extracting [===============> ] 22.84MB/71.91MB 96e38c8865ba Extracting [===============> ] 22.84MB/71.91MB 110a13bd01fb Downloading [============================================> ] 63.26MB/71.86MB c124ba1a8b26 Downloading [===> ] 5.946MB/91.87MB dcc0c3b2850c Downloading [===> ] 5.946MB/76.12MB 96e38c8865ba Extracting [====================> ] 28.97MB/71.91MB 96e38c8865ba Extracting [====================> ] 28.97MB/71.91MB 96e38c8865ba Extracting [====================> ] 28.97MB/71.91MB 110a13bd01fb Downloading [================================================> ] 69.75MB/71.86MB 110a13bd01fb Download complete c124ba1a8b26 Downloading [=====> ] 9.19MB/91.87MB 6394804c2196 Downloading [==================================================>] 1.299kB/1.299kB 6394804c2196 Verifying Checksum 6394804c2196 Download complete dcc0c3b2850c Downloading [=====> ] 8.65MB/76.12MB 96e38c8865ba Extracting [========================> ] 34.54MB/71.91MB 96e38c8865ba Extracting [========================> ] 34.54MB/71.91MB 96e38c8865ba Extracting [========================> ] 34.54MB/71.91MB f18232174bc9 Downloading [> ] 48.06kB/3.642MB 110a13bd01fb Extracting [> ] 557.1kB/71.86MB c124ba1a8b26 Downloading [======> ] 12.43MB/91.87MB dcc0c3b2850c Downloading [=======> ] 11.89MB/76.12MB 96e38c8865ba Extracting [===========================> ] 38.99MB/71.91MB 96e38c8865ba Extracting [===========================> ] 38.99MB/71.91MB 96e38c8865ba Extracting [===========================> ] 38.99MB/71.91MB f18232174bc9 Downloading [=============================================> ] 3.292MB/3.642MB f18232174bc9 Verifying Checksum f18232174bc9 Download complete f18232174bc9 Extracting [> ] 65.54kB/3.642MB 110a13bd01fb Extracting [===> ] 5.014MB/71.86MB 9183b65e90ee Downloading [==================================================>] 141B/141B 9183b65e90ee Verifying Checksum 9183b65e90ee Download complete c124ba1a8b26 Downloading [========> ] 15.68MB/91.87MB 3f8d5c908dcc Downloading [> ] 48.06kB/3.524MB 96e38c8865ba Extracting [==============================> ] 43.45MB/71.91MB 96e38c8865ba Extracting [==============================> ] 43.45MB/71.91MB dcc0c3b2850c Downloading [=========> ] 14.6MB/76.12MB 96e38c8865ba Extracting [==============================> ] 43.45MB/71.91MB f18232174bc9 Extracting [=====> ] 393.2kB/3.642MB 110a13bd01fb Extracting [======> ] 10.03MB/71.86MB c124ba1a8b26 Downloading [===========> ] 20.54MB/91.87MB 3f8d5c908dcc Download complete 96e38c8865ba Extracting [================================> ] 46.79MB/71.91MB 96e38c8865ba Extracting [================================> ] 46.79MB/71.91MB 96e38c8865ba Extracting [================================> ] 46.79MB/71.91MB dcc0c3b2850c Downloading [============> ] 19.46MB/76.12MB 30bb92ff0608 Downloading [> ] 97.22kB/8.735MB f18232174bc9 Extracting [==================================================>] 3.642MB/3.642MB f18232174bc9 Extracting [==================================================>] 3.642MB/3.642MB 110a13bd01fb Extracting [=========> ] 13.93MB/71.86MB c124ba1a8b26 Downloading [===============> ] 27.57MB/91.87MB f18232174bc9 Pull complete 9183b65e90ee Extracting [==================================================>] 141B/141B 9183b65e90ee Extracting [==================================================>] 141B/141B dcc0c3b2850c Downloading [================> ] 25.41MB/76.12MB 30bb92ff0608 Downloading [=====> ] 883.6kB/8.735MB 96e38c8865ba Extracting [====================================> ] 51.81MB/71.91MB 96e38c8865ba Extracting [====================================> ] 51.81MB/71.91MB 96e38c8865ba Extracting [====================================> ] 51.81MB/71.91MB 110a13bd01fb Extracting [============> ] 17.83MB/71.86MB c124ba1a8b26 Downloading [===================> ] 36.22MB/91.87MB dcc0c3b2850c Downloading [====================> ] 31.9MB/76.12MB 30bb92ff0608 Downloading [===========> ] 1.965MB/8.735MB 9183b65e90ee Pull complete 3f8d5c908dcc Extracting [> ] 65.54kB/3.524MB 96e38c8865ba Extracting [=======================================> ] 56.26MB/71.91MB 96e38c8865ba Extracting [=======================================> ] 56.26MB/71.91MB 96e38c8865ba Extracting [=======================================> ] 56.26MB/71.91MB 110a13bd01fb Extracting [===============> ] 22.28MB/71.86MB c124ba1a8b26 Downloading [=========================> ] 45.96MB/91.87MB dcc0c3b2850c Downloading [==========================> ] 41.09MB/76.12MB 30bb92ff0608 Downloading [==================> ] 3.243MB/8.735MB 3f8d5c908dcc Extracting [====> ] 327.7kB/3.524MB 96e38c8865ba Extracting [=========================================> ] 60.16MB/71.91MB 96e38c8865ba Extracting [=========================================> ] 60.16MB/71.91MB 96e38c8865ba Extracting [=========================================> ] 60.16MB/71.91MB 110a13bd01fb Extracting [===================> ] 27.85MB/71.86MB c124ba1a8b26 Downloading [===============================> ] 57.31MB/91.87MB 3f8d5c908dcc Extracting [==================================================>] 3.524MB/3.524MB 3f8d5c908dcc Extracting [==================================================>] 3.524MB/3.524MB dcc0c3b2850c Downloading [================================> ] 49.74MB/76.12MB 30bb92ff0608 Downloading [===========================> ] 4.718MB/8.735MB 96e38c8865ba Extracting [============================================> ] 64.06MB/71.91MB 96e38c8865ba Extracting [============================================> ] 64.06MB/71.91MB 96e38c8865ba Extracting [============================================> ] 64.06MB/71.91MB 110a13bd01fb Extracting [======================> ] 32.87MB/71.86MB c124ba1a8b26 Downloading [====================================> ] 67.58MB/91.87MB 30bb92ff0608 Downloading [====================================> ] 6.29MB/8.735MB dcc0c3b2850c Downloading [======================================> ] 58.93MB/76.12MB 3f8d5c908dcc Pull complete 96e38c8865ba Extracting [================================================> ] 69.07MB/71.91MB 96e38c8865ba Extracting [================================================> ] 69.07MB/71.91MB 96e38c8865ba Extracting [================================================> ] 69.07MB/71.91MB 110a13bd01fb Extracting [==========================> ] 38.44MB/71.86MB c124ba1a8b26 Downloading [==========================================> ] 77.86MB/91.87MB 30bb92ff0608 Downloading [==============================================> ] 8.158MB/8.735MB dcc0c3b2850c Downloading [=============================================> ] 68.66MB/76.12MB 96e38c8865ba Extracting [==================================================>] 71.91MB/71.91MB 96e38c8865ba Extracting [==================================================>] 71.91MB/71.91MB 96e38c8865ba Extracting [==================================================>] 71.91MB/71.91MB 30bb92ff0608 Verifying Checksum 30bb92ff0608 Download complete 30bb92ff0608 Extracting [> ] 98.3kB/8.735MB 110a13bd01fb Extracting [==============================> ] 43.45MB/71.86MB 807a2e881ecd Downloading [==> ] 3.01kB/58.07kB dcc0c3b2850c Verifying Checksum dcc0c3b2850c Download complete c124ba1a8b26 Downloading [===============================================> ] 88.13MB/91.87MB 807a2e881ecd Downloading [==================================================>] 58.07kB/58.07kB 807a2e881ecd Download complete 96e38c8865ba Pull complete 96e38c8865ba Pull complete 96e38c8865ba Pull complete c124ba1a8b26 Verifying Checksum c124ba1a8b26 Download complete 4a4d0948b0bf Downloading [=====> ] 3.01kB/27.78kB 4a4d0948b0bf Downloading [==================================================>] 27.78kB/27.78kB 4a4d0948b0bf Verifying Checksum 4a4d0948b0bf Download complete 795b910b71c0 Extracting [> ] 32.77kB/2.323MB 5e06c6bed798 Extracting [==================================================>] 296B/296B e5d7009d9e55 Extracting [==================================================>] 295B/295B e5d7009d9e55 Extracting [==================================================>] 295B/295B 5e06c6bed798 Extracting [==================================================>] 296B/296B 7009d5001b77 Downloading [============> ] 3.01kB/11.92kB 7009d5001b77 Download complete 04f6155c873d Downloading [> ] 539.6kB/107.3MB 30bb92ff0608 Extracting [======> ] 1.18MB/8.735MB 538deb30e80c Downloading [==================================================>] 1.225kB/1.225kB 538deb30e80c Verifying Checksum 538deb30e80c Download complete 85dde7dceb0a Downloading [> ] 539.6kB/63.48MB 110a13bd01fb Extracting [================================> ] 46.79MB/71.86MB 795b910b71c0 Extracting [=========> ] 458.8kB/2.323MB 795b910b71c0 Extracting [==================================================>] 2.323MB/2.323MB 795b910b71c0 Extracting [==================================================>] 2.323MB/2.323MB 04f6155c873d Downloading [====> ] 9.731MB/107.3MB 30bb92ff0608 Extracting [=========================> ] 4.424MB/8.735MB 5e06c6bed798 Pull complete 684be6598fc9 Extracting [============> ] 32.77kB/127.5kB 684be6598fc9 Extracting [==================================================>] 127.5kB/127.5kB e5d7009d9e55 Pull complete 795b910b71c0 Pull complete 1ec5fb03eaee Extracting [============> ] 32.77kB/127kB 1ec5fb03eaee Extracting [==================================================>] 127kB/127kB eca0188f477e Downloading [> ] 375.7kB/37.17MB 1ec5fb03eaee Extracting [==================================================>] 127kB/127kB 85dde7dceb0a Downloading [==> ] 2.702MB/63.48MB 110a13bd01fb Extracting [===================================> ] 51.25MB/71.86MB 04f6155c873d Downloading [=========> ] 19.46MB/107.3MB 30bb92ff0608 Extracting [===============================================> ] 8.258MB/8.735MB 30bb92ff0608 Extracting [==================================================>] 8.735MB/8.735MB 110a13bd01fb Extracting [=====================================> ] 54.03MB/71.86MB 684be6598fc9 Pull complete 1ec5fb03eaee Pull complete 85dde7dceb0a Downloading [===> ] 4.865MB/63.48MB 0d92cad902ba Extracting [==================================================>] 1.148kB/1.148kB 0d92cad902ba Extracting [==================================================>] 1.148kB/1.148kB d3165a332ae3 Extracting [==================================================>] 1.328kB/1.328kB d3165a332ae3 Extracting [==================================================>] 1.328kB/1.328kB d1bdb495a7aa Extracting [> ] 557.1kB/58.78MB 30bb92ff0608 Pull complete 807a2e881ecd Extracting [============================> ] 32.77kB/58.07kB eca0188f477e Downloading [=> ] 1.129MB/37.17MB 807a2e881ecd Extracting [==================================================>] 58.07kB/58.07kB 04f6155c873d Downloading [=============> ] 29.2MB/107.3MB 85dde7dceb0a Downloading [=====> ] 7.028MB/63.48MB 110a13bd01fb Extracting [=======================================> ] 56.82MB/71.86MB d1bdb495a7aa Extracting [=========> ] 10.58MB/58.78MB 04f6155c873d Downloading [==================> ] 39.47MB/107.3MB eca0188f477e Downloading [==> ] 1.883MB/37.17MB 85dde7dceb0a Downloading [=======> ] 9.19MB/63.48MB 110a13bd01fb Extracting [==========================================> ] 61.28MB/71.86MB d1bdb495a7aa Extracting [=====================> ] 25.07MB/58.78MB 04f6155c873d Downloading [=======================> ] 50.28MB/107.3MB eca0188f477e Downloading [===> ] 2.637MB/37.17MB 110a13bd01fb Extracting [================================================> ] 69.07MB/71.86MB d1bdb495a7aa Extracting [=================================> ] 38.99MB/58.78MB 85dde7dceb0a Downloading [=========> ] 11.89MB/63.48MB 04f6155c873d Downloading [===========================> ] 58.93MB/107.3MB eca0188f477e Downloading [====> ] 3.39MB/37.17MB 110a13bd01fb Extracting [==================================================>] 71.86MB/71.86MB 110a13bd01fb Extracting [==================================================>] 71.86MB/71.86MB d1bdb495a7aa Extracting [========================================> ] 47.35MB/58.78MB 85dde7dceb0a Downloading [===========> ] 14.6MB/63.48MB 0d92cad902ba Pull complete 04f6155c873d Downloading [=================================> ] 70.83MB/107.3MB eca0188f477e Downloading [======> ] 4.521MB/37.17MB d1bdb495a7aa Extracting [============================================> ] 51.81MB/58.78MB 85dde7dceb0a Downloading [=============> ] 17.3MB/63.48MB 04f6155c873d Downloading [======================================> ] 82.72MB/107.3MB d1bdb495a7aa Extracting [==================================================>] 58.78MB/58.78MB eca0188f477e Downloading [=======> ] 5.651MB/37.17MB 85dde7dceb0a Downloading [===============> ] 20MB/63.48MB 04f6155c873d Downloading [===========================================> ] 93.54MB/107.3MB d3165a332ae3 Pull complete 807a2e881ecd Pull complete 110a13bd01fb Pull complete d1bdb495a7aa Pull complete 0444d3911dbb Extracting [==================================================>] 1.2kB/1.2kB 0444d3911dbb Extracting [==================================================>] 1.2kB/1.2kB 12cf1ed9c784 Extracting [> ] 163.8kB/14.64MB 4a4d0948b0bf Extracting [==================================================>] 27.78kB/27.78kB 4a4d0948b0bf Extracting [==================================================>] 27.78kB/27.78kB eca0188f477e Downloading [=========> ] 6.782MB/37.17MB 85dde7dceb0a Downloading [=================> ] 22.71MB/63.48MB 04f6155c873d Downloading [================================================> ] 104.3MB/107.3MB dcc0c3b2850c Extracting [> ] 557.1kB/76.12MB 04f6155c873d Verifying Checksum 04f6155c873d Download complete c124ba1a8b26 Extracting [> ] 557.1kB/91.87MB 12cf1ed9c784 Extracting [===> ] 1.147MB/14.64MB e444bcd4d577 Downloading [==================================================>] 279B/279B e444bcd4d577 Verifying Checksum e444bcd4d577 Download complete eca0188f477e Downloading [===========> ] 8.289MB/37.17MB 0444d3911dbb Pull complete b801adf990e2 Extracting [==================================================>] 1.17kB/1.17kB b801adf990e2 Extracting [==================================================>] 1.17kB/1.17kB 4a4d0948b0bf Pull complete 85dde7dceb0a Downloading [====================> ] 26.49MB/63.48MB dcc0c3b2850c Extracting [======> ] 10.58MB/76.12MB c124ba1a8b26 Extracting [=====> ] 9.47MB/91.87MB eabd8714fec9 Downloading [> ] 539.6kB/375MB 12cf1ed9c784 Extracting [================> ] 4.915MB/14.64MB eca0188f477e Downloading [=============> ] 10.17MB/37.17MB 04f6155c873d Extracting [> ] 557.1kB/107.3MB 85dde7dceb0a Downloading [========================> ] 30.82MB/63.48MB dcc0c3b2850c Extracting [============> ] 18.94MB/76.12MB c124ba1a8b26 Extracting [==========> ] 19.5MB/91.87MB b801adf990e2 Pull complete 12cf1ed9c784 Extracting [=====================> ] 6.39MB/14.64MB eabd8714fec9 Downloading [> ] 6.487MB/375MB xacml-pdp Pulled eca0188f477e Downloading [================> ] 12.06MB/37.17MB 04f6155c873d Extracting [=> ] 2.785MB/107.3MB dcc0c3b2850c Extracting [=================> ] 26.18MB/76.12MB 85dde7dceb0a Downloading [===========================> ] 35.14MB/63.48MB c124ba1a8b26 Extracting [=============> ] 25.62MB/91.87MB 12cf1ed9c784 Extracting [==========================> ] 7.7MB/14.64MB eabd8714fec9 Downloading [==> ] 16.22MB/375MB eca0188f477e Downloading [==================> ] 13.94MB/37.17MB dcc0c3b2850c Extracting [========================> ] 37.88MB/76.12MB 85dde7dceb0a Downloading [===============================> ] 39.47MB/63.48MB 04f6155c873d Extracting [==> ] 6.128MB/107.3MB c124ba1a8b26 Extracting [=================> ] 32.87MB/91.87MB eabd8714fec9 Downloading [===> ] 23.79MB/375MB 12cf1ed9c784 Extracting [=============================> ] 8.684MB/14.64MB eca0188f477e Downloading [=====================> ] 15.83MB/37.17MB dcc0c3b2850c Extracting [================================> ] 49.02MB/76.12MB 85dde7dceb0a Downloading [==================================> ] 44.33MB/63.48MB 04f6155c873d Extracting [====> ] 10.03MB/107.3MB c124ba1a8b26 Extracting [========================> ] 45.68MB/91.87MB eabd8714fec9 Downloading [====> ] 31.36MB/375MB 12cf1ed9c784 Extracting [======================================> ] 11.3MB/14.64MB eca0188f477e Downloading [========================> ] 18.09MB/37.17MB dcc0c3b2850c Extracting [====================================> ] 56.26MB/76.12MB 85dde7dceb0a Downloading [======================================> ] 48.66MB/63.48MB c124ba1a8b26 Extracting [==============================> ] 56.82MB/91.87MB 04f6155c873d Extracting [======> ] 13.37MB/107.3MB eabd8714fec9 Downloading [=====> ] 38.93MB/375MB 12cf1ed9c784 Extracting [===========================================> ] 12.62MB/14.64MB 12cf1ed9c784 Extracting [==================================================>] 14.64MB/14.64MB eca0188f477e Downloading [===========================> ] 20.35MB/37.17MB dcc0c3b2850c Extracting [===========================================> ] 66.29MB/76.12MB 85dde7dceb0a Downloading [==========================================> ] 53.53MB/63.48MB 12cf1ed9c784 Pull complete d4108afce2f7 Extracting [==================================================>] 1.073kB/1.073kB d4108afce2f7 Extracting [==================================================>] 1.073kB/1.073kB c124ba1a8b26 Extracting [====================================> ] 66.85MB/91.87MB 04f6155c873d Extracting [=======> ] 16.15MB/107.3MB eabd8714fec9 Downloading [======> ] 46.5MB/375MB eca0188f477e Downloading [==============================> ] 22.61MB/37.17MB dcc0c3b2850c Extracting [==================================================>] 76.12MB/76.12MB 85dde7dceb0a Downloading [=============================================> ] 58.39MB/63.48MB dcc0c3b2850c Pull complete eb7cda286a15 Extracting [==================================================>] 1.119kB/1.119kB eb7cda286a15 Extracting [==================================================>] 1.119kB/1.119kB c124ba1a8b26 Extracting [==========================================> ] 77.99MB/91.87MB 04f6155c873d Extracting [========> ] 17.27MB/107.3MB eabd8714fec9 Downloading [=======> ] 54.07MB/375MB d4108afce2f7 Pull complete 07255172bfd8 Extracting [==================================================>] 5.24kB/5.24kB 07255172bfd8 Extracting [==================================================>] 5.24kB/5.24kB eca0188f477e Downloading [=================================> ] 25.25MB/37.17MB 85dde7dceb0a Downloading [=================================================> ] 63.26MB/63.48MB 85dde7dceb0a Verifying Checksum 85dde7dceb0a Download complete c124ba1a8b26 Extracting [=================================================> ] 91.36MB/91.87MB c124ba1a8b26 Extracting [==================================================>] 91.87MB/91.87MB 45fd2fec8a19 Downloading [==================================================>] 1.103kB/1.103kB 45fd2fec8a19 Verifying Checksum 45fd2fec8a19 Download complete 04f6155c873d Extracting [========> ] 18.38MB/107.3MB eabd8714fec9 Downloading [========> ] 62.72MB/375MB c124ba1a8b26 Pull complete 6394804c2196 Extracting [==================================================>] 1.299kB/1.299kB 6394804c2196 Extracting [==================================================>] 1.299kB/1.299kB eb7cda286a15 Pull complete eca0188f477e Downloading [=======================================> ] 29.39MB/37.17MB api Pulled 07255172bfd8 Pull complete 22c948928e79 Extracting [==================================================>] 1.031kB/1.031kB 22c948928e79 Extracting [==================================================>] 1.031kB/1.031kB 8f10199ed94b Downloading [> ] 97.22kB/8.768MB eabd8714fec9 Downloading [==========> ] 78.4MB/375MB 04f6155c873d Extracting [==========> ] 22.84MB/107.3MB 6394804c2196 Pull complete eca0188f477e Downloading [==============================================> ] 34.67MB/37.17MB pap Pulled 8f10199ed94b Downloading [============================> ] 4.914MB/8.768MB 22c948928e79 Pull complete e92d65bf8445 Extracting [==================================================>] 1.034kB/1.034kB e92d65bf8445 Extracting [==================================================>] 1.034kB/1.034kB eca0188f477e Verifying Checksum eca0188f477e Download complete eabd8714fec9 Downloading [============> ] 91.37MB/375MB 04f6155c873d Extracting [=============> ] 28.97MB/107.3MB 8f10199ed94b Verifying Checksum 8f10199ed94b Download complete f963a77d2726 Downloading [=======> ] 3.01kB/21.44kB f963a77d2726 Downloading [==================================================>] 21.44kB/21.44kB f963a77d2726 Download complete e92d65bf8445 Pull complete 7910fddefabc Extracting [==================================================>] 19.51kB/19.51kB 7910fddefabc Extracting [==================================================>] 19.51kB/19.51kB 79161a3f5362 Downloading [================================> ] 3.011kB/4.656kB 79161a3f5362 Downloading [==================================================>] 4.656kB/4.656kB 79161a3f5362 Verifying Checksum 79161a3f5362 Download complete eca0188f477e Extracting [> ] 393.2kB/37.17MB eabd8714fec9 Downloading [==============> ] 105.4MB/375MB 9c266ba63f51 Downloading [==================================================>] 1.105kB/1.105kB 9c266ba63f51 Verifying Checksum 9c266ba63f51 Download complete 04f6155c873d Extracting [================> ] 34.54MB/107.3MB f3a82e9f1761 Downloading [> ] 457.7kB/44.41MB 2e8a7df9c2ee Download complete eca0188f477e Extracting [======> ] 4.719MB/37.17MB 10f05dd8b1db Downloading [==================================================>] 98B/98B 10f05dd8b1db Verifying Checksum 10f05dd8b1db Download complete eabd8714fec9 Downloading [================> ] 121.7MB/375MB f3a82e9f1761 Downloading [====> ] 3.669MB/44.41MB 04f6155c873d Extracting [=================> ] 38.44MB/107.3MB 41dac8b43ba6 Download complete eca0188f477e Extracting [=========> ] 7.078MB/37.17MB 71a9f6a9ab4d Downloading [> ] 3.009kB/230.6kB 7910fddefabc Pull complete policy-db-migrator Pulled eabd8714fec9 Downloading [=================> ] 131.9MB/375MB 71a9f6a9ab4d Downloading [==================================================>] 230.6kB/230.6kB 71a9f6a9ab4d Verifying Checksum 71a9f6a9ab4d Download complete f3a82e9f1761 Downloading [==========> ] 9.174MB/44.41MB 04f6155c873d Extracting [===================> ] 41.22MB/107.3MB eca0188f477e Extracting [===============> ] 11.4MB/37.17MB eabd8714fec9 Downloading [===================> ] 147.1MB/375MB f3a82e9f1761 Downloading [=============> ] 12.39MB/44.41MB 04f6155c873d Extracting [=====================> ] 46.24MB/107.3MB eca0188f477e Extracting [=======================> ] 17.3MB/37.17MB f3a82e9f1761 Downloading [==============> ] 12.84MB/44.41MB eabd8714fec9 Downloading [===================> ] 148.7MB/375MB 04f6155c873d Extracting [=======================> ] 51.25MB/107.3MB da3ed5db7103 Downloading [> ] 539.6kB/127.4MB eca0188f477e Extracting [================================> ] 24.38MB/37.17MB eabd8714fec9 Downloading [====================> ] 153.5MB/375MB 04f6155c873d Extracting [=========================> ] 55.71MB/107.3MB f3a82e9f1761 Downloading [=================> ] 15.14MB/44.41MB eca0188f477e Extracting [=========================================> ] 30.67MB/37.17MB eabd8714fec9 Downloading [====================> ] 155.2MB/375MB da3ed5db7103 Downloading [> ] 1.621MB/127.4MB 04f6155c873d Extracting [============================> ] 60.16MB/107.3MB f3a82e9f1761 Downloading [===================> ] 16.97MB/44.41MB eca0188f477e Extracting [==============================================> ] 34.6MB/37.17MB eabd8714fec9 Downloading [=====================> ] 161.7MB/375MB 04f6155c873d Extracting [==============================> ] 64.62MB/107.3MB eca0188f477e Extracting [==================================================>] 37.17MB/37.17MB f3a82e9f1761 Downloading [======================> ] 19.73MB/44.41MB da3ed5db7103 Downloading [=> ] 3.243MB/127.4MB eabd8714fec9 Downloading [=====================> ] 163.3MB/375MB 04f6155c873d Extracting [==============================> ] 66.29MB/107.3MB f3a82e9f1761 Downloading [=======================> ] 21.1MB/44.41MB eca0188f477e Pull complete e444bcd4d577 Extracting [==================================================>] 279B/279B e444bcd4d577 Extracting [==================================================>] 279B/279B da3ed5db7103 Downloading [==> ] 6.487MB/127.4MB eabd8714fec9 Downloading [=======================> ] 175.2MB/375MB 04f6155c873d Extracting [================================> ] 70.19MB/107.3MB f3a82e9f1761 Downloading [================================> ] 28.9MB/44.41MB da3ed5db7103 Downloading [=====> ] 14.06MB/127.4MB eabd8714fec9 Downloading [=========================> ] 189.2MB/375MB e444bcd4d577 Pull complete 04f6155c873d Extracting [==================================> ] 74.65MB/107.3MB f3a82e9f1761 Downloading [=====================================> ] 33.49MB/44.41MB da3ed5db7103 Downloading [========> ] 22.71MB/127.4MB eabd8714fec9 Downloading [===========================> ] 206MB/375MB 04f6155c873d Extracting [====================================> ] 79.1MB/107.3MB f3a82e9f1761 Downloading [===========================================> ] 38.99MB/44.41MB da3ed5db7103 Downloading [==========> ] 27.57MB/127.4MB eabd8714fec9 Downloading [=============================> ] 224.4MB/375MB 04f6155c873d Extracting [=======================================> ] 84.12MB/107.3MB f3a82e9f1761 Downloading [=================================================> ] 44.04MB/44.41MB f3a82e9f1761 Verifying Checksum f3a82e9f1761 Download complete da3ed5db7103 Downloading [============> ] 32.44MB/127.4MB eabd8714fec9 Downloading [================================> ] 241.7MB/375MB c955f6e31a04 Downloading [===========================================> ] 3.011kB/3.446kB c955f6e31a04 Downloading [==================================================>] 3.446kB/3.446kB c955f6e31a04 Verifying Checksum c955f6e31a04 Download complete 04f6155c873d Extracting [===========================================> ] 93.59MB/107.3MB 9fa9226be034 Downloading [> ] 15.3kB/783kB da3ed5db7103 Downloading [==============> ] 37.31MB/127.4MB eabd8714fec9 Downloading [==================================> ] 260.1MB/375MB 9fa9226be034 Downloading [==================================================>] 783kB/783kB 9fa9226be034 Verifying Checksum 9fa9226be034 Download complete 9fa9226be034 Extracting [==> ] 32.77kB/783kB 04f6155c873d Extracting [===============================================> ] 101.4MB/107.3MB 1617e25568b2 Downloading [=> ] 15.3kB/480.9kB eabd8714fec9 Downloading [=====================================> ] 280.6MB/375MB da3ed5db7103 Downloading [=================> ] 43.79MB/127.4MB 1617e25568b2 Download complete 9fa9226be034 Extracting [=======================> ] 360.4kB/783kB 9fa9226be034 Extracting [==================================================>] 783kB/783kB 9fa9226be034 Extracting [==================================================>] 783kB/783kB eabd8714fec9 Downloading [=======================================> ] 297.9MB/375MB da3ed5db7103 Downloading [===================> ] 50.28MB/127.4MB 04f6155c873d Extracting [================================================> ] 104.2MB/107.3MB 9fa9226be034 Pull complete 1617e25568b2 Extracting [===> ] 32.77kB/480.9kB 6ac0e4adf315 Downloading [> ] 539.6kB/62.07MB eabd8714fec9 Downloading [=========================================> ] 312.5MB/375MB da3ed5db7103 Downloading [======================> ] 57.85MB/127.4MB 1617e25568b2 Extracting [========================================> ] 393.2kB/480.9kB 04f6155c873d Extracting [=================================================> ] 105.3MB/107.3MB 6ac0e4adf315 Downloading [=> ] 1.621MB/62.07MB 1617e25568b2 Extracting [==================================================>] 480.9kB/480.9kB 1617e25568b2 Extracting [==================================================>] 480.9kB/480.9kB eabd8714fec9 Downloading [===========================================> ] 328.2MB/375MB da3ed5db7103 Downloading [===========================> ] 69.75MB/127.4MB 04f6155c873d Extracting [==================================================>] 107.3MB/107.3MB 6ac0e4adf315 Downloading [===> ] 3.784MB/62.07MB 04f6155c873d Pull complete eabd8714fec9 Downloading [============================================> ] 337.4MB/375MB da3ed5db7103 Downloading [==============================> ] 76.77MB/127.4MB 1617e25568b2 Pull complete 6ac0e4adf315 Downloading [======> ] 7.568MB/62.07MB eabd8714fec9 Downloading [===============================================> ] 353.6MB/375MB da3ed5db7103 Downloading [====================================> ] 93.54MB/127.4MB 85dde7dceb0a Extracting [> ] 557.1kB/63.48MB eabd8714fec9 Downloading [=================================================> ] 372.5MB/375MB eabd8714fec9 Verifying Checksum eabd8714fec9 Download complete 6ac0e4adf315 Downloading [=======> ] 9.731MB/62.07MB da3ed5db7103 Downloading [==========================================> ] 109.2MB/127.4MB 85dde7dceb0a Extracting [=> ] 1.671MB/63.48MB f3b09c502777 Downloading [> ] 539.6kB/56.52MB eabd8714fec9 Extracting [> ] 557.1kB/375MB da3ed5db7103 Downloading [================================================> ] 124.4MB/127.4MB 6ac0e4adf315 Downloading [===========> ] 14.06MB/62.07MB da3ed5db7103 Verifying Checksum da3ed5db7103 Download complete 408012a7b118 Downloading [==================================================>] 637B/637B 408012a7b118 Download complete eabd8714fec9 Extracting [=> ] 14.48MB/375MB f3b09c502777 Downloading [=====> ] 5.946MB/56.52MB 85dde7dceb0a Extracting [==> ] 2.785MB/63.48MB 44986281b8b9 Downloading [=====================================> ] 3.011kB/4.022kB 44986281b8b9 Downloading [==================================================>] 4.022kB/4.022kB 44986281b8b9 Verifying Checksum 44986281b8b9 Download complete 6ac0e4adf315 Downloading [==============> ] 17.84MB/62.07MB bf70c5107ab5 Downloading [==================================================>] 1.44kB/1.44kB bf70c5107ab5 Verifying Checksum bf70c5107ab5 Download complete f3b09c502777 Downloading [===============> ] 17.3MB/56.52MB eabd8714fec9 Extracting [==> ] 21.73MB/375MB 1ccde423731d Downloading [==> ] 3.01kB/61.44kB 1ccde423731d Downloading [==================================================>] 61.44kB/61.44kB 1ccde423731d Verifying Checksum 1ccde423731d Download complete 6ac0e4adf315 Downloading [=================> ] 21.63MB/62.07MB 85dde7dceb0a Extracting [===> ] 4.456MB/63.48MB 7221d93db8a9 Downloading [==================================================>] 100B/100B 7221d93db8a9 Download complete f3b09c502777 Downloading [=====================> ] 24.33MB/56.52MB 7df673c7455d Downloading [==================================================>] 694B/694B 7df673c7455d Verifying Checksum 7df673c7455d Download complete eabd8714fec9 Extracting [===> ] 23.95MB/375MB 6ac0e4adf315 Downloading [====================> ] 24.87MB/62.07MB 85dde7dceb0a Extracting [===> ] 5.014MB/63.48MB 2d429b9e73a6 Downloading [> ] 293.8kB/29.13MB f3b09c502777 Downloading [===========================> ] 31.36MB/56.52MB eabd8714fec9 Extracting [====> ] 32.31MB/375MB 6ac0e4adf315 Downloading [======================> ] 28.11MB/62.07MB 85dde7dceb0a Extracting [======> ] 7.799MB/63.48MB 2d429b9e73a6 Downloading [========> ] 4.718MB/29.13MB f3b09c502777 Downloading [=================================> ] 37.85MB/56.52MB eabd8714fec9 Extracting [======> ] 45.68MB/375MB 6ac0e4adf315 Downloading [===========================> ] 33.52MB/62.07MB 2d429b9e73a6 Downloading [======================> ] 13.27MB/29.13MB 85dde7dceb0a Extracting [=======> ] 10.03MB/63.48MB f3b09c502777 Downloading [=========================================> ] 46.5MB/56.52MB eabd8714fec9 Extracting [=======> ] 54.59MB/375MB 6ac0e4adf315 Downloading [==============================> ] 38.39MB/62.07MB 2d429b9e73a6 Downloading [==================================> ] 20.05MB/29.13MB f3b09c502777 Downloading [==============================================> ] 52.44MB/56.52MB 85dde7dceb0a Extracting [==========> ] 12.81MB/63.48MB eabd8714fec9 Extracting [=========> ] 69.07MB/375MB 6ac0e4adf315 Downloading [=================================> ] 41.09MB/62.07MB f3b09c502777 Verifying Checksum f3b09c502777 Download complete 2d429b9e73a6 Downloading [===========================================> ] 25.36MB/29.13MB eabd8714fec9 Extracting [==========> ] 81.89MB/375MB 85dde7dceb0a Extracting [============> ] 16.15MB/63.48MB 46eab5b44a35 Downloading [==================================================>] 1.168kB/1.168kB 46eab5b44a35 Verifying Checksum 46eab5b44a35 Download complete 2d429b9e73a6 Verifying Checksum 2d429b9e73a6 Download complete 6ac0e4adf315 Downloading [======================================> ] 48.12MB/62.07MB 01e0882c90d9 Downloading [> ] 15.3kB/1.447MB c4d302cc468d Downloading [> ] 48.06kB/4.534MB 85dde7dceb0a Extracting [=============> ] 17.27MB/63.48MB eabd8714fec9 Extracting [===========> ] 89.69MB/375MB c4d302cc468d Verifying Checksum c4d302cc468d Download complete 2d429b9e73a6 Extracting [> ] 294.9kB/29.13MB 6ac0e4adf315 Downloading [================================================> ] 60.55MB/62.07MB 01e0882c90d9 Downloading [=================================================> ] 1.424MB/1.447MB 01e0882c90d9 Verifying Checksum 01e0882c90d9 Download complete 6ac0e4adf315 Verifying Checksum 6ac0e4adf315 Download complete ed54a7dee1d8 Downloading [> ] 15.3kB/1.196MB 12c5c803443f Downloading [==================================================>] 116B/116B 12c5c803443f Verifying Checksum 12c5c803443f Download complete 531ee2cf3c0c Downloading [> ] 80.83kB/8.066MB e27c75a98748 Downloading [===============================================> ] 3.011kB/3.144kB e27c75a98748 Download complete eabd8714fec9 Extracting [=============> ] 98.6MB/375MB ed54a7dee1d8 Verifying Checksum ed54a7dee1d8 Download complete 85dde7dceb0a Extracting [================> ] 20.61MB/63.48MB a83b68436f09 Downloading [===============> ] 3.011kB/9.919kB a83b68436f09 Downloading [==================================================>] 9.919kB/9.919kB a83b68436f09 Verifying Checksum a83b68436f09 Download complete e73cb4a42719 Downloading [> ] 539.6kB/109.1MB 787d6bee9571 Download complete 2d429b9e73a6 Extracting [======> ] 3.539MB/29.13MB 13ff0988aaea Downloading [==================================================>] 167B/167B 13ff0988aaea Verifying Checksum 13ff0988aaea Download complete 4b82842ab819 Downloading [===========================> ] 3.011kB/5.415kB 4b82842ab819 Downloading [==================================================>] 5.415kB/5.415kB 4b82842ab819 Verifying Checksum 4b82842ab819 Download complete 531ee2cf3c0c Downloading [=========================================> ] 6.716MB/8.066MB 6ac0e4adf315 Extracting [> ] 557.1kB/62.07MB 531ee2cf3c0c Verifying Checksum 531ee2cf3c0c Download complete 7e568a0dc8fb Downloading [==================================================>] 184B/184B 7e568a0dc8fb Verifying Checksum 7e568a0dc8fb Download complete eabd8714fec9 Extracting [==============> ] 105.8MB/375MB e73cb4a42719 Downloading [==> ] 5.946MB/109.1MB 1e017ebebdbd Downloading [> ] 375.7kB/37.19MB 55f2b468da67 Downloading [> ] 539.6kB/257.9MB 85dde7dceb0a Extracting [==================> ] 23.4MB/63.48MB 2d429b9e73a6 Extracting [===========> ] 6.783MB/29.13MB 6ac0e4adf315 Extracting [===> ] 3.899MB/62.07MB e73cb4a42719 Downloading [=====> ] 12.98MB/109.1MB 1e017ebebdbd Downloading [====> ] 3.014MB/37.19MB eabd8714fec9 Extracting [==============> ] 110.9MB/375MB 85dde7dceb0a Extracting [====================> ] 26.18MB/63.48MB 55f2b468da67 Downloading [> ] 4.324MB/257.9MB 2d429b9e73a6 Extracting [================> ] 9.732MB/29.13MB 6ac0e4adf315 Extracting [=====> ] 6.685MB/62.07MB e73cb4a42719 Downloading [===========> ] 24.87MB/109.1MB 85dde7dceb0a Extracting [======================> ] 28.41MB/63.48MB 1e017ebebdbd Downloading [============> ] 9.043MB/37.19MB 55f2b468da67 Downloading [==> ] 11.35MB/257.9MB eabd8714fec9 Extracting [===============> ] 113.6MB/375MB 2d429b9e73a6 Extracting [======================> ] 13.27MB/29.13MB 6ac0e4adf315 Extracting [========> ] 10.03MB/62.07MB e73cb4a42719 Downloading [=================> ] 38.39MB/109.1MB 55f2b468da67 Downloading [===> ] 18.38MB/257.9MB 1e017ebebdbd Downloading [====================> ] 15.07MB/37.19MB eabd8714fec9 Extracting [===============> ] 117MB/375MB 85dde7dceb0a Extracting [========================> ] 30.64MB/63.48MB 2d429b9e73a6 Extracting [================================> ] 19.17MB/29.13MB 6ac0e4adf315 Extracting [==========> ] 13.37MB/62.07MB e73cb4a42719 Downloading [=======================> ] 51.9MB/109.1MB 55f2b468da67 Downloading [====> ] 24.87MB/257.9MB 1e017ebebdbd Downloading [============================> ] 21.1MB/37.19MB eabd8714fec9 Extracting [================> ] 121.4MB/375MB 85dde7dceb0a Extracting [=========================> ] 32.87MB/63.48MB 2d429b9e73a6 Extracting [=======================================> ] 23MB/29.13MB 6ac0e4adf315 Extracting [=============> ] 17.27MB/62.07MB e73cb4a42719 Downloading [=============================> ] 63.8MB/109.1MB 55f2b468da67 Downloading [======> ] 31.9MB/257.9MB 1e017ebebdbd Downloading [====================================> ] 27.13MB/37.19MB eabd8714fec9 Extracting [================> ] 124.8MB/375MB 85dde7dceb0a Extracting [============================> ] 35.65MB/63.48MB 2d429b9e73a6 Extracting [==========================================> ] 24.77MB/29.13MB e73cb4a42719 Downloading [====================================> ] 78.94MB/109.1MB 6ac0e4adf315 Extracting [==================> ] 23.4MB/62.07MB 55f2b468da67 Downloading [========> ] 41.63MB/257.9MB 1e017ebebdbd Verifying Checksum 1e017ebebdbd Download complete eabd8714fec9 Extracting [=================> ] 128.1MB/375MB 82bfc142787e Downloading [> ] 97.22kB/8.613MB 85dde7dceb0a Extracting [==============================> ] 38.44MB/63.48MB 2d429b9e73a6 Extracting [===============================================> ] 27.72MB/29.13MB e73cb4a42719 Downloading [===========================================> ] 94.62MB/109.1MB 55f2b468da67 Downloading [==========> ] 52.44MB/257.9MB eabd8714fec9 Extracting [=================> ] 131.5MB/375MB 6ac0e4adf315 Extracting [===================> ] 24.51MB/62.07MB 82bfc142787e Downloading [============> ] 2.162MB/8.613MB 1e017ebebdbd Extracting [> ] 393.2kB/37.19MB 85dde7dceb0a Extracting [================================> ] 41.78MB/63.48MB e73cb4a42719 Verifying Checksum e73cb4a42719 Download complete 55f2b468da67 Downloading [============> ] 63.8MB/257.9MB 46baca71a4ef Downloading [========> ] 3.01kB/18.11kB 46baca71a4ef Downloading [==================================================>] 18.11kB/18.11kB 46baca71a4ef Verifying Checksum 46baca71a4ef Download complete 6ac0e4adf315 Extracting [======================> ] 27.85MB/62.07MB 82bfc142787e Downloading [=====================> ] 3.636MB/8.613MB eabd8714fec9 Extracting [==================> ] 135.9MB/375MB b0e0ef7895f4 Downloading [> ] 375.7kB/37.01MB 2d429b9e73a6 Extracting [================================================> ] 28.31MB/29.13MB 1e017ebebdbd Extracting [=====> ] 3.932MB/37.19MB 85dde7dceb0a Extracting [===================================> ] 45.12MB/63.48MB 55f2b468da67 Downloading [==============> ] 76.23MB/257.9MB 2d429b9e73a6 Extracting [==================================================>] 29.13MB/29.13MB 82bfc142787e Downloading [===================================> ] 6.094MB/8.613MB b0e0ef7895f4 Downloading [===> ] 2.637MB/37.01MB 6ac0e4adf315 Extracting [=========================> ] 31.2MB/62.07MB eabd8714fec9 Extracting [==================> ] 139.3MB/375MB 1e017ebebdbd Extracting [========> ] 6.685MB/37.19MB 85dde7dceb0a Extracting [=====================================> ] 47.91MB/63.48MB 55f2b468da67 Downloading [=================> ] 90.29MB/257.9MB 6ac0e4adf315 Extracting [============================> ] 35.09MB/62.07MB 82bfc142787e Downloading [=============================================> ] 7.863MB/8.613MB eabd8714fec9 Extracting [==================> ] 140.9MB/375MB b0e0ef7895f4 Downloading [=====> ] 4.144MB/37.01MB 55f2b468da67 Downloading [=================> ] 90.83MB/257.9MB 1e017ebebdbd Extracting [===========> ] 8.651MB/37.19MB 82bfc142787e Verifying Checksum 82bfc142787e Download complete 85dde7dceb0a Extracting [======================================> ] 48.46MB/63.48MB c0c90eeb8aca Downloading [==================================================>] 1.105kB/1.105kB c0c90eeb8aca Verifying Checksum c0c90eeb8aca Download complete 2d429b9e73a6 Pull complete 46eab5b44a35 Extracting [==================================================>] 1.168kB/1.168kB 46eab5b44a35 Extracting [==================================================>] 1.168kB/1.168kB 5cfb27c10ea5 Downloading [==================================================>] 852B/852B 5cfb27c10ea5 Verifying Checksum 5cfb27c10ea5 Download complete 40a5eed61bb0 Downloading [==================================================>] 98B/98B 40a5eed61bb0 Verifying Checksum 40a5eed61bb0 Download complete e040ea11fa10 Downloading [==================================================>] 173B/173B e040ea11fa10 Verifying Checksum e040ea11fa10 Download complete 6ac0e4adf315 Extracting [=================================> ] 41.78MB/62.07MB 09d5a3f70313 Downloading [> ] 539.6kB/109.2MB b0e0ef7895f4 Downloading [===============> ] 11.68MB/37.01MB eabd8714fec9 Extracting [===================> ] 143.2MB/375MB 55f2b468da67 Downloading [===================> ] 101.1MB/257.9MB 85dde7dceb0a Extracting [=======================================> ] 50.14MB/63.48MB 1e017ebebdbd Extracting [==============> ] 10.62MB/37.19MB 85dde7dceb0a Extracting [========================================> ] 51.25MB/63.48MB 09d5a3f70313 Downloading [==> ] 5.946MB/109.2MB 55f2b468da67 Downloading [=====================> ] 109.8MB/257.9MB b0e0ef7895f4 Downloading [===========================> ] 20.72MB/37.01MB 1e017ebebdbd Extracting [================> ] 12.58MB/37.19MB 6ac0e4adf315 Extracting [===========================================> ] 54.03MB/62.07MB 46eab5b44a35 Pull complete eabd8714fec9 Extracting [===================> ] 145.4MB/375MB 85dde7dceb0a Extracting [=========================================> ] 52.92MB/63.48MB 09d5a3f70313 Downloading [=====> ] 12.43MB/109.2MB 55f2b468da67 Downloading [=======================> ] 120.6MB/257.9MB b0e0ef7895f4 Downloading [=======================================> ] 29.39MB/37.01MB 1e017ebebdbd Extracting [====================> ] 15.34MB/37.19MB 6ac0e4adf315 Extracting [================================================> ] 60.72MB/62.07MB b0e0ef7895f4 Verifying Checksum b0e0ef7895f4 Download complete eabd8714fec9 Extracting [===================> ] 148.2MB/375MB 356f5c2c843b Downloading [=========================================> ] 3.011kB/3.623kB 356f5c2c843b Downloading [==================================================>] 3.623kB/3.623kB 356f5c2c843b Verifying Checksum 356f5c2c843b Download complete 09d5a3f70313 Downloading [=========> ] 20.54MB/109.2MB 55f2b468da67 Downloading [=========================> ] 131.4MB/257.9MB 1e017ebebdbd Extracting [========================> ] 18.09MB/37.19MB 85dde7dceb0a Extracting [===========================================> ] 55.71MB/63.48MB 6ac0e4adf315 Extracting [==================================================>] 62.07MB/62.07MB eabd8714fec9 Extracting [===================> ] 149.8MB/375MB 09d5a3f70313 Downloading [==============> ] 31.36MB/109.2MB 55f2b468da67 Downloading [===========================> ] 142.2MB/257.9MB 1e017ebebdbd Extracting [=============================> ] 22.02MB/37.19MB eabd8714fec9 Extracting [====================> ] 152.6MB/375MB c4d302cc468d Extracting [> ] 65.54kB/4.534MB 55f2b468da67 Downloading [=============================> ] 151.9MB/257.9MB 1e017ebebdbd Extracting [================================> ] 23.99MB/37.19MB 09d5a3f70313 Downloading [===================> ] 43.25MB/109.2MB 85dde7dceb0a Extracting [==============================================> ] 59.6MB/63.48MB eabd8714fec9 Extracting [====================> ] 154.3MB/375MB 55f2b468da67 Downloading [===============================> ] 163.8MB/257.9MB 09d5a3f70313 Downloading [=========================> ] 55.69MB/109.2MB c4d302cc468d Extracting [===> ] 327.7kB/4.534MB 1e017ebebdbd Extracting [==================================> ] 25.95MB/37.19MB 85dde7dceb0a Extracting [=================================================> ] 62.39MB/63.48MB 09d5a3f70313 Downloading [===============================> ] 68.66MB/109.2MB 55f2b468da67 Downloading [===================================> ] 181.1MB/257.9MB eabd8714fec9 Extracting [====================> ] 156.5MB/375MB 85dde7dceb0a Extracting [==================================================>] 63.48MB/63.48MB 85dde7dceb0a Extracting [==================================================>] 63.48MB/63.48MB c4d302cc468d Extracting [==================================================>] 4.534MB/4.534MB 1e017ebebdbd Extracting [======================================> ] 28.31MB/37.19MB 09d5a3f70313 Downloading [====================================> ] 78.94MB/109.2MB 55f2b468da67 Downloading [======================================> ] 197.3MB/257.9MB eabd8714fec9 Extracting [=====================> ] 158.8MB/375MB 1e017ebebdbd Extracting [============================================> ] 33.03MB/37.19MB 09d5a3f70313 Downloading [=========================================> ] 89.75MB/109.2MB 55f2b468da67 Downloading [========================================> ] 207.6MB/257.9MB eabd8714fec9 Extracting [=====================> ] 162.1MB/375MB 1e017ebebdbd Extracting [===============================================> ] 35MB/37.19MB 09d5a3f70313 Downloading [===============================================> ] 104.3MB/109.2MB 55f2b468da67 Downloading [===========================================> ] 223.3MB/257.9MB 1e017ebebdbd Extracting [================================================> ] 35.78MB/37.19MB eabd8714fec9 Extracting [======================> ] 167.1MB/375MB 09d5a3f70313 Verifying Checksum 09d5a3f70313 Download complete 55f2b468da67 Downloading [===========================================> ] 224.4MB/257.9MB 1e017ebebdbd Extracting [================================================> ] 36.18MB/37.19MB eabd8714fec9 Extracting [======================> ] 172.1MB/375MB 1e017ebebdbd Extracting [==================================================>] 37.19MB/37.19MB 55f2b468da67 Downloading [==============================================> ] 239.5MB/257.9MB 6ac0e4adf315 Pull complete 55f2b468da67 Downloading [===============================================> ] 243.8MB/257.9MB eabd8714fec9 Extracting [========================> ] 182.7MB/375MB 85dde7dceb0a Pull complete 55f2b468da67 Downloading [=================================================> ] 255.7MB/257.9MB eabd8714fec9 Extracting [=========================> ] 191.6MB/375MB 55f2b468da67 Verifying Checksum 55f2b468da67 Download complete f3b09c502777 Extracting [> ] 557.1kB/56.52MB c4d302cc468d Pull complete 1e017ebebdbd Pull complete eabd8714fec9 Extracting [===========================> ] 205.6MB/375MB f3b09c502777 Extracting [====> ] 5.014MB/56.52MB 7009d5001b77 Extracting [==================================================>] 11.92kB/11.92kB 7009d5001b77 Extracting [==================================================>] 11.92kB/11.92kB eabd8714fec9 Extracting [============================> ] 216.7MB/375MB f3b09c502777 Extracting [=======> ] 8.356MB/56.52MB eabd8714fec9 Extracting [=============================> ] 221.2MB/375MB f3b09c502777 Extracting [==========> ] 11.7MB/56.52MB eabd8714fec9 Extracting [==============================> ] 226.2MB/375MB f3b09c502777 Extracting [==============> ] 16.71MB/56.52MB eabd8714fec9 Extracting [==============================> ] 230.1MB/375MB f3b09c502777 Extracting [=================> ] 20.05MB/56.52MB eabd8714fec9 Extracting [===============================> ] 234MB/375MB f3b09c502777 Extracting [==================> ] 20.61MB/56.52MB eabd8714fec9 Extracting [===============================> ] 237.3MB/375MB f3b09c502777 Extracting [========================> ] 27.3MB/56.52MB eabd8714fec9 Extracting [================================> ] 242.3MB/375MB f3b09c502777 Extracting [=====================================> ] 42.34MB/56.52MB eabd8714fec9 Extracting [=================================> ] 247.9MB/375MB f3b09c502777 Extracting [===============================================> ] 53.48MB/56.52MB eabd8714fec9 Extracting [=================================> ] 252.9MB/375MB f3b09c502777 Extracting [==================================================>] 56.52MB/56.52MB eabd8714fec9 Extracting [==================================> ] 256.8MB/375MB eabd8714fec9 Extracting [===================================> ] 264MB/375MB 01e0882c90d9 Extracting [=> ] 32.77kB/1.447MB eabd8714fec9 Extracting [===================================> ] 269.1MB/375MB 01e0882c90d9 Extracting [=======================================> ] 1.147MB/1.447MB 01e0882c90d9 Extracting [==================================================>] 1.447MB/1.447MB eabd8714fec9 Extracting [===================================> ] 269.6MB/375MB 7009d5001b77 Pull complete eabd8714fec9 Extracting [====================================> ] 270.7MB/375MB 55f2b468da67 Extracting [> ] 557.1kB/257.9MB 55f2b468da67 Extracting [=> ] 8.913MB/257.9MB eabd8714fec9 Extracting [====================================> ] 271.8MB/375MB 55f2b468da67 Extracting [===> ] 20.05MB/257.9MB eabd8714fec9 Extracting [====================================> ] 273.5MB/375MB eabd8714fec9 Extracting [====================================> ] 275.7MB/375MB 55f2b468da67 Extracting [====> ] 24.51MB/257.9MB eabd8714fec9 Extracting [=====================================> ] 281.3MB/375MB 538deb30e80c Extracting [==================================================>] 1.225kB/1.225kB 538deb30e80c Extracting [==================================================>] 1.225kB/1.225kB f3b09c502777 Pull complete 55f2b468da67 Extracting [=======> ] 39.55MB/257.9MB eabd8714fec9 Extracting [======================================> ] 285.2MB/375MB 55f2b468da67 Extracting [=========> ] 49.02MB/257.9MB eabd8714fec9 Extracting [======================================> ] 290.8MB/375MB 55f2b468da67 Extracting [============> ] 65.73MB/257.9MB eabd8714fec9 Extracting [=======================================> ] 294.1MB/375MB 55f2b468da67 Extracting [===============> ] 78.54MB/257.9MB 01e0882c90d9 Pull complete 55f2b468da67 Extracting [===============> ] 80.77MB/257.9MB eabd8714fec9 Extracting [=======================================> ] 295.2MB/375MB 55f2b468da67 Extracting [=================> ] 89.13MB/257.9MB eabd8714fec9 Extracting [=======================================> ] 296.9MB/375MB 55f2b468da67 Extracting [===================> ] 102.5MB/257.9MB eabd8714fec9 Extracting [========================================> ] 300.3MB/375MB 55f2b468da67 Extracting [=====================> ] 109.2MB/257.9MB eabd8714fec9 Extracting [========================================> ] 303MB/375MB 55f2b468da67 Extracting [======================> ] 114.8MB/257.9MB 408012a7b118 Extracting [==================================================>] 637B/637B 408012a7b118 Extracting [==================================================>] 637B/637B eabd8714fec9 Extracting [========================================> ] 305.8MB/375MB 55f2b468da67 Extracting [=======================> ] 119.2MB/257.9MB eabd8714fec9 Extracting [=========================================> ] 307.5MB/375MB 55f2b468da67 Extracting [========================> ] 125.3MB/257.9MB eabd8714fec9 Extracting [=========================================> ] 310.8MB/375MB 55f2b468da67 Extracting [=========================> ] 132MB/257.9MB eabd8714fec9 Extracting [=========================================> ] 312.5MB/375MB 55f2b468da67 Extracting [=========================> ] 133.1MB/257.9MB 531ee2cf3c0c Extracting [> ] 98.3kB/8.066MB 55f2b468da67 Extracting [==========================> ] 135.4MB/257.9MB eabd8714fec9 Extracting [=========================================> ] 313.6MB/375MB 538deb30e80c Pull complete eabd8714fec9 Extracting [=========================================> ] 314.2MB/375MB 55f2b468da67 Extracting [==========================> ] 137MB/257.9MB 408012a7b118 Pull complete 531ee2cf3c0c Extracting [=> ] 294.9kB/8.066MB eabd8714fec9 Extracting [==========================================> ] 315.9MB/375MB 531ee2cf3c0c Extracting [===================> ] 3.146MB/8.066MB 55f2b468da67 Extracting [===========================> ] 141.5MB/257.9MB eabd8714fec9 Extracting [==========================================> ] 318.6MB/375MB 55f2b468da67 Extracting [============================> ] 145.4MB/257.9MB 531ee2cf3c0c Extracting [============================> ] 4.522MB/8.066MB eabd8714fec9 Extracting [==========================================> ] 319.2MB/375MB 531ee2cf3c0c Extracting [=====================================> ] 5.997MB/8.066MB 55f2b468da67 Extracting [=============================> ] 149.8MB/257.9MB eabd8714fec9 Extracting [===========================================> ] 322.5MB/375MB 531ee2cf3c0c Extracting [=================================================> ] 8.061MB/8.066MB 531ee2cf3c0c Extracting [==================================================>] 8.066MB/8.066MB 55f2b468da67 Extracting [=============================> ] 153.7MB/257.9MB eabd8714fec9 Extracting [===========================================> ] 325.9MB/375MB 44986281b8b9 Extracting [==================================================>] 4.022kB/4.022kB 44986281b8b9 Extracting [==================================================>] 4.022kB/4.022kB 55f2b468da67 Extracting [==============================> ] 157.1MB/257.9MB eabd8714fec9 Extracting [===========================================> ] 328.7MB/375MB 55f2b468da67 Extracting [===============================> ] 162.7MB/257.9MB 55f2b468da67 Extracting [===============================> ] 163.2MB/257.9MB eabd8714fec9 Extracting [============================================> ] 330.9MB/375MB 55f2b468da67 Extracting [================================> ] 165.4MB/257.9MB eabd8714fec9 Extracting [============================================> ] 332MB/375MB 55f2b468da67 Extracting [================================> ] 169.3MB/257.9MB eabd8714fec9 Extracting [============================================> ] 334.8MB/375MB 55f2b468da67 Extracting [=================================> ] 170.5MB/257.9MB eabd8714fec9 Extracting [============================================> ] 337MB/375MB 55f2b468da67 Extracting [=================================> ] 171MB/257.9MB 44986281b8b9 Pull complete 531ee2cf3c0c Pull complete eabd8714fec9 Extracting [=============================================> ] 340.4MB/375MB 55f2b468da67 Extracting [=================================> ] 172.7MB/257.9MB 55f2b468da67 Extracting [=================================> ] 173.2MB/257.9MB grafana Pulled eabd8714fec9 Extracting [=============================================> ] 340.9MB/375MB ed54a7dee1d8 Extracting [=> ] 32.77kB/1.196MB 55f2b468da67 Extracting [=================================> ] 173.8MB/257.9MB eabd8714fec9 Extracting [=============================================> ] 341.5MB/375MB ed54a7dee1d8 Extracting [============> ] 294.9kB/1.196MB ed54a7dee1d8 Extracting [==================================================>] 1.196MB/1.196MB ed54a7dee1d8 Extracting [==================================================>] 1.196MB/1.196MB 55f2b468da67 Extracting [=================================> ] 174.4MB/257.9MB bf70c5107ab5 Extracting [==================================================>] 1.44kB/1.44kB bf70c5107ab5 Extracting [==================================================>] 1.44kB/1.44kB 55f2b468da67 Extracting [==================================> ] 175.5MB/257.9MB eabd8714fec9 Extracting [=============================================> ] 342MB/375MB ed54a7dee1d8 Pull complete 55f2b468da67 Extracting [==================================> ] 177.7MB/257.9MB 55f2b468da67 Extracting [===================================> ] 181.6MB/257.9MB eabd8714fec9 Extracting [=============================================> ] 342.6MB/375MB eabd8714fec9 Extracting [=============================================> ] 343.1MB/375MB 55f2b468da67 Extracting [====================================> ] 186.6MB/257.9MB 55f2b468da67 Extracting [====================================> ] 189.4MB/257.9MB eabd8714fec9 Extracting [==============================================> ] 345.4MB/375MB 55f2b468da67 Extracting [=====================================> ] 193.9MB/257.9MB eabd8714fec9 Extracting [==============================================> ] 345.9MB/375MB eabd8714fec9 Extracting [==============================================> ] 350.4MB/375MB 55f2b468da67 Extracting [======================================> ] 196.1MB/257.9MB eabd8714fec9 Extracting [===============================================> ] 353.7MB/375MB 55f2b468da67 Extracting [======================================> ] 196.6MB/257.9MB 12c5c803443f Extracting [==================================================>] 116B/116B 12c5c803443f Extracting [==================================================>] 116B/116B 55f2b468da67 Extracting [======================================> ] 198.3MB/257.9MB eabd8714fec9 Extracting [===============================================> ] 357.1MB/375MB 55f2b468da67 Extracting [======================================> ] 200MB/257.9MB 55f2b468da67 Extracting [======================================> ] 200.5MB/257.9MB bf70c5107ab5 Pull complete eabd8714fec9 Extracting [===============================================> ] 357.6MB/375MB 55f2b468da67 Extracting [=======================================> ] 201.7MB/257.9MB eabd8714fec9 Extracting [================================================> ] 361.5MB/375MB 55f2b468da67 Extracting [=======================================> ] 202.2MB/257.9MB eabd8714fec9 Extracting [================================================> ] 367.1MB/375MB 55f2b468da67 Extracting [=======================================> ] 203.9MB/257.9MB eabd8714fec9 Extracting [=================================================> ] 371.6MB/375MB 55f2b468da67 Extracting [========================================> ] 206.7MB/257.9MB eabd8714fec9 Extracting [==================================================>] 375MB/375MB 55f2b468da67 Extracting [========================================> ] 207.8MB/257.9MB 1ccde423731d Extracting [==========================> ] 32.77kB/61.44kB 1ccde423731d Extracting [==================================================>] 61.44kB/61.44kB 55f2b468da67 Extracting [========================================> ] 210MB/257.9MB 12c5c803443f Pull complete e27c75a98748 Extracting [==================================================>] 3.144kB/3.144kB e27c75a98748 Extracting [==================================================>] 3.144kB/3.144kB 55f2b468da67 Extracting [========================================> ] 210.6MB/257.9MB 55f2b468da67 Extracting [=========================================> ] 211.7MB/257.9MB 1ccde423731d Pull complete 7221d93db8a9 Extracting [==================================================>] 100B/100B 7221d93db8a9 Extracting [==================================================>] 100B/100B eabd8714fec9 Pull complete e27c75a98748 Pull complete 45fd2fec8a19 Extracting [==================================================>] 1.103kB/1.103kB 45fd2fec8a19 Extracting [==================================================>] 1.103kB/1.103kB 55f2b468da67 Extracting [=========================================> ] 213.9MB/257.9MB e73cb4a42719 Extracting [> ] 557.1kB/109.1MB 55f2b468da67 Extracting [==========================================> ] 217.8MB/257.9MB 45fd2fec8a19 Pull complete 7221d93db8a9 Pull complete 7df673c7455d Extracting [==================================================>] 694B/694B 7df673c7455d Extracting [==================================================>] 694B/694B 8f10199ed94b Extracting [> ] 98.3kB/8.768MB e73cb4a42719 Extracting [=> ] 2.785MB/109.1MB 55f2b468da67 Extracting [==========================================> ] 220MB/257.9MB 8f10199ed94b Extracting [==========================> ] 4.719MB/8.768MB e73cb4a42719 Extracting [===> ] 6.685MB/109.1MB 7df673c7455d Pull complete 55f2b468da67 Extracting [===========================================> ] 223.4MB/257.9MB prometheus Pulled 8f10199ed94b Extracting [================================================> ] 8.454MB/8.768MB 8f10199ed94b Extracting [==================================================>] 8.768MB/8.768MB e73cb4a42719 Extracting [====> ] 9.47MB/109.1MB 55f2b468da67 Extracting [===========================================> ] 225.6MB/257.9MB 8f10199ed94b Pull complete f963a77d2726 Extracting [==================================================>] 21.44kB/21.44kB f963a77d2726 Extracting [==================================================>] 21.44kB/21.44kB 55f2b468da67 Extracting [============================================> ] 227.3MB/257.9MB e73cb4a42719 Extracting [=====> ] 12.81MB/109.1MB f963a77d2726 Pull complete 55f2b468da67 Extracting [============================================> ] 228.4MB/257.9MB e73cb4a42719 Extracting [=======> ] 16.15MB/109.1MB f3a82e9f1761 Extracting [> ] 458.8kB/44.41MB e73cb4a42719 Extracting [========> ] 19.5MB/109.1MB 55f2b468da67 Extracting [============================================> ] 230.6MB/257.9MB f3a82e9f1761 Extracting [===========> ] 10.09MB/44.41MB e73cb4a42719 Extracting [==========> ] 23.95MB/109.1MB 55f2b468da67 Extracting [=============================================> ] 232.8MB/257.9MB f3a82e9f1761 Extracting [======================> ] 19.73MB/44.41MB e73cb4a42719 Extracting [============> ] 27.3MB/109.1MB f3a82e9f1761 Extracting [==================================> ] 30.74MB/44.41MB 55f2b468da67 Extracting [=============================================> ] 236.7MB/257.9MB e73cb4a42719 Extracting [===============> ] 32.87MB/109.1MB f3a82e9f1761 Extracting [===============================================> ] 42.21MB/44.41MB 55f2b468da67 Extracting [==============================================> ] 242.3MB/257.9MB f3a82e9f1761 Extracting [==================================================>] 44.41MB/44.41MB e73cb4a42719 Extracting [=================> ] 38.44MB/109.1MB e73cb4a42719 Extracting [====================> ] 45.68MB/109.1MB 55f2b468da67 Extracting [===============================================> ] 244.5MB/257.9MB e73cb4a42719 Extracting [=======================> ] 51.25MB/109.1MB 55f2b468da67 Extracting [=================================================> ] 253.5MB/257.9MB 55f2b468da67 Extracting [=================================================> ] 256.2MB/257.9MB e73cb4a42719 Extracting [========================> ] 53.48MB/109.1MB e73cb4a42719 Extracting [=========================> ] 56.26MB/109.1MB 55f2b468da67 Extracting [==================================================>] 257.9MB/257.9MB 55f2b468da67 Extracting [==================================================>] 257.9MB/257.9MB e73cb4a42719 Extracting [===========================> ] 60.16MB/109.1MB e73cb4a42719 Extracting [==============================> ] 66.85MB/109.1MB e73cb4a42719 Extracting [================================> ] 70.75MB/109.1MB e73cb4a42719 Extracting [===================================> ] 76.87MB/109.1MB e73cb4a42719 Extracting [=====================================> ] 80.77MB/109.1MB e73cb4a42719 Extracting [========================================> ] 87.46MB/109.1MB e73cb4a42719 Extracting [==========================================> ] 92.47MB/109.1MB e73cb4a42719 Extracting [===========================================> ] 95.81MB/109.1MB e73cb4a42719 Extracting [============================================> ] 98.04MB/109.1MB e73cb4a42719 Extracting [==============================================> ] 101.4MB/109.1MB e73cb4a42719 Extracting [===============================================> ] 104.2MB/109.1MB e73cb4a42719 Extracting [================================================> ] 106.4MB/109.1MB e73cb4a42719 Extracting [=================================================> ] 107.5MB/109.1MB e73cb4a42719 Extracting [=================================================> ] 108.6MB/109.1MB e73cb4a42719 Extracting [==================================================>] 109.1MB/109.1MB f3a82e9f1761 Pull complete 55f2b468da67 Pull complete 79161a3f5362 Extracting [==================================================>] 4.656kB/4.656kB 79161a3f5362 Extracting [==================================================>] 4.656kB/4.656kB e73cb4a42719 Pull complete 82bfc142787e Extracting [> ] 98.3kB/8.613MB a83b68436f09 Extracting [==================================================>] 9.919kB/9.919kB a83b68436f09 Extracting [==================================================>] 9.919kB/9.919kB 79161a3f5362 Pull complete 9c266ba63f51 Extracting [==================================================>] 1.105kB/1.105kB 9c266ba63f51 Extracting [==================================================>] 1.105kB/1.105kB 82bfc142787e Extracting [=====================> ] 3.736MB/8.613MB 82bfc142787e Extracting [==================================================>] 8.613MB/8.613MB a83b68436f09 Pull complete 787d6bee9571 Extracting [==================================================>] 127B/127B 787d6bee9571 Extracting [==================================================>] 127B/127B 9c266ba63f51 Pull complete 82bfc142787e Pull complete 2e8a7df9c2ee Extracting [==================================================>] 851B/851B 2e8a7df9c2ee Extracting [==================================================>] 851B/851B 46baca71a4ef Extracting [==================================================>] 18.11kB/18.11kB 46baca71a4ef Extracting [==================================================>] 18.11kB/18.11kB 787d6bee9571 Pull complete 13ff0988aaea Extracting [==================================================>] 167B/167B 13ff0988aaea Extracting [==================================================>] 167B/167B 2e8a7df9c2ee Pull complete 46baca71a4ef Pull complete 10f05dd8b1db Extracting [==================================================>] 98B/98B 10f05dd8b1db Extracting [==================================================>] 98B/98B 13ff0988aaea Pull complete b0e0ef7895f4 Extracting [> ] 393.2kB/37.01MB 4b82842ab819 Extracting [==================================================>] 5.415kB/5.415kB 4b82842ab819 Extracting [==================================================>] 5.415kB/5.415kB 10f05dd8b1db Pull complete 41dac8b43ba6 Extracting [==================================================>] 171B/171B 41dac8b43ba6 Extracting [==================================================>] 171B/171B b0e0ef7895f4 Extracting [=====================> ] 15.73MB/37.01MB 4b82842ab819 Pull complete 7e568a0dc8fb Extracting [==================================================>] 184B/184B 7e568a0dc8fb Extracting [==================================================>] 184B/184B 41dac8b43ba6 Pull complete 71a9f6a9ab4d Extracting [=======> ] 32.77kB/230.6kB b0e0ef7895f4 Extracting [=================================> ] 25.17MB/37.01MB 71a9f6a9ab4d Extracting [==================================================>] 230.6kB/230.6kB 7e568a0dc8fb Pull complete b0e0ef7895f4 Extracting [==================================================>] 37.01MB/37.01MB 71a9f6a9ab4d Pull complete postgres Pulled b0e0ef7895f4 Pull complete c0c90eeb8aca Extracting [==================================================>] 1.105kB/1.105kB c0c90eeb8aca Extracting [==================================================>] 1.105kB/1.105kB da3ed5db7103 Extracting [> ] 557.1kB/127.4MB c0c90eeb8aca Pull complete 5cfb27c10ea5 Extracting [==================================================>] 852B/852B 5cfb27c10ea5 Extracting [==================================================>] 852B/852B da3ed5db7103 Extracting [======> ] 16.15MB/127.4MB 5cfb27c10ea5 Pull complete 40a5eed61bb0 Extracting [==================================================>] 98B/98B 40a5eed61bb0 Extracting [==================================================>] 98B/98B da3ed5db7103 Extracting [============> ] 31.75MB/127.4MB da3ed5db7103 Extracting [==================> ] 46.79MB/127.4MB 40a5eed61bb0 Pull complete e040ea11fa10 Extracting [==================================================>] 173B/173B e040ea11fa10 Extracting [==================================================>] 173B/173B da3ed5db7103 Extracting [=========================> ] 64.62MB/127.4MB e040ea11fa10 Pull complete da3ed5db7103 Extracting [==============================> ] 77.99MB/127.4MB 09d5a3f70313 Extracting [> ] 557.1kB/109.2MB da3ed5db7103 Extracting [====================================> ] 93.59MB/127.4MB 09d5a3f70313 Extracting [======> ] 14.48MB/109.2MB da3ed5db7103 Extracting [===========================================> ] 110.9MB/127.4MB 09d5a3f70313 Extracting [============> ] 27.85MB/109.2MB da3ed5db7103 Extracting [===============================================> ] 120.3MB/127.4MB 09d5a3f70313 Extracting [===================> ] 42.34MB/109.2MB da3ed5db7103 Extracting [=================================================> ] 125.9MB/127.4MB da3ed5db7103 Extracting [==================================================>] 127.4MB/127.4MB 09d5a3f70313 Extracting [===========================> ] 59.6MB/109.2MB da3ed5db7103 Pull complete c955f6e31a04 Extracting [==================================================>] 3.446kB/3.446kB c955f6e31a04 Extracting [==================================================>] 3.446kB/3.446kB 09d5a3f70313 Extracting [=================================> ] 73.53MB/109.2MB 09d5a3f70313 Extracting [========================================> ] 88.57MB/109.2MB c955f6e31a04 Pull complete zookeeper Pulled 09d5a3f70313 Extracting [===============================================> ] 104.2MB/109.2MB 09d5a3f70313 Extracting [==================================================>] 109.2MB/109.2MB 09d5a3f70313 Extracting [==================================================>] 109.2MB/109.2MB 09d5a3f70313 Pull complete 356f5c2c843b Extracting [==================================================>] 3.623kB/3.623kB 356f5c2c843b Extracting [==================================================>] 3.623kB/3.623kB 356f5c2c843b Pull complete kafka Pulled Network compose_default Creating Network compose_default Created Container postgres Creating Container zookeeper Creating Container prometheus Creating Container zookeeper Created Container kafka Creating Container prometheus Created Container grafana Creating Container postgres Created Container policy-db-migrator Creating Container policy-db-migrator Created Container policy-api Creating Container grafana Created Container kafka Created Container policy-api Created Container policy-pap Creating Container policy-pap Created Container policy-xacml-pdp Creating Container policy-xacml-pdp Created Container postgres Starting Container zookeeper Starting Container prometheus Starting Container postgres Started Container policy-db-migrator Starting Container policy-db-migrator Started Container policy-api Starting Container zookeeper Started Container kafka Starting Container prometheus Started Container grafana Starting Container grafana Started Container policy-api Started Container kafka Started Container policy-pap Starting Container policy-pap Started Container policy-xacml-pdp Starting Container policy-xacml-pdp Started Prometheus server: http://localhost:30259 Grafana server: http://localhost:30269 Waiting 1 minute for xacml-pdp to start... Checking if REST port 30004 is open on localhost ... IMAGE NAMES STATUS nexus3.onap.org:10001/onap/policy-xacml-pdp:4.2.1-SNAPSHOT policy-xacml-pdp Up About a minute nexus3.onap.org:10001/onap/policy-pap:4.2.1-SNAPSHOT policy-pap Up About a minute nexus3.onap.org:10001/onap/policy-api:4.2.1-SNAPSHOT policy-api Up About a minute nexus3.onap.org:10001/confluentinc/cp-kafka:7.4.9 kafka Up About a minute nexus3.onap.org:10001/grafana/grafana:latest grafana Up About a minute nexus3.onap.org:10001/confluentinc/cp-zookeeper:latest zookeeper Up About a minute nexus3.onap.org:10001/prom/prometheus:latest prometheus Up About a minute nexus3.onap.org:10001/library/postgres:16.4 postgres Up About a minute Cloning into '/w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp/csit/resources/tests/models'... Building robot framework docker image sha256:d703360be706d5312ed5bbee3337ecf8baeddc9ec0107f65af66dcda89023422 top - 15:22:13 up 5 min, 0 users, load average: 2.23, 1.82, 0.80 Tasks: 230 total, 1 running, 150 sleeping, 0 stopped, 0 zombie %Cpu(s): 12.8 us, 2.9 sy, 0.0 ni, 77.2 id, 7.0 wa, 0.0 hi, 0.1 si, 0.1 st total used free shared buff/cache available Mem: 31G 2.5G 21G 27M 7.1G 28G Swap: 1.0G 0B 1.0G IMAGE NAMES STATUS nexus3.onap.org:10001/onap/policy-xacml-pdp:4.2.1-SNAPSHOT policy-xacml-pdp Up About a minute nexus3.onap.org:10001/onap/policy-pap:4.2.1-SNAPSHOT policy-pap Up About a minute nexus3.onap.org:10001/onap/policy-api:4.2.1-SNAPSHOT policy-api Up About a minute nexus3.onap.org:10001/confluentinc/cp-kafka:7.4.9 kafka Up About a minute nexus3.onap.org:10001/grafana/grafana:latest grafana Up About a minute nexus3.onap.org:10001/confluentinc/cp-zookeeper:latest zookeeper Up About a minute nexus3.onap.org:10001/prom/prometheus:latest prometheus Up About a minute nexus3.onap.org:10001/library/postgres:16.4 postgres Up About a minute CONTAINER ID NAME CPU % MEM USAGE / LIMIT MEM % NET I/O BLOCK I/O PIDS 5cfb8a39c6ba policy-xacml-pdp 0.73% 173.3MiB / 31.41GiB 0.54% 44.8kB / 55.4kB 0B / 4.1kB 51 15287384e37a policy-pap 0.91% 545.7MiB / 31.41GiB 1.70% 2.13MB / 1.06MB 0B / 139MB 68 7382b4734a19 policy-api 0.09% 439.4MiB / 31.41GiB 1.37% 1.15MB / 986kB 0B / 8.19kB 57 97f6c82b4ff6 kafka 3.05% 392.6MiB / 31.41GiB 1.22% 187kB / 176kB 8.19kB / 635kB 83 e16f2b999244 grafana 0.20% 106.3MiB / 31.41GiB 0.33% 19MB / 213kB 0B / 30.6MB 20 b73308842e34 zookeeper 0.09% 87.33MiB / 31.41GiB 0.27% 59.2kB / 52.1kB 0B / 414kB 63 896ceee32cb8 prometheus 0.00% 20.6MiB / 31.41GiB 0.06% 63kB / 3.62kB 225kB / 0B 14 ef97e30b608b postgres 0.00% 86.18MiB / 31.41GiB 0.27% 2.56MB / 3.75MB 4.1kB / 158MB 26 Container policy-csit Creating Container policy-csit Created Attaching to policy-csit policy-csit | Invoking the robot tests from: xacml-pdp-test.robot xacml-pdp-slas.robot policy-csit | Run Robot test policy-csit | ROBOT_VARIABLES=-v DATA:/opt/robotworkspace/models/models-examples/src/main/resources/policies policy-csit | -v NODETEMPLATES:/opt/robotworkspace/models/models-examples/src/main/resources/nodetemplates policy-csit | -v POLICY_API_IP:policy-api:6969 policy-csit | -v POLICY_RUNTIME_ACM_IP:policy-clamp-runtime-acm:6969 policy-csit | -v POLICY_PARTICIPANT_SIM_IP:policy-clamp-ac-sim-ppnt:6969 policy-csit | -v POLICY_PAP_IP:policy-pap:6969 policy-csit | -v APEX_IP:policy-apex-pdp:6969 policy-csit | -v APEX_EVENTS_IP:policy-apex-pdp:23324 policy-csit | -v KAFKA_IP:kafka:9092 policy-csit | -v PROMETHEUS_IP:prometheus:9090 policy-csit | -v POLICY_PDPX_IP:policy-xacml-pdp:6969 policy-csit | -v POLICY_OPA_IP:policy-opa-pdp:8282 policy-csit | -v POLICY_DROOLS_IP:policy-drools-pdp:9696 policy-csit | -v DROOLS_IP:policy-drools-apps:6969 policy-csit | -v DROOLS_IP_2:policy-drools-apps:9696 policy-csit | -v TEMP_FOLDER:/tmp/distribution policy-csit | -v DISTRIBUTION_IP:policy-distribution:6969 policy-csit | -v TEST_ENV:docker policy-csit | -v JAEGER_IP:jaeger:16686 policy-csit | Starting Robot test suites ... policy-csit | ============================================================================== policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas policy-csit | ============================================================================== policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas.Xacml-Pdp-Test policy-csit | ============================================================================== policy-csit | Healthcheck :: Verify policy xacml-pdp health check | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | Metrics :: Verify policy-xacml-pdp is exporting prometheus metrics | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | MakeTopics :: Creates the Policy topics | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | ExecuteXacmlPolicy | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas.Xacml-Pdp-Test | PASS | policy-csit | 4 tests, 4 passed, 0 failed policy-csit | ============================================================================== policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas.Xacml-Pdp-Slas policy-csit | ============================================================================== policy-csit | WaitForPrometheusServer :: Sleep time to wait for Prometheus serve... | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | ValidatePolicyDecisionsTotalCounter :: Validate policy decision co... | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas.Xacml-Pdp-Slas | PASS | policy-csit | 2 tests, 2 passed, 0 failed policy-csit | ============================================================================== policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas | PASS | policy-csit | 6 tests, 6 passed, 0 failed policy-csit | ============================================================================== policy-csit | Output: /tmp/results/output.xml policy-csit | Log: /tmp/results/log.html policy-csit | Report: /tmp/results/report.html policy-csit | RESULT: 0 policy-csit exited with code 0 IMAGE NAMES STATUS nexus3.onap.org:10001/onap/policy-xacml-pdp:4.2.1-SNAPSHOT policy-xacml-pdp Up 3 minutes nexus3.onap.org:10001/onap/policy-pap:4.2.1-SNAPSHOT policy-pap Up 3 minutes nexus3.onap.org:10001/onap/policy-api:4.2.1-SNAPSHOT policy-api Up 3 minutes nexus3.onap.org:10001/confluentinc/cp-kafka:7.4.9 kafka Up 3 minutes nexus3.onap.org:10001/grafana/grafana:latest grafana Up 3 minutes nexus3.onap.org:10001/confluentinc/cp-zookeeper:latest zookeeper Up 3 minutes nexus3.onap.org:10001/prom/prometheus:latest prometheus Up 3 minutes nexus3.onap.org:10001/library/postgres:16.4 postgres Up 3 minutes Shut down started! Collecting logs from docker compose containers... grafana | logger=settings t=2025-06-18T15:20:26.38310824Z level=info msg="Starting Grafana" version=12.0.2 commit=5bda17e7c1cb313eb96266f2fdda73a6b35c3977 branch=HEAD compiled=2025-06-18T15:20:26Z grafana | logger=settings t=2025-06-18T15:20:26.383874138Z level=info msg="Config loaded from" file=/usr/share/grafana/conf/defaults.ini grafana | logger=settings t=2025-06-18T15:20:26.383898568Z level=info msg="Config loaded from" file=/etc/grafana/grafana.ini grafana | logger=settings t=2025-06-18T15:20:26.383902638Z level=info msg="Config overridden from command line" arg="default.paths.data=/var/lib/grafana" grafana | logger=settings t=2025-06-18T15:20:26.383907418Z level=info msg="Config overridden from command line" arg="default.paths.logs=/var/log/grafana" grafana | logger=settings t=2025-06-18T15:20:26.383910668Z level=info msg="Config overridden from command line" arg="default.paths.plugins=/var/lib/grafana/plugins" grafana | logger=settings t=2025-06-18T15:20:26.383942508Z level=info msg="Config overridden from command line" arg="default.paths.provisioning=/etc/grafana/provisioning" grafana | logger=settings t=2025-06-18T15:20:26.383946708Z level=info msg="Config overridden from command line" arg="default.log.mode=console" grafana | logger=settings t=2025-06-18T15:20:26.383974589Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_DATA=/var/lib/grafana" grafana | logger=settings t=2025-06-18T15:20:26.383981549Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_LOGS=/var/log/grafana" grafana | logger=settings t=2025-06-18T15:20:26.383985789Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_PLUGINS=/var/lib/grafana/plugins" grafana | logger=settings t=2025-06-18T15:20:26.384023259Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_PROVISIONING=/etc/grafana/provisioning" grafana | logger=settings t=2025-06-18T15:20:26.384042169Z level=info msg=Target target=[all] grafana | logger=settings t=2025-06-18T15:20:26.38410617Z level=info msg="Path Home" path=/usr/share/grafana grafana | logger=settings t=2025-06-18T15:20:26.38411133Z level=info msg="Path Data" path=/var/lib/grafana grafana | logger=settings t=2025-06-18T15:20:26.38413622Z level=info msg="Path Logs" path=/var/log/grafana grafana | logger=settings t=2025-06-18T15:20:26.38414332Z level=info msg="Path Plugins" path=/var/lib/grafana/plugins grafana | logger=settings t=2025-06-18T15:20:26.38414772Z level=info msg="Path Provisioning" path=/etc/grafana/provisioning grafana | logger=settings t=2025-06-18T15:20:26.38415136Z level=info msg="App mode production" grafana | logger=featuremgmt t=2025-06-18T15:20:26.384560354Z level=info msg=FeatureToggles prometheusUsesCombobox=true grafanaconThemes=true alertingRuleRecoverDeleted=true prometheusAzureOverrideAudience=true panelMonitoring=true cloudWatchNewLabelParsing=true cloudWatchCrossAccountQuerying=true kubernetesClientDashboardsFolders=true externalCorePlugins=true awsAsyncQueryCaching=true recordedQueriesMulti=true alertingInsights=true ssoSettingsApi=true alertingQueryAndExpressionsStepMode=true dashgpt=true addFieldFromCalculationStatFunctions=true formatString=true azureMonitorPrometheusExemplars=true groupToNestedTableTransformation=true reportingUseRawTimeRange=true tlsMemcached=true azureMonitorEnableUserAuth=true logsExploreTableVisualisation=true cloudWatchRoundUpEndTime=true lokiStructuredMetadata=true preinstallAutoUpdate=true pinNavItems=true dashboardScene=true newDashboardSharingComponent=true logsPanelControls=true kubernetesPlaylists=true recoveryThreshold=true alertingUIOptimizeReducer=true alertingApiServer=true alertingRulePermanentlyDelete=true angularDeprecationUI=true newFiltersUI=true lokiLabelNamesQueryApi=true correlations=true influxdbBackendMigration=true dashboardSceneSolo=true onPremToCloudMigrations=true alertingNotificationsStepMode=true annotationPermissionUpdate=true lokiQuerySplitting=true nestedFolders=true useSessionStorageForRedirection=true logRowsPopoverMenu=true unifiedStorageSearchPermissionFiltering=true transformationsRedesign=true alertingSimplifiedRouting=true newPDFRendering=true dashboardSceneForViewers=true ssoSettingsSAML=true logsInfiniteScrolling=true logsContextDatasourceUi=true lokiQueryHints=true unifiedRequestLog=true dataplaneFrontendFallback=true pluginsDetailsRightPanel=true failWrongDSUID=true promQLScope=true alertRuleRestore=true publicDashboardsScene=true alertingRuleVersionHistoryRestore=true grafana | logger=sqlstore t=2025-06-18T15:20:26.384665805Z level=info msg="Connecting to DB" dbtype=sqlite3 grafana | logger=sqlstore t=2025-06-18T15:20:26.384723376Z level=info msg="Creating SQLite database file" path=/var/lib/grafana/grafana.db grafana | logger=migrator t=2025-06-18T15:20:26.386429703Z level=info msg="Locking database" grafana | logger=migrator t=2025-06-18T15:20:26.386440953Z level=info msg="Starting DB migrations" grafana | logger=migrator t=2025-06-18T15:20:26.387164119Z level=info msg="Executing migration" id="create migration_log table" grafana | logger=migrator t=2025-06-18T15:20:26.388056538Z level=info msg="Migration successfully executed" id="create migration_log table" duration=890.849µs grafana | logger=migrator t=2025-06-18T15:20:26.400720832Z level=info msg="Executing migration" id="create user table" grafana | logger=migrator t=2025-06-18T15:20:26.401751132Z level=info msg="Migration successfully executed" id="create user table" duration=1.030699ms grafana | logger=migrator t=2025-06-18T15:20:26.411192654Z level=info msg="Executing migration" id="add unique index user.login" grafana | logger=migrator t=2025-06-18T15:20:26.412003162Z level=info msg="Migration successfully executed" id="add unique index user.login" duration=808.858µs grafana | logger=migrator t=2025-06-18T15:20:26.421730416Z level=info msg="Executing migration" id="add unique index user.email" grafana | logger=migrator t=2025-06-18T15:20:26.42317108Z level=info msg="Migration successfully executed" id="add unique index user.email" duration=1.438584ms grafana | logger=migrator t=2025-06-18T15:20:26.45286098Z level=info msg="Executing migration" id="drop index UQE_user_login - v1" grafana | logger=migrator t=2025-06-18T15:20:26.454207053Z level=info msg="Migration successfully executed" id="drop index UQE_user_login - v1" duration=1.342293ms grafana | logger=migrator t=2025-06-18T15:20:26.472237848Z level=info msg="Executing migration" id="drop index UQE_user_email - v1" grafana | logger=migrator t=2025-06-18T15:20:26.473945835Z level=info msg="Migration successfully executed" id="drop index UQE_user_email - v1" duration=1.708827ms grafana | logger=migrator t=2025-06-18T15:20:26.488059593Z level=info msg="Executing migration" id="Rename table user to user_v1 - v1" grafana | logger=migrator t=2025-06-18T15:20:26.493179183Z level=info msg="Migration successfully executed" id="Rename table user to user_v1 - v1" duration=5.11712ms grafana | logger=migrator t=2025-06-18T15:20:26.521555149Z level=info msg="Executing migration" id="create user table v2" grafana | logger=migrator t=2025-06-18T15:20:26.523136825Z level=info msg="Migration successfully executed" id="create user table v2" duration=1.582166ms grafana | logger=migrator t=2025-06-18T15:20:26.54827956Z level=info msg="Executing migration" id="create index UQE_user_login - v2" grafana | logger=migrator t=2025-06-18T15:20:26.549640133Z level=info msg="Migration successfully executed" id="create index UQE_user_login - v2" duration=1.355373ms grafana | logger=migrator t=2025-06-18T15:20:26.556697782Z level=info msg="Executing migration" id="create index UQE_user_email - v2" grafana | logger=migrator t=2025-06-18T15:20:26.557956184Z level=info msg="Migration successfully executed" id="create index UQE_user_email - v2" duration=1.257982ms grafana | logger=migrator t=2025-06-18T15:20:26.567338515Z level=info msg="Executing migration" id="copy data_source v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:26.56769338Z level=info msg="Migration successfully executed" id="copy data_source v1 to v2" duration=354.115µs grafana | logger=migrator t=2025-06-18T15:20:26.576168552Z level=info msg="Executing migration" id="Drop old table user_v1" grafana | logger=migrator t=2025-06-18T15:20:26.577090681Z level=info msg="Migration successfully executed" id="Drop old table user_v1" duration=922.449µs grafana | logger=migrator t=2025-06-18T15:20:26.586166499Z level=info msg="Executing migration" id="Add column help_flags1 to user table" grafana | logger=migrator t=2025-06-18T15:20:26.587829796Z level=info msg="Migration successfully executed" id="Add column help_flags1 to user table" duration=1.660517ms grafana | logger=migrator t=2025-06-18T15:20:26.59653838Z level=info msg="Executing migration" id="Update user table charset" grafana | logger=migrator t=2025-06-18T15:20:26.596592031Z level=info msg="Migration successfully executed" id="Update user table charset" duration=54.211µs grafana | logger=migrator t=2025-06-18T15:20:26.607827631Z level=info msg="Executing migration" id="Add last_seen_at column to user" grafana | logger=migrator t=2025-06-18T15:20:26.609566197Z level=info msg="Migration successfully executed" id="Add last_seen_at column to user" duration=1.738456ms grafana | logger=migrator t=2025-06-18T15:20:26.639464059Z level=info msg="Executing migration" id="Add missing user data" grafana | logger=migrator t=2025-06-18T15:20:26.639937683Z level=info msg="Migration successfully executed" id="Add missing user data" duration=475.585µs grafana | logger=migrator t=2025-06-18T15:20:26.68258929Z level=info msg="Executing migration" id="Add is_disabled column to user" grafana | logger=migrator t=2025-06-18T15:20:26.684308556Z level=info msg="Migration successfully executed" id="Add is_disabled column to user" duration=1.721726ms grafana | logger=migrator t=2025-06-18T15:20:26.705407502Z level=info msg="Executing migration" id="Add index user.login/user.email" grafana | logger=migrator t=2025-06-18T15:20:26.706884456Z level=info msg="Migration successfully executed" id="Add index user.login/user.email" duration=1.476914ms grafana | logger=migrator t=2025-06-18T15:20:26.713971415Z level=info msg="Executing migration" id="Add is_service_account column to user" grafana | logger=migrator t=2025-06-18T15:20:26.716372159Z level=info msg="Migration successfully executed" id="Add is_service_account column to user" duration=2.402484ms grafana | logger=migrator t=2025-06-18T15:20:26.720179145Z level=info msg="Executing migration" id="Update is_service_account column to nullable" grafana | logger=migrator t=2025-06-18T15:20:26.728493687Z level=info msg="Migration successfully executed" id="Update is_service_account column to nullable" duration=8.313402ms grafana | logger=migrator t=2025-06-18T15:20:26.733314364Z level=info msg="Executing migration" id="Add uid column to user" grafana | logger=migrator t=2025-06-18T15:20:26.735692836Z level=info msg="Migration successfully executed" id="Add uid column to user" duration=2.377272ms grafana | logger=migrator t=2025-06-18T15:20:26.740030739Z level=info msg="Executing migration" id="Update uid column values for users" grafana | logger=migrator t=2025-06-18T15:20:26.740329102Z level=info msg="Migration successfully executed" id="Update uid column values for users" duration=297.123µs grafana | logger=migrator t=2025-06-18T15:20:26.744839006Z level=info msg="Executing migration" id="Add unique index user_uid" grafana | logger=migrator t=2025-06-18T15:20:26.745664574Z level=info msg="Migration successfully executed" id="Add unique index user_uid" duration=825.308µs grafana | logger=migrator t=2025-06-18T15:20:26.749300789Z level=info msg="Executing migration" id="Add is_provisioned column to user" grafana | logger=migrator t=2025-06-18T15:20:26.750562412Z level=info msg="Migration successfully executed" id="Add is_provisioned column to user" duration=1.261073ms grafana | logger=migrator t=2025-06-18T15:20:26.755617311Z level=info msg="Executing migration" id="update login field with orgid to allow for multiple service accounts with same name across orgs" grafana | logger=migrator t=2025-06-18T15:20:26.756117166Z level=info msg="Migration successfully executed" id="update login field with orgid to allow for multiple service accounts with same name across orgs" duration=506.435µs grafana | logger=migrator t=2025-06-18T15:20:26.762591509Z level=info msg="Executing migration" id="update service accounts login field orgid to appear only once" grafana | logger=migrator t=2025-06-18T15:20:26.763547668Z level=info msg="Migration successfully executed" id="update service accounts login field orgid to appear only once" duration=955.589µs grafana | logger=migrator t=2025-06-18T15:20:26.768218444Z level=info msg="Executing migration" id="update login and email fields to lowercase" grafana | logger=migrator t=2025-06-18T15:20:26.768952301Z level=info msg="Migration successfully executed" id="update login and email fields to lowercase" duration=732.977µs grafana | logger=migrator t=2025-06-18T15:20:26.772782378Z level=info msg="Executing migration" id="update login and email fields to lowercase2" grafana | logger=migrator t=2025-06-18T15:20:26.773264373Z level=info msg="Migration successfully executed" id="update login and email fields to lowercase2" duration=481.155µs grafana | logger=migrator t=2025-06-18T15:20:26.776564255Z level=info msg="Executing migration" id="create temp user table v1-7" grafana | logger=migrator t=2025-06-18T15:20:26.777562114Z level=info msg="Migration successfully executed" id="create temp user table v1-7" duration=996.799µs grafana | logger=migrator t=2025-06-18T15:20:26.783372982Z level=info msg="Executing migration" id="create index IDX_temp_user_email - v1-7" grafana | logger=migrator t=2025-06-18T15:20:26.78420774Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_email - v1-7" duration=834.118µs grafana | logger=migrator t=2025-06-18T15:20:26.794315458Z level=info msg="Executing migration" id="create index IDX_temp_user_org_id - v1-7" grafana | logger=migrator t=2025-06-18T15:20:26.795865823Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_org_id - v1-7" duration=1.549016ms grafana | logger=migrator t=2025-06-18T15:20:26.799898622Z level=info msg="Executing migration" id="create index IDX_temp_user_code - v1-7" grafana | logger=migrator t=2025-06-18T15:20:26.8006655Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_code - v1-7" duration=764.248µs grafana | logger=migrator t=2025-06-18T15:20:26.805960442Z level=info msg="Executing migration" id="create index IDX_temp_user_status - v1-7" grafana | logger=migrator t=2025-06-18T15:20:26.807222054Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_status - v1-7" duration=1.260222ms grafana | logger=migrator t=2025-06-18T15:20:26.811603707Z level=info msg="Executing migration" id="Update temp_user table charset" grafana | logger=migrator t=2025-06-18T15:20:26.811658817Z level=info msg="Migration successfully executed" id="Update temp_user table charset" duration=55.57µs grafana | logger=migrator t=2025-06-18T15:20:26.815516965Z level=info msg="Executing migration" id="drop index IDX_temp_user_email - v1" grafana | logger=migrator t=2025-06-18T15:20:26.816334532Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_email - v1" duration=818.457µs grafana | logger=migrator t=2025-06-18T15:20:26.837132576Z level=info msg="Executing migration" id="drop index IDX_temp_user_org_id - v1" grafana | logger=migrator t=2025-06-18T15:20:26.838435488Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_org_id - v1" duration=1.303842ms grafana | logger=migrator t=2025-06-18T15:20:26.865490352Z level=info msg="Executing migration" id="drop index IDX_temp_user_code - v1" grafana | logger=migrator t=2025-06-18T15:20:26.866495221Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_code - v1" duration=1.009169ms grafana | logger=migrator t=2025-06-18T15:20:26.870496751Z level=info msg="Executing migration" id="drop index IDX_temp_user_status - v1" grafana | logger=migrator t=2025-06-18T15:20:26.871297039Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_status - v1" duration=800.138µs grafana | logger=migrator t=2025-06-18T15:20:26.875277517Z level=info msg="Executing migration" id="Rename table temp_user to temp_user_tmp_qwerty - v1" grafana | logger=migrator t=2025-06-18T15:20:26.878800252Z level=info msg="Migration successfully executed" id="Rename table temp_user to temp_user_tmp_qwerty - v1" duration=3.522855ms grafana | logger=migrator t=2025-06-18T15:20:26.887209414Z level=info msg="Executing migration" id="create temp_user v2" grafana | logger=migrator t=2025-06-18T15:20:26.888298664Z level=info msg="Migration successfully executed" id="create temp_user v2" duration=1.09152ms grafana | logger=migrator t=2025-06-18T15:20:26.892095781Z level=info msg="Executing migration" id="create index IDX_temp_user_email - v2" grafana | logger=migrator t=2025-06-18T15:20:26.892859439Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_email - v2" duration=761.038µs grafana | logger=migrator t=2025-06-18T15:20:26.896170691Z level=info msg="Executing migration" id="create index IDX_temp_user_org_id - v2" grafana | logger=migrator t=2025-06-18T15:20:26.896924798Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_org_id - v2" duration=753.157µs grafana | logger=migrator t=2025-06-18T15:20:26.903085868Z level=info msg="Executing migration" id="create index IDX_temp_user_code - v2" grafana | logger=migrator t=2025-06-18T15:20:26.904325871Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_code - v2" duration=1.240153ms grafana | logger=migrator t=2025-06-18T15:20:26.908253399Z level=info msg="Executing migration" id="create index IDX_temp_user_status - v2" grafana | logger=migrator t=2025-06-18T15:20:26.909912495Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_status - v2" duration=1.661166ms grafana | logger=migrator t=2025-06-18T15:20:26.915496039Z level=info msg="Executing migration" id="copy temp_user v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:26.915960174Z level=info msg="Migration successfully executed" id="copy temp_user v1 to v2" duration=461.525µs grafana | logger=migrator t=2025-06-18T15:20:26.920226205Z level=info msg="Executing migration" id="drop temp_user_tmp_qwerty" grafana | logger=migrator t=2025-06-18T15:20:26.921327106Z level=info msg="Migration successfully executed" id="drop temp_user_tmp_qwerty" duration=1.100091ms grafana | logger=migrator t=2025-06-18T15:20:26.926420636Z level=info msg="Executing migration" id="Set created for temp users that will otherwise prematurely expire" grafana | logger=migrator t=2025-06-18T15:20:26.927244364Z level=info msg="Migration successfully executed" id="Set created for temp users that will otherwise prematurely expire" duration=823.478µs grafana | logger=migrator t=2025-06-18T15:20:26.932725577Z level=info msg="Executing migration" id="create star table" grafana | logger=migrator t=2025-06-18T15:20:26.933592925Z level=info msg="Migration successfully executed" id="create star table" duration=866.728µs grafana | logger=migrator t=2025-06-18T15:20:26.938483203Z level=info msg="Executing migration" id="add unique index star.user_id_dashboard_id" grafana | logger=migrator t=2025-06-18T15:20:26.939333011Z level=info msg="Migration successfully executed" id="add unique index star.user_id_dashboard_id" duration=851.588µs grafana | logger=migrator t=2025-06-18T15:20:26.943796355Z level=info msg="Executing migration" id="Add column dashboard_uid in star" grafana | logger=migrator t=2025-06-18T15:20:26.94531815Z level=info msg="Migration successfully executed" id="Add column dashboard_uid in star" duration=1.521655ms grafana | logger=migrator t=2025-06-18T15:20:26.950666632Z level=info msg="Executing migration" id="Add column org_id in star" grafana | logger=migrator t=2025-06-18T15:20:26.953006565Z level=info msg="Migration successfully executed" id="Add column org_id in star" duration=2.338423ms grafana | logger=migrator t=2025-06-18T15:20:26.95755485Z level=info msg="Executing migration" id="Add column updated in star" grafana | logger=migrator t=2025-06-18T15:20:26.960054874Z level=info msg="Migration successfully executed" id="Add column updated in star" duration=2.499014ms grafana | logger=migrator t=2025-06-18T15:20:26.96482024Z level=info msg="Executing migration" id="add index in star table on dashboard_uid, org_id and user_id columns" grafana | logger=migrator t=2025-06-18T15:20:26.965715059Z level=info msg="Migration successfully executed" id="add index in star table on dashboard_uid, org_id and user_id columns" duration=894.419µs grafana | logger=migrator t=2025-06-18T15:20:26.969025521Z level=info msg="Executing migration" id="create org table v1" grafana | logger=migrator t=2025-06-18T15:20:26.96992423Z level=info msg="Migration successfully executed" id="create org table v1" duration=896.219µs grafana | logger=migrator t=2025-06-18T15:20:26.975675986Z level=info msg="Executing migration" id="create index UQE_org_name - v1" grafana | logger=migrator t=2025-06-18T15:20:26.97708381Z level=info msg="Migration successfully executed" id="create index UQE_org_name - v1" duration=1.407694ms grafana | logger=migrator t=2025-06-18T15:20:26.980998478Z level=info msg="Executing migration" id="create org_user table v1" grafana | logger=migrator t=2025-06-18T15:20:26.981621274Z level=info msg="Migration successfully executed" id="create org_user table v1" duration=622.726µs grafana | logger=migrator t=2025-06-18T15:20:27.00272896Z level=info msg="Executing migration" id="create index IDX_org_user_org_id - v1" grafana | logger=migrator t=2025-06-18T15:20:27.004295234Z level=info msg="Migration successfully executed" id="create index IDX_org_user_org_id - v1" duration=1.569065ms grafana | logger=migrator t=2025-06-18T15:20:27.019751495Z level=info msg="Executing migration" id="create index UQE_org_user_org_id_user_id - v1" grafana | logger=migrator t=2025-06-18T15:20:27.021420481Z level=info msg="Migration successfully executed" id="create index UQE_org_user_org_id_user_id - v1" duration=1.668546ms grafana | logger=migrator t=2025-06-18T15:20:27.027231778Z level=info msg="Executing migration" id="create index IDX_org_user_user_id - v1" grafana | logger=migrator t=2025-06-18T15:20:27.028280927Z level=info msg="Migration successfully executed" id="create index IDX_org_user_user_id - v1" duration=1.052709ms grafana | logger=migrator t=2025-06-18T15:20:27.032414447Z level=info msg="Executing migration" id="Update org table charset" grafana | logger=migrator t=2025-06-18T15:20:27.032448817Z level=info msg="Migration successfully executed" id="Update org table charset" duration=35.16µs grafana | logger=migrator t=2025-06-18T15:20:27.035508348Z level=info msg="Executing migration" id="Update org_user table charset" grafana | logger=migrator t=2025-06-18T15:20:27.035540718Z level=info msg="Migration successfully executed" id="Update org_user table charset" duration=32.17µs grafana | logger=migrator t=2025-06-18T15:20:27.039509456Z level=info msg="Executing migration" id="Migrate all Read Only Viewers to Viewers" grafana | logger=migrator t=2025-06-18T15:20:27.039787528Z level=info msg="Migration successfully executed" id="Migrate all Read Only Viewers to Viewers" duration=280.312µs grafana | logger=migrator t=2025-06-18T15:20:27.053950815Z level=info msg="Executing migration" id="create dashboard table" grafana | logger=migrator t=2025-06-18T15:20:27.05539751Z level=info msg="Migration successfully executed" id="create dashboard table" duration=1.444384ms grafana | logger=migrator t=2025-06-18T15:20:27.059337417Z level=info msg="Executing migration" id="add index dashboard.account_id" grafana | logger=migrator t=2025-06-18T15:20:27.060241506Z level=info msg="Migration successfully executed" id="add index dashboard.account_id" duration=903.869µs grafana | logger=migrator t=2025-06-18T15:20:27.063834011Z level=info msg="Executing migration" id="add unique index dashboard_account_id_slug" grafana | logger=migrator t=2025-06-18T15:20:27.06479038Z level=info msg="Migration successfully executed" id="add unique index dashboard_account_id_slug" duration=955.839µs grafana | logger=migrator t=2025-06-18T15:20:27.068272953Z level=info msg="Executing migration" id="create dashboard_tag table" grafana | logger=migrator t=2025-06-18T15:20:27.069254213Z level=info msg="Migration successfully executed" id="create dashboard_tag table" duration=979.48µs grafana | logger=migrator t=2025-06-18T15:20:27.074879918Z level=info msg="Executing migration" id="add unique index dashboard_tag.dasboard_id_term" grafana | logger=migrator t=2025-06-18T15:20:27.076090539Z level=info msg="Migration successfully executed" id="add unique index dashboard_tag.dasboard_id_term" duration=1.207951ms grafana | logger=migrator t=2025-06-18T15:20:27.083709643Z level=info msg="Executing migration" id="drop index UQE_dashboard_tag_dashboard_id_term - v1" grafana | logger=migrator t=2025-06-18T15:20:27.085268937Z level=info msg="Migration successfully executed" id="drop index UQE_dashboard_tag_dashboard_id_term - v1" duration=1.557174ms grafana | logger=migrator t=2025-06-18T15:20:27.088612539Z level=info msg="Executing migration" id="Rename table dashboard to dashboard_v1 - v1" grafana | logger=migrator t=2025-06-18T15:20:27.095300845Z level=info msg="Migration successfully executed" id="Rename table dashboard to dashboard_v1 - v1" duration=6.689806ms grafana | logger=migrator t=2025-06-18T15:20:27.100158652Z level=info msg="Executing migration" id="create dashboard v2" grafana | logger=migrator t=2025-06-18T15:20:27.100796278Z level=info msg="Migration successfully executed" id="create dashboard v2" duration=636.756µs grafana | logger=migrator t=2025-06-18T15:20:27.10420521Z level=info msg="Executing migration" id="create index IDX_dashboard_org_id - v2" grafana | logger=migrator t=2025-06-18T15:20:27.105086729Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_org_id - v2" duration=881.469µs grafana | logger=migrator t=2025-06-18T15:20:27.110019536Z level=info msg="Executing migration" id="create index UQE_dashboard_org_id_slug - v2" grafana | logger=migrator t=2025-06-18T15:20:27.110905275Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_org_id_slug - v2" duration=883.789µs grafana | logger=migrator t=2025-06-18T15:20:27.116338177Z level=info msg="Executing migration" id="copy dashboard v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:27.117072435Z level=info msg="Migration successfully executed" id="copy dashboard v1 to v2" duration=733.288µs grafana | logger=migrator t=2025-06-18T15:20:27.121072323Z level=info msg="Executing migration" id="drop table dashboard_v1" grafana | logger=migrator t=2025-06-18T15:20:27.122401667Z level=info msg="Migration successfully executed" id="drop table dashboard_v1" duration=1.327874ms grafana | logger=migrator t=2025-06-18T15:20:27.127638446Z level=info msg="Executing migration" id="alter dashboard.data to mediumtext v1" grafana | logger=migrator t=2025-06-18T15:20:27.127669147Z level=info msg="Migration successfully executed" id="alter dashboard.data to mediumtext v1" duration=32.401µs grafana | logger=migrator t=2025-06-18T15:20:27.132525324Z level=info msg="Executing migration" id="Add column updated_by in dashboard - v2" grafana | logger=migrator t=2025-06-18T15:20:27.134392602Z level=info msg="Migration successfully executed" id="Add column updated_by in dashboard - v2" duration=1.865338ms grafana | logger=migrator t=2025-06-18T15:20:27.137869705Z level=info msg="Executing migration" id="Add column created_by in dashboard - v2" grafana | logger=migrator t=2025-06-18T15:20:27.139590612Z level=info msg="Migration successfully executed" id="Add column created_by in dashboard - v2" duration=1.720287ms grafana | logger=migrator t=2025-06-18T15:20:27.145066075Z level=info msg="Executing migration" id="Add column gnetId in dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.147250736Z level=info msg="Migration successfully executed" id="Add column gnetId in dashboard" duration=2.183481ms grafana | logger=migrator t=2025-06-18T15:20:27.151404246Z level=info msg="Executing migration" id="Add index for gnetId in dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.152303305Z level=info msg="Migration successfully executed" id="Add index for gnetId in dashboard" duration=898.689µs grafana | logger=migrator t=2025-06-18T15:20:27.19000187Z level=info msg="Executing migration" id="Add column plugin_id in dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.192480514Z level=info msg="Migration successfully executed" id="Add column plugin_id in dashboard" duration=2.479124ms grafana | logger=migrator t=2025-06-18T15:20:27.198176129Z level=info msg="Executing migration" id="Add index for plugin_id in dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.199507731Z level=info msg="Migration successfully executed" id="Add index for plugin_id in dashboard" duration=1.331182ms grafana | logger=migrator t=2025-06-18T15:20:27.203591541Z level=info msg="Executing migration" id="Add index for dashboard_id in dashboard_tag" grafana | logger=migrator t=2025-06-18T15:20:27.204894313Z level=info msg="Migration successfully executed" id="Add index for dashboard_id in dashboard_tag" duration=1.302882ms grafana | logger=migrator t=2025-06-18T15:20:27.208256265Z level=info msg="Executing migration" id="Update dashboard table charset" grafana | logger=migrator t=2025-06-18T15:20:27.208280926Z level=info msg="Migration successfully executed" id="Update dashboard table charset" duration=25.231µs grafana | logger=migrator t=2025-06-18T15:20:27.228937886Z level=info msg="Executing migration" id="Update dashboard_tag table charset" grafana | logger=migrator t=2025-06-18T15:20:27.228981826Z level=info msg="Migration successfully executed" id="Update dashboard_tag table charset" duration=44.84µs grafana | logger=migrator t=2025-06-18T15:20:27.234089505Z level=info msg="Executing migration" id="Add column folder_id in dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.237305296Z level=info msg="Migration successfully executed" id="Add column folder_id in dashboard" duration=3.212731ms grafana | logger=migrator t=2025-06-18T15:20:27.241998611Z level=info msg="Executing migration" id="Add column isFolder in dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.24505326Z level=info msg="Migration successfully executed" id="Add column isFolder in dashboard" duration=3.055119ms grafana | logger=migrator t=2025-06-18T15:20:27.251328032Z level=info msg="Executing migration" id="Add column has_acl in dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.253496282Z level=info msg="Migration successfully executed" id="Add column has_acl in dashboard" duration=2.16751ms grafana | logger=migrator t=2025-06-18T15:20:27.260371739Z level=info msg="Executing migration" id="Add column uid in dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.26252283Z level=info msg="Migration successfully executed" id="Add column uid in dashboard" duration=2.150241ms grafana | logger=migrator t=2025-06-18T15:20:27.266140075Z level=info msg="Executing migration" id="Update uid column values in dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.266346696Z level=info msg="Migration successfully executed" id="Update uid column values in dashboard" duration=206.981µs grafana | logger=migrator t=2025-06-18T15:20:27.26986487Z level=info msg="Executing migration" id="Add unique index dashboard_org_id_uid" grafana | logger=migrator t=2025-06-18T15:20:27.270627317Z level=info msg="Migration successfully executed" id="Add unique index dashboard_org_id_uid" duration=761.797µs grafana | logger=migrator t=2025-06-18T15:20:27.276411304Z level=info msg="Executing migration" id="Remove unique index org_id_slug" grafana | logger=migrator t=2025-06-18T15:20:27.277193801Z level=info msg="Migration successfully executed" id="Remove unique index org_id_slug" duration=780.567µs grafana | logger=migrator t=2025-06-18T15:20:27.281680504Z level=info msg="Executing migration" id="Update dashboard title length" grafana | logger=migrator t=2025-06-18T15:20:27.281718394Z level=info msg="Migration successfully executed" id="Update dashboard title length" duration=39.01µs grafana | logger=migrator t=2025-06-18T15:20:27.286348319Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_title_folder_id" grafana | logger=migrator t=2025-06-18T15:20:27.287627932Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_title_folder_id" duration=1.278583ms grafana | logger=migrator t=2025-06-18T15:20:27.29359223Z level=info msg="Executing migration" id="create dashboard_provisioning" grafana | logger=migrator t=2025-06-18T15:20:27.294304766Z level=info msg="Migration successfully executed" id="create dashboard_provisioning" duration=712.256µs grafana | logger=migrator t=2025-06-18T15:20:27.298021102Z level=info msg="Executing migration" id="Rename table dashboard_provisioning to dashboard_provisioning_tmp_qwerty - v1" grafana | logger=migrator t=2025-06-18T15:20:27.306822637Z level=info msg="Migration successfully executed" id="Rename table dashboard_provisioning to dashboard_provisioning_tmp_qwerty - v1" duration=8.805075ms grafana | logger=migrator t=2025-06-18T15:20:27.325530538Z level=info msg="Executing migration" id="create dashboard_provisioning v2" grafana | logger=migrator t=2025-06-18T15:20:27.326851081Z level=info msg="Migration successfully executed" id="create dashboard_provisioning v2" duration=1.325353ms grafana | logger=migrator t=2025-06-18T15:20:27.33198118Z level=info msg="Executing migration" id="create index IDX_dashboard_provisioning_dashboard_id - v2" grafana | logger=migrator t=2025-06-18T15:20:27.333653076Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_provisioning_dashboard_id - v2" duration=1.672606ms grafana | logger=migrator t=2025-06-18T15:20:27.359260074Z level=info msg="Executing migration" id="create index IDX_dashboard_provisioning_dashboard_id_name - v2" grafana | logger=migrator t=2025-06-18T15:20:27.360953431Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_provisioning_dashboard_id_name - v2" duration=1.696837ms grafana | logger=migrator t=2025-06-18T15:20:27.364820658Z level=info msg="Executing migration" id="copy dashboard_provisioning v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:27.365204311Z level=info msg="Migration successfully executed" id="copy dashboard_provisioning v1 to v2" duration=383.303µs grafana | logger=migrator t=2025-06-18T15:20:27.37131682Z level=info msg="Executing migration" id="drop dashboard_provisioning_tmp_qwerty" grafana | logger=migrator t=2025-06-18T15:20:27.37227754Z level=info msg="Migration successfully executed" id="drop dashboard_provisioning_tmp_qwerty" duration=959.89µs grafana | logger=migrator t=2025-06-18T15:20:27.376108387Z level=info msg="Executing migration" id="Add check_sum column" grafana | logger=migrator t=2025-06-18T15:20:27.379856382Z level=info msg="Migration successfully executed" id="Add check_sum column" duration=3.747255ms grafana | logger=migrator t=2025-06-18T15:20:27.383834461Z level=info msg="Executing migration" id="Add index for dashboard_title" grafana | logger=migrator t=2025-06-18T15:20:27.385099473Z level=info msg="Migration successfully executed" id="Add index for dashboard_title" duration=1.260172ms grafana | logger=migrator t=2025-06-18T15:20:27.391231313Z level=info msg="Executing migration" id="delete tags for deleted dashboards" grafana | logger=migrator t=2025-06-18T15:20:27.391427415Z level=info msg="Migration successfully executed" id="delete tags for deleted dashboards" duration=196.942µs grafana | logger=migrator t=2025-06-18T15:20:27.394234552Z level=info msg="Executing migration" id="delete stars for deleted dashboards" grafana | logger=migrator t=2025-06-18T15:20:27.394627716Z level=info msg="Migration successfully executed" id="delete stars for deleted dashboards" duration=392.154µs grafana | logger=migrator t=2025-06-18T15:20:27.39820936Z level=info msg="Executing migration" id="Add index for dashboard_is_folder" grafana | logger=migrator t=2025-06-18T15:20:27.399768166Z level=info msg="Migration successfully executed" id="Add index for dashboard_is_folder" duration=1.559075ms grafana | logger=migrator t=2025-06-18T15:20:27.404239108Z level=info msg="Executing migration" id="Add isPublic for dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.408327188Z level=info msg="Migration successfully executed" id="Add isPublic for dashboard" duration=4.08768ms grafana | logger=migrator t=2025-06-18T15:20:27.413631269Z level=info msg="Executing migration" id="Add deleted for dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.416366205Z level=info msg="Migration successfully executed" id="Add deleted for dashboard" duration=2.721266ms grafana | logger=migrator t=2025-06-18T15:20:27.420119982Z level=info msg="Executing migration" id="Add index for deleted" grafana | logger=migrator t=2025-06-18T15:20:27.420943089Z level=info msg="Migration successfully executed" id="Add index for deleted" duration=824.317µs grafana | logger=migrator t=2025-06-18T15:20:27.42408202Z level=info msg="Executing migration" id="Add column dashboard_uid in dashboard_tag" grafana | logger=migrator t=2025-06-18T15:20:27.425876487Z level=info msg="Migration successfully executed" id="Add column dashboard_uid in dashboard_tag" duration=1.793897ms grafana | logger=migrator t=2025-06-18T15:20:27.431261739Z level=info msg="Executing migration" id="Add column org_id in dashboard_tag" grafana | logger=migrator t=2025-06-18T15:20:27.434928324Z level=info msg="Migration successfully executed" id="Add column org_id in dashboard_tag" duration=3.666285ms grafana | logger=migrator t=2025-06-18T15:20:27.439169336Z level=info msg="Executing migration" id="Add missing dashboard_uid and org_id to dashboard_tag" grafana | logger=migrator t=2025-06-18T15:20:27.439857732Z level=info msg="Migration successfully executed" id="Add missing dashboard_uid and org_id to dashboard_tag" duration=689.776µs grafana | logger=migrator t=2025-06-18T15:20:27.44383483Z level=info msg="Executing migration" id="Add apiVersion for dashboard" grafana | logger=migrator t=2025-06-18T15:20:27.446304334Z level=info msg="Migration successfully executed" id="Add apiVersion for dashboard" duration=2.468644ms grafana | logger=migrator t=2025-06-18T15:20:27.451572055Z level=info msg="Executing migration" id="Add index for dashboard_uid on dashboard_tag table" grafana | logger=migrator t=2025-06-18T15:20:27.452523374Z level=info msg="Migration successfully executed" id="Add index for dashboard_uid on dashboard_tag table" duration=950.789µs grafana | logger=migrator t=2025-06-18T15:20:27.455868687Z level=info msg="Executing migration" id="Add missing dashboard_uid and org_id to star" grafana | logger=migrator t=2025-06-18T15:20:27.456468393Z level=info msg="Migration successfully executed" id="Add missing dashboard_uid and org_id to star" duration=599.086µs grafana | logger=migrator t=2025-06-18T15:20:27.459732274Z level=info msg="Executing migration" id="create data_source table" grafana | logger=migrator t=2025-06-18T15:20:27.460735463Z level=info msg="Migration successfully executed" id="create data_source table" duration=1.003499ms grafana | logger=migrator t=2025-06-18T15:20:27.497321437Z level=info msg="Executing migration" id="add index data_source.account_id" grafana | logger=migrator t=2025-06-18T15:20:27.498871272Z level=info msg="Migration successfully executed" id="add index data_source.account_id" duration=1.552575ms grafana | logger=migrator t=2025-06-18T15:20:27.503178373Z level=info msg="Executing migration" id="add unique index data_source.account_id_name" grafana | logger=migrator t=2025-06-18T15:20:27.504434756Z level=info msg="Migration successfully executed" id="add unique index data_source.account_id_name" duration=1.257233ms grafana | logger=migrator t=2025-06-18T15:20:27.530848261Z level=info msg="Executing migration" id="drop index IDX_data_source_account_id - v1" grafana | logger=migrator t=2025-06-18T15:20:27.532168354Z level=info msg="Migration successfully executed" id="drop index IDX_data_source_account_id - v1" duration=1.322073ms grafana | logger=migrator t=2025-06-18T15:20:27.537498406Z level=info msg="Executing migration" id="drop index UQE_data_source_account_id_name - v1" grafana | logger=migrator t=2025-06-18T15:20:27.538312843Z level=info msg="Migration successfully executed" id="drop index UQE_data_source_account_id_name - v1" duration=814.677µs grafana | logger=migrator t=2025-06-18T15:20:27.543991438Z level=info msg="Executing migration" id="Rename table data_source to data_source_v1 - v1" grafana | logger=migrator t=2025-06-18T15:20:27.554683631Z level=info msg="Migration successfully executed" id="Rename table data_source to data_source_v1 - v1" duration=10.689463ms grafana | logger=migrator t=2025-06-18T15:20:27.558896402Z level=info msg="Executing migration" id="create data_source table v2" grafana | logger=migrator t=2025-06-18T15:20:27.559515458Z level=info msg="Migration successfully executed" id="create data_source table v2" duration=619.036µs grafana | logger=migrator t=2025-06-18T15:20:27.563104763Z level=info msg="Executing migration" id="create index IDX_data_source_org_id - v2" grafana | logger=migrator t=2025-06-18T15:20:27.563754679Z level=info msg="Migration successfully executed" id="create index IDX_data_source_org_id - v2" duration=649.626µs grafana | logger=migrator t=2025-06-18T15:20:27.568938609Z level=info msg="Executing migration" id="create index UQE_data_source_org_id_name - v2" grafana | logger=migrator t=2025-06-18T15:20:27.570243341Z level=info msg="Migration successfully executed" id="create index UQE_data_source_org_id_name - v2" duration=1.303572ms grafana | logger=migrator t=2025-06-18T15:20:27.576119249Z level=info msg="Executing migration" id="Drop old table data_source_v1 #2" grafana | logger=migrator t=2025-06-18T15:20:27.576983697Z level=info msg="Migration successfully executed" id="Drop old table data_source_v1 #2" duration=863.768µs grafana | logger=migrator t=2025-06-18T15:20:27.581258088Z level=info msg="Executing migration" id="Add column with_credentials" grafana | logger=migrator t=2025-06-18T15:20:27.585501969Z level=info msg="Migration successfully executed" id="Add column with_credentials" duration=4.240231ms grafana | logger=migrator t=2025-06-18T15:20:27.590965102Z level=info msg="Executing migration" id="Add secure json data column" grafana | logger=migrator t=2025-06-18T15:20:27.594189513Z level=info msg="Migration successfully executed" id="Add secure json data column" duration=3.219271ms grafana | logger=migrator t=2025-06-18T15:20:27.647558088Z level=info msg="Executing migration" id="Update data_source table charset" grafana | logger=migrator t=2025-06-18T15:20:27.647699489Z level=info msg="Migration successfully executed" id="Update data_source table charset" duration=144.951µs grafana | logger=migrator t=2025-06-18T15:20:27.655549925Z level=info msg="Executing migration" id="Update initial version to 1" grafana | logger=migrator t=2025-06-18T15:20:27.655861528Z level=info msg="Migration successfully executed" id="Update initial version to 1" duration=313.663µs grafana | logger=migrator t=2025-06-18T15:20:27.663183259Z level=info msg="Executing migration" id="Add read_only data column" grafana | logger=migrator t=2025-06-18T15:20:27.670015295Z level=info msg="Migration successfully executed" id="Add read_only data column" duration=6.826546ms grafana | logger=migrator t=2025-06-18T15:20:27.688948428Z level=info msg="Executing migration" id="Migrate logging ds to loki ds" grafana | logger=migrator t=2025-06-18T15:20:27.689339482Z level=info msg="Migration successfully executed" id="Migrate logging ds to loki ds" duration=392.144µs grafana | logger=migrator t=2025-06-18T15:20:27.701292447Z level=info msg="Executing migration" id="Update json_data with nulls" grafana | logger=migrator t=2025-06-18T15:20:27.701877153Z level=info msg="Migration successfully executed" id="Update json_data with nulls" duration=583.796µs grafana | logger=migrator t=2025-06-18T15:20:27.707541337Z level=info msg="Executing migration" id="Add uid column" grafana | logger=migrator t=2025-06-18T15:20:27.710400095Z level=info msg="Migration successfully executed" id="Add uid column" duration=2.858148ms grafana | logger=migrator t=2025-06-18T15:20:27.715703166Z level=info msg="Executing migration" id="Update uid value" grafana | logger=migrator t=2025-06-18T15:20:27.715942948Z level=info msg="Migration successfully executed" id="Update uid value" duration=239.522µs grafana | logger=migrator t=2025-06-18T15:20:27.726322609Z level=info msg="Executing migration" id="Add unique index datasource_org_id_uid" grafana | logger=migrator t=2025-06-18T15:20:27.727814424Z level=info msg="Migration successfully executed" id="Add unique index datasource_org_id_uid" duration=1.493645ms grafana | logger=migrator t=2025-06-18T15:20:27.735793051Z level=info msg="Executing migration" id="add unique index datasource_org_id_is_default" grafana | logger=migrator t=2025-06-18T15:20:27.736666679Z level=info msg="Migration successfully executed" id="add unique index datasource_org_id_is_default" duration=873.028µs grafana | logger=migrator t=2025-06-18T15:20:27.742510305Z level=info msg="Executing migration" id="Add is_prunable column" grafana | logger=migrator t=2025-06-18T15:20:27.746636115Z level=info msg="Migration successfully executed" id="Add is_prunable column" duration=4.12401ms grafana | logger=migrator t=2025-06-18T15:20:27.751429871Z level=info msg="Executing migration" id="Add api_version column" grafana | logger=migrator t=2025-06-18T15:20:27.754180948Z level=info msg="Migration successfully executed" id="Add api_version column" duration=2.751397ms grafana | logger=migrator t=2025-06-18T15:20:27.757201087Z level=info msg="Executing migration" id="Update secure_json_data column to MediumText" grafana | logger=migrator t=2025-06-18T15:20:27.757217947Z level=info msg="Migration successfully executed" id="Update secure_json_data column to MediumText" duration=17.46µs grafana | logger=migrator t=2025-06-18T15:20:27.762353767Z level=info msg="Executing migration" id="create api_key table" grafana | logger=migrator t=2025-06-18T15:20:27.763072744Z level=info msg="Migration successfully executed" id="create api_key table" duration=719.137µs grafana | logger=migrator t=2025-06-18T15:20:27.768068692Z level=info msg="Executing migration" id="add index api_key.account_id" grafana | logger=migrator t=2025-06-18T15:20:27.769062221Z level=info msg="Migration successfully executed" id="add index api_key.account_id" duration=991.779µs grafana | logger=migrator t=2025-06-18T15:20:27.773514505Z level=info msg="Executing migration" id="add index api_key.key" grafana | logger=migrator t=2025-06-18T15:20:27.774841347Z level=info msg="Migration successfully executed" id="add index api_key.key" duration=1.326332ms grafana | logger=migrator t=2025-06-18T15:20:27.780969827Z level=info msg="Executing migration" id="add index api_key.account_id_name" grafana | logger=migrator t=2025-06-18T15:20:27.781824695Z level=info msg="Migration successfully executed" id="add index api_key.account_id_name" duration=854.218µs grafana | logger=migrator t=2025-06-18T15:20:27.785144067Z level=info msg="Executing migration" id="drop index IDX_api_key_account_id - v1" grafana | logger=migrator t=2025-06-18T15:20:27.785934164Z level=info msg="Migration successfully executed" id="drop index IDX_api_key_account_id - v1" duration=789.927µs grafana | logger=migrator t=2025-06-18T15:20:27.811856655Z level=info msg="Executing migration" id="drop index UQE_api_key_key - v1" grafana | logger=migrator t=2025-06-18T15:20:27.813214378Z level=info msg="Migration successfully executed" id="drop index UQE_api_key_key - v1" duration=1.358153ms grafana | logger=migrator t=2025-06-18T15:20:27.86624649Z level=info msg="Executing migration" id="drop index UQE_api_key_account_id_name - v1" grafana | logger=migrator t=2025-06-18T15:20:27.867722315Z level=info msg="Migration successfully executed" id="drop index UQE_api_key_account_id_name - v1" duration=1.477645ms grafana | logger=migrator t=2025-06-18T15:20:27.881526718Z level=info msg="Executing migration" id="Rename table api_key to api_key_v1 - v1" grafana | logger=migrator t=2025-06-18T15:20:27.889794117Z level=info msg="Migration successfully executed" id="Rename table api_key to api_key_v1 - v1" duration=8.266049ms grafana | logger=migrator t=2025-06-18T15:20:27.89932822Z level=info msg="Executing migration" id="create api_key table v2" grafana | logger=migrator t=2025-06-18T15:20:27.900676333Z level=info msg="Migration successfully executed" id="create api_key table v2" duration=1.358042ms grafana | logger=migrator t=2025-06-18T15:20:27.910549948Z level=info msg="Executing migration" id="create index IDX_api_key_org_id - v2" grafana | logger=migrator t=2025-06-18T15:20:27.911322916Z level=info msg="Migration successfully executed" id="create index IDX_api_key_org_id - v2" duration=773.237µs grafana | logger=migrator t=2025-06-18T15:20:27.924878037Z level=info msg="Executing migration" id="create index UQE_api_key_key - v2" grafana | logger=migrator t=2025-06-18T15:20:27.926417372Z level=info msg="Migration successfully executed" id="create index UQE_api_key_key - v2" duration=1.541806ms grafana | logger=migrator t=2025-06-18T15:20:27.942302535Z level=info msg="Executing migration" id="create index UQE_api_key_org_id_name - v2" grafana | logger=migrator t=2025-06-18T15:20:27.943604668Z level=info msg="Migration successfully executed" id="create index UQE_api_key_org_id_name - v2" duration=1.307704ms grafana | logger=migrator t=2025-06-18T15:20:27.952328562Z level=info msg="Executing migration" id="copy api_key v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:27.952654415Z level=info msg="Migration successfully executed" id="copy api_key v1 to v2" duration=326.253µs grafana | logger=migrator t=2025-06-18T15:20:27.980170981Z level=info msg="Executing migration" id="Drop old table api_key_v1" grafana | logger=migrator t=2025-06-18T15:20:27.98116202Z level=info msg="Migration successfully executed" id="Drop old table api_key_v1" duration=994.019µs grafana | logger=migrator t=2025-06-18T15:20:27.989625722Z level=info msg="Executing migration" id="Update api_key table charset" grafana | logger=migrator t=2025-06-18T15:20:27.989658642Z level=info msg="Migration successfully executed" id="Update api_key table charset" duration=33.73µs grafana | logger=migrator t=2025-06-18T15:20:27.995835981Z level=info msg="Executing migration" id="Add expires to api_key table" grafana | logger=migrator t=2025-06-18T15:20:28.000639728Z level=info msg="Migration successfully executed" id="Add expires to api_key table" duration=4.802457ms grafana | logger=migrator t=2025-06-18T15:20:28.034190081Z level=info msg="Executing migration" id="Add service account foreign key" grafana | logger=migrator t=2025-06-18T15:20:28.041257277Z level=info msg="Migration successfully executed" id="Add service account foreign key" duration=7.065946ms grafana | logger=migrator t=2025-06-18T15:20:28.044994691Z level=info msg="Executing migration" id="set service account foreign key to nil if 0" grafana | logger=migrator t=2025-06-18T15:20:28.045215943Z level=info msg="Migration successfully executed" id="set service account foreign key to nil if 0" duration=221.472µs grafana | logger=migrator t=2025-06-18T15:20:28.049579434Z level=info msg="Executing migration" id="Add last_used_at to api_key table" grafana | logger=migrator t=2025-06-18T15:20:28.053086776Z level=info msg="Migration successfully executed" id="Add last_used_at to api_key table" duration=3.507592ms grafana | logger=migrator t=2025-06-18T15:20:28.065320889Z level=info msg="Executing migration" id="Add is_revoked column to api_key table" grafana | logger=migrator t=2025-06-18T15:20:28.069733251Z level=info msg="Migration successfully executed" id="Add is_revoked column to api_key table" duration=4.414941ms grafana | logger=migrator t=2025-06-18T15:20:28.08168585Z level=info msg="Executing migration" id="create dashboard_snapshot table v4" grafana | logger=migrator t=2025-06-18T15:20:28.082520868Z level=info msg="Migration successfully executed" id="create dashboard_snapshot table v4" duration=837.098µs grafana | logger=migrator t=2025-06-18T15:20:28.094038464Z level=info msg="Executing migration" id="drop table dashboard_snapshot_v4 #1" grafana | logger=migrator t=2025-06-18T15:20:28.094969244Z level=info msg="Migration successfully executed" id="drop table dashboard_snapshot_v4 #1" duration=931.22µs grafana | logger=migrator t=2025-06-18T15:20:28.122827472Z level=info msg="Executing migration" id="create dashboard_snapshot table v5 #2" grafana | logger=migrator t=2025-06-18T15:20:28.12380255Z level=info msg="Migration successfully executed" id="create dashboard_snapshot table v5 #2" duration=977.598µs grafana | logger=migrator t=2025-06-18T15:20:28.128846327Z level=info msg="Executing migration" id="create index UQE_dashboard_snapshot_key - v5" grafana | logger=migrator t=2025-06-18T15:20:28.13028238Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_snapshot_key - v5" duration=1.435203ms grafana | logger=migrator t=2025-06-18T15:20:28.139423995Z level=info msg="Executing migration" id="create index UQE_dashboard_snapshot_delete_key - v5" grafana | logger=migrator t=2025-06-18T15:20:28.140696377Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_snapshot_delete_key - v5" duration=1.269382ms grafana | logger=migrator t=2025-06-18T15:20:28.147002245Z level=info msg="Executing migration" id="create index IDX_dashboard_snapshot_user_id - v5" grafana | logger=migrator t=2025-06-18T15:20:28.147978604Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_snapshot_user_id - v5" duration=979.739µs grafana | logger=migrator t=2025-06-18T15:20:28.157066819Z level=info msg="Executing migration" id="alter dashboard_snapshot to mediumtext v2" grafana | logger=migrator t=2025-06-18T15:20:28.157092109Z level=info msg="Migration successfully executed" id="alter dashboard_snapshot to mediumtext v2" duration=27.6µs grafana | logger=migrator t=2025-06-18T15:20:28.160810433Z level=info msg="Executing migration" id="Update dashboard_snapshot table charset" grafana | logger=migrator t=2025-06-18T15:20:28.160834253Z level=info msg="Migration successfully executed" id="Update dashboard_snapshot table charset" duration=24.67µs grafana | logger=migrator t=2025-06-18T15:20:28.16702299Z level=info msg="Executing migration" id="Add column external_delete_url to dashboard_snapshots table" grafana | logger=migrator t=2025-06-18T15:20:28.170045249Z level=info msg="Migration successfully executed" id="Add column external_delete_url to dashboard_snapshots table" duration=3.020649ms grafana | logger=migrator t=2025-06-18T15:20:28.174214107Z level=info msg="Executing migration" id="Add encrypted dashboard json column" grafana | logger=migrator t=2025-06-18T15:20:28.177019283Z level=info msg="Migration successfully executed" id="Add encrypted dashboard json column" duration=2.804736ms grafana | logger=migrator t=2025-06-18T15:20:28.206564137Z level=info msg="Executing migration" id="Change dashboard_encrypted column to MEDIUMBLOB" grafana | logger=migrator t=2025-06-18T15:20:28.206631597Z level=info msg="Migration successfully executed" id="Change dashboard_encrypted column to MEDIUMBLOB" duration=66.62µs grafana | logger=migrator t=2025-06-18T15:20:28.21341735Z level=info msg="Executing migration" id="create quota table v1" grafana | logger=migrator t=2025-06-18T15:20:28.214297927Z level=info msg="Migration successfully executed" id="create quota table v1" duration=881.908µs grafana | logger=migrator t=2025-06-18T15:20:28.218432006Z level=info msg="Executing migration" id="create index UQE_quota_org_id_user_id_target - v1" grafana | logger=migrator t=2025-06-18T15:20:28.219259124Z level=info msg="Migration successfully executed" id="create index UQE_quota_org_id_user_id_target - v1" duration=826.618µs grafana | logger=migrator t=2025-06-18T15:20:28.225005247Z level=info msg="Executing migration" id="Update quota table charset" grafana | logger=migrator t=2025-06-18T15:20:28.225029877Z level=info msg="Migration successfully executed" id="Update quota table charset" duration=23.75µs grafana | logger=migrator t=2025-06-18T15:20:28.232504017Z level=info msg="Executing migration" id="create plugin_setting table" grafana | logger=migrator t=2025-06-18T15:20:28.233818428Z level=info msg="Migration successfully executed" id="create plugin_setting table" duration=1.314031ms grafana | logger=migrator t=2025-06-18T15:20:28.24690964Z level=info msg="Executing migration" id="create index UQE_plugin_setting_org_id_plugin_id - v1" grafana | logger=migrator t=2025-06-18T15:20:28.248409663Z level=info msg="Migration successfully executed" id="create index UQE_plugin_setting_org_id_plugin_id - v1" duration=1.499983ms grafana | logger=migrator t=2025-06-18T15:20:28.255422728Z level=info msg="Executing migration" id="Add column plugin_version to plugin_settings" grafana | logger=migrator t=2025-06-18T15:20:28.260481825Z level=info msg="Migration successfully executed" id="Add column plugin_version to plugin_settings" duration=5.052727ms grafana | logger=migrator t=2025-06-18T15:20:28.2825398Z level=info msg="Executing migration" id="Update plugin_setting table charset" grafana | logger=migrator t=2025-06-18T15:20:28.28261309Z level=info msg="Migration successfully executed" id="Update plugin_setting table charset" duration=77.46µs grafana | logger=migrator t=2025-06-18T15:20:28.288213922Z level=info msg="Executing migration" id="update NULL org_id to 1" grafana | logger=migrator t=2025-06-18T15:20:28.288663716Z level=info msg="Migration successfully executed" id="update NULL org_id to 1" duration=450.274µs grafana | logger=migrator t=2025-06-18T15:20:28.294717202Z level=info msg="Executing migration" id="make org_id NOT NULL and DEFAULT VALUE 1" grafana | logger=migrator t=2025-06-18T15:20:28.307846034Z level=info msg="Migration successfully executed" id="make org_id NOT NULL and DEFAULT VALUE 1" duration=13.126082ms grafana | logger=migrator t=2025-06-18T15:20:28.313016532Z level=info msg="Executing migration" id="create session table" grafana | logger=migrator t=2025-06-18T15:20:28.313760598Z level=info msg="Migration successfully executed" id="create session table" duration=745.126µs grafana | logger=migrator t=2025-06-18T15:20:28.318831046Z level=info msg="Executing migration" id="Drop old table playlist table" grafana | logger=migrator t=2025-06-18T15:20:28.318942717Z level=info msg="Migration successfully executed" id="Drop old table playlist table" duration=112.751µs grafana | logger=migrator t=2025-06-18T15:20:28.32588517Z level=info msg="Executing migration" id="Drop old table playlist_item table" grafana | logger=migrator t=2025-06-18T15:20:28.325968621Z level=info msg="Migration successfully executed" id="Drop old table playlist_item table" duration=83.631µs grafana | logger=migrator t=2025-06-18T15:20:28.33014697Z level=info msg="Executing migration" id="create playlist table v2" grafana | logger=migrator t=2025-06-18T15:20:28.330797306Z level=info msg="Migration successfully executed" id="create playlist table v2" duration=652.596µs grafana | logger=migrator t=2025-06-18T15:20:28.338110044Z level=info msg="Executing migration" id="create playlist item table v2" grafana | logger=migrator t=2025-06-18T15:20:28.339030402Z level=info msg="Migration successfully executed" id="create playlist item table v2" duration=923.388µs grafana | logger=migrator t=2025-06-18T15:20:28.343518683Z level=info msg="Executing migration" id="Update playlist table charset" grafana | logger=migrator t=2025-06-18T15:20:28.343556924Z level=info msg="Migration successfully executed" id="Update playlist table charset" duration=39.071µs grafana | logger=migrator t=2025-06-18T15:20:28.348412069Z level=info msg="Executing migration" id="Update playlist_item table charset" grafana | logger=migrator t=2025-06-18T15:20:28.348445769Z level=info msg="Migration successfully executed" id="Update playlist_item table charset" duration=37.54µs grafana | logger=migrator t=2025-06-18T15:20:28.35284707Z level=info msg="Executing migration" id="Add playlist column created_at" grafana | logger=migrator t=2025-06-18T15:20:28.356890817Z level=info msg="Migration successfully executed" id="Add playlist column created_at" duration=4.043337ms grafana | logger=migrator t=2025-06-18T15:20:28.37878175Z level=info msg="Executing migration" id="Add playlist column updated_at" grafana | logger=migrator t=2025-06-18T15:20:28.382532374Z level=info msg="Migration successfully executed" id="Add playlist column updated_at" duration=3.750004ms grafana | logger=migrator t=2025-06-18T15:20:28.388602571Z level=info msg="Executing migration" id="drop preferences table v2" grafana | logger=migrator t=2025-06-18T15:20:28.388721732Z level=info msg="Migration successfully executed" id="drop preferences table v2" duration=119.611µs grafana | logger=migrator t=2025-06-18T15:20:28.392541977Z level=info msg="Executing migration" id="drop preferences table v3" grafana | logger=migrator t=2025-06-18T15:20:28.392650508Z level=info msg="Migration successfully executed" id="drop preferences table v3" duration=108.871µs grafana | logger=migrator t=2025-06-18T15:20:28.39610586Z level=info msg="Executing migration" id="create preferences table v3" grafana | logger=migrator t=2025-06-18T15:20:28.396807767Z level=info msg="Migration successfully executed" id="create preferences table v3" duration=701.866µs grafana | logger=migrator t=2025-06-18T15:20:28.40252769Z level=info msg="Executing migration" id="Update preferences table charset" grafana | logger=migrator t=2025-06-18T15:20:28.40256174Z level=info msg="Migration successfully executed" id="Update preferences table charset" duration=35.28µs grafana | logger=migrator t=2025-06-18T15:20:28.411426912Z level=info msg="Executing migration" id="Add column team_id in preferences" grafana | logger=migrator t=2025-06-18T15:20:28.415850023Z level=info msg="Migration successfully executed" id="Add column team_id in preferences" duration=4.422091ms grafana | logger=migrator t=2025-06-18T15:20:28.432553177Z level=info msg="Executing migration" id="Update team_id column values in preferences" grafana | logger=migrator t=2025-06-18T15:20:28.433047152Z level=info msg="Migration successfully executed" id="Update team_id column values in preferences" duration=493.585µs grafana | logger=migrator t=2025-06-18T15:20:28.438436382Z level=info msg="Executing migration" id="Add column week_start in preferences" grafana | logger=migrator t=2025-06-18T15:20:28.441832214Z level=info msg="Migration successfully executed" id="Add column week_start in preferences" duration=3.395352ms grafana | logger=migrator t=2025-06-18T15:20:28.446629988Z level=info msg="Executing migration" id="Add column preferences.json_data" grafana | logger=migrator t=2025-06-18T15:20:28.449911438Z level=info msg="Migration successfully executed" id="Add column preferences.json_data" duration=3.28052ms grafana | logger=migrator t=2025-06-18T15:20:28.453401021Z level=info msg="Executing migration" id="alter preferences.json_data to mediumtext v1" grafana | logger=migrator t=2025-06-18T15:20:28.453418971Z level=info msg="Migration successfully executed" id="alter preferences.json_data to mediumtext v1" duration=18.79µs grafana | logger=migrator t=2025-06-18T15:20:28.457227156Z level=info msg="Executing migration" id="Add preferences index org_id" grafana | logger=migrator t=2025-06-18T15:20:28.458120594Z level=info msg="Migration successfully executed" id="Add preferences index org_id" duration=888.278µs grafana | logger=migrator t=2025-06-18T15:20:28.466701163Z level=info msg="Executing migration" id="Add preferences index user_id" grafana | logger=migrator t=2025-06-18T15:20:28.468251658Z level=info msg="Migration successfully executed" id="Add preferences index user_id" duration=1.550015ms grafana | logger=migrator t=2025-06-18T15:20:28.472333805Z level=info msg="Executing migration" id="create alert table v1" grafana | logger=migrator t=2025-06-18T15:20:28.474042182Z level=info msg="Migration successfully executed" id="create alert table v1" duration=1.705327ms grafana | logger=migrator t=2025-06-18T15:20:28.492584714Z level=info msg="Executing migration" id="add index alert org_id & id " grafana | logger=migrator t=2025-06-18T15:20:28.494195448Z level=info msg="Migration successfully executed" id="add index alert org_id & id " duration=1.610355ms grafana | logger=migrator t=2025-06-18T15:20:28.518600364Z level=info msg="Executing migration" id="add index alert state" grafana | logger=migrator t=2025-06-18T15:20:28.520131388Z level=info msg="Migration successfully executed" id="add index alert state" duration=1.541484ms grafana | logger=migrator t=2025-06-18T15:20:28.545146129Z level=info msg="Executing migration" id="add index alert dashboard_id" grafana | logger=migrator t=2025-06-18T15:20:28.547105088Z level=info msg="Migration successfully executed" id="add index alert dashboard_id" duration=1.959899ms grafana | logger=migrator t=2025-06-18T15:20:28.557700966Z level=info msg="Executing migration" id="Create alert_rule_tag table v1" grafana | logger=migrator t=2025-06-18T15:20:28.558390912Z level=info msg="Migration successfully executed" id="Create alert_rule_tag table v1" duration=690.216µs grafana | logger=migrator t=2025-06-18T15:20:28.563752902Z level=info msg="Executing migration" id="Add unique index alert_rule_tag.alert_id_tag_id" grafana | logger=migrator t=2025-06-18T15:20:28.564482168Z level=info msg="Migration successfully executed" id="Add unique index alert_rule_tag.alert_id_tag_id" duration=731.836µs grafana | logger=migrator t=2025-06-18T15:20:28.585627275Z level=info msg="Executing migration" id="drop index UQE_alert_rule_tag_alert_id_tag_id - v1" grafana | logger=migrator t=2025-06-18T15:20:28.587239029Z level=info msg="Migration successfully executed" id="drop index UQE_alert_rule_tag_alert_id_tag_id - v1" duration=1.613655ms grafana | logger=migrator t=2025-06-18T15:20:28.591776941Z level=info msg="Executing migration" id="Rename table alert_rule_tag to alert_rule_tag_v1 - v1" grafana | logger=migrator t=2025-06-18T15:20:28.606403546Z level=info msg="Migration successfully executed" id="Rename table alert_rule_tag to alert_rule_tag_v1 - v1" duration=14.615295ms grafana | logger=migrator t=2025-06-18T15:20:28.613322611Z level=info msg="Executing migration" id="Create alert_rule_tag table v2" grafana | logger=migrator t=2025-06-18T15:20:28.614179108Z level=info msg="Migration successfully executed" id="Create alert_rule_tag table v2" duration=857.257µs grafana | logger=migrator t=2025-06-18T15:20:28.618908703Z level=info msg="Executing migration" id="create index UQE_alert_rule_tag_alert_id_tag_id - Add unique index alert_rule_tag.alert_id_tag_id V2" grafana | logger=migrator t=2025-06-18T15:20:28.620359485Z level=info msg="Migration successfully executed" id="create index UQE_alert_rule_tag_alert_id_tag_id - Add unique index alert_rule_tag.alert_id_tag_id V2" duration=1.450213ms grafana | logger=migrator t=2025-06-18T15:20:28.634121793Z level=info msg="Executing migration" id="copy alert_rule_tag v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:28.634515236Z level=info msg="Migration successfully executed" id="copy alert_rule_tag v1 to v2" duration=394.203µs grafana | logger=migrator t=2025-06-18T15:20:28.643625521Z level=info msg="Executing migration" id="drop table alert_rule_tag_v1" grafana | logger=migrator t=2025-06-18T15:20:28.644347568Z level=info msg="Migration successfully executed" id="drop table alert_rule_tag_v1" duration=727.177µs grafana | logger=migrator t=2025-06-18T15:20:28.651481643Z level=info msg="Executing migration" id="create alert_notification table v1" grafana | logger=migrator t=2025-06-18T15:20:28.652337312Z level=info msg="Migration successfully executed" id="create alert_notification table v1" duration=855.529µs grafana | logger=migrator t=2025-06-18T15:20:28.659911321Z level=info msg="Executing migration" id="Add column is_default" grafana | logger=migrator t=2025-06-18T15:20:28.664361063Z level=info msg="Migration successfully executed" id="Add column is_default" duration=4.451722ms grafana | logger=migrator t=2025-06-18T15:20:28.669511331Z level=info msg="Executing migration" id="Add column frequency" grafana | logger=migrator t=2025-06-18T15:20:28.673425406Z level=info msg="Migration successfully executed" id="Add column frequency" duration=3.898755ms grafana | logger=migrator t=2025-06-18T15:20:28.678636555Z level=info msg="Executing migration" id="Add column send_reminder" grafana | logger=migrator t=2025-06-18T15:20:28.682924415Z level=info msg="Migration successfully executed" id="Add column send_reminder" duration=4.28755ms grafana | logger=migrator t=2025-06-18T15:20:28.711255747Z level=info msg="Executing migration" id="Add column disable_resolve_message" grafana | logger=migrator t=2025-06-18T15:20:28.7160153Z level=info msg="Migration successfully executed" id="Add column disable_resolve_message" duration=4.760113ms grafana | logger=migrator t=2025-06-18T15:20:28.739929452Z level=info msg="Executing migration" id="add index alert_notification org_id & name" grafana | logger=migrator t=2025-06-18T15:20:28.741978071Z level=info msg="Migration successfully executed" id="add index alert_notification org_id & name" duration=2.048059ms grafana | logger=migrator t=2025-06-18T15:20:28.751258097Z level=info msg="Executing migration" id="Update alert table charset" grafana | logger=migrator t=2025-06-18T15:20:28.751317628Z level=info msg="Migration successfully executed" id="Update alert table charset" duration=62.221µs grafana | logger=migrator t=2025-06-18T15:20:28.76122685Z level=info msg="Executing migration" id="Update alert_notification table charset" grafana | logger=migrator t=2025-06-18T15:20:28.76126428Z level=info msg="Migration successfully executed" id="Update alert_notification table charset" duration=39.91µs grafana | logger=migrator t=2025-06-18T15:20:28.770241283Z level=info msg="Executing migration" id="create notification_journal table v1" grafana | logger=migrator t=2025-06-18T15:20:28.77314337Z level=info msg="Migration successfully executed" id="create notification_journal table v1" duration=2.899486ms grafana | logger=migrator t=2025-06-18T15:20:28.781458717Z level=info msg="Executing migration" id="add index notification_journal org_id & alert_id & notifier_id" grafana | logger=migrator t=2025-06-18T15:20:28.782450186Z level=info msg="Migration successfully executed" id="add index notification_journal org_id & alert_id & notifier_id" duration=1.019759ms grafana | logger=migrator t=2025-06-18T15:20:28.790554201Z level=info msg="Executing migration" id="drop alert_notification_journal" grafana | logger=migrator t=2025-06-18T15:20:28.791277637Z level=info msg="Migration successfully executed" id="drop alert_notification_journal" duration=724.996µs grafana | logger=migrator t=2025-06-18T15:20:28.804697562Z level=info msg="Executing migration" id="create alert_notification_state table v1" grafana | logger=migrator t=2025-06-18T15:20:28.805556149Z level=info msg="Migration successfully executed" id="create alert_notification_state table v1" duration=854.737µs grafana | logger=migrator t=2025-06-18T15:20:28.813302701Z level=info msg="Executing migration" id="add index alert_notification_state org_id & alert_id & notifier_id" grafana | logger=migrator t=2025-06-18T15:20:28.814585803Z level=info msg="Migration successfully executed" id="add index alert_notification_state org_id & alert_id & notifier_id" duration=1.287462ms grafana | logger=migrator t=2025-06-18T15:20:28.82511769Z level=info msg="Executing migration" id="Add for to alert table" grafana | logger=migrator t=2025-06-18T15:20:28.832452738Z level=info msg="Migration successfully executed" id="Add for to alert table" duration=7.332418ms grafana | logger=migrator t=2025-06-18T15:20:28.84666864Z level=info msg="Executing migration" id="Add column uid in alert_notification" grafana | logger=migrator t=2025-06-18T15:20:28.851579526Z level=info msg="Migration successfully executed" id="Add column uid in alert_notification" duration=4.907006ms grafana | logger=migrator t=2025-06-18T15:20:28.907799736Z level=info msg="Executing migration" id="Update uid column values in alert_notification" grafana | logger=migrator t=2025-06-18T15:20:28.908096808Z level=info msg="Migration successfully executed" id="Update uid column values in alert_notification" duration=298.992µs grafana | logger=migrator t=2025-06-18T15:20:28.916638037Z level=info msg="Executing migration" id="Add unique index alert_notification_org_id_uid" grafana | logger=migrator t=2025-06-18T15:20:28.918187942Z level=info msg="Migration successfully executed" id="Add unique index alert_notification_org_id_uid" duration=1.549615ms grafana | logger=migrator t=2025-06-18T15:20:28.925033445Z level=info msg="Executing migration" id="Remove unique index org_id_name" grafana | logger=migrator t=2025-06-18T15:20:28.925944693Z level=info msg="Migration successfully executed" id="Remove unique index org_id_name" duration=906.958µs grafana | logger=migrator t=2025-06-18T15:20:28.933231001Z level=info msg="Executing migration" id="Add column secure_settings in alert_notification" grafana | logger=migrator t=2025-06-18T15:20:28.940148175Z level=info msg="Migration successfully executed" id="Add column secure_settings in alert_notification" duration=6.916664ms grafana | logger=migrator t=2025-06-18T15:20:28.946861517Z level=info msg="Executing migration" id="alter alert.settings to mediumtext" grafana | logger=migrator t=2025-06-18T15:20:28.946887168Z level=info msg="Migration successfully executed" id="alter alert.settings to mediumtext" duration=27.751µs grafana | logger=migrator t=2025-06-18T15:20:28.955394556Z level=info msg="Executing migration" id="Add non-unique index alert_notification_state_alert_id" grafana | logger=migrator t=2025-06-18T15:20:28.956351715Z level=info msg="Migration successfully executed" id="Add non-unique index alert_notification_state_alert_id" duration=956.909µs grafana | logger=migrator t=2025-06-18T15:20:28.961639474Z level=info msg="Executing migration" id="Add non-unique index alert_rule_tag_alert_id" grafana | logger=migrator t=2025-06-18T15:20:28.963204869Z level=info msg="Migration successfully executed" id="Add non-unique index alert_rule_tag_alert_id" duration=1.560675ms grafana | logger=migrator t=2025-06-18T15:20:28.975389191Z level=info msg="Executing migration" id="Drop old annotation table v4" grafana | logger=migrator t=2025-06-18T15:20:28.975531952Z level=info msg="Migration successfully executed" id="Drop old annotation table v4" duration=143.291µs grafana | logger=migrator t=2025-06-18T15:20:28.982185494Z level=info msg="Executing migration" id="create annotation table v5" grafana | logger=migrator t=2025-06-18T15:20:28.983668998Z level=info msg="Migration successfully executed" id="create annotation table v5" duration=1.483474ms grafana | logger=migrator t=2025-06-18T15:20:28.990940325Z level=info msg="Executing migration" id="add index annotation 0 v3" grafana | logger=migrator t=2025-06-18T15:20:28.99259244Z level=info msg="Migration successfully executed" id="add index annotation 0 v3" duration=1.653205ms grafana | logger=migrator t=2025-06-18T15:20:28.996931951Z level=info msg="Executing migration" id="add index annotation 1 v3" grafana | logger=migrator t=2025-06-18T15:20:28.998075291Z level=info msg="Migration successfully executed" id="add index annotation 1 v3" duration=1.14478ms grafana | logger=migrator t=2025-06-18T15:20:29.006424328Z level=info msg="Executing migration" id="add index annotation 2 v3" grafana | logger=migrator t=2025-06-18T15:20:29.00758669Z level=info msg="Migration successfully executed" id="add index annotation 2 v3" duration=1.163571ms grafana | logger=migrator t=2025-06-18T15:20:29.015841246Z level=info msg="Executing migration" id="add index annotation 3 v3" grafana | logger=migrator t=2025-06-18T15:20:29.017232959Z level=info msg="Migration successfully executed" id="add index annotation 3 v3" duration=1.393643ms grafana | logger=migrator t=2025-06-18T15:20:29.026498675Z level=info msg="Executing migration" id="add index annotation 4 v3" grafana | logger=migrator t=2025-06-18T15:20:29.027689667Z level=info msg="Migration successfully executed" id="add index annotation 4 v3" duration=1.190412ms grafana | logger=migrator t=2025-06-18T15:20:29.083101392Z level=info msg="Executing migration" id="Update annotation table charset" grafana | logger=migrator t=2025-06-18T15:20:29.083168413Z level=info msg="Migration successfully executed" id="Update annotation table charset" duration=70.621µs grafana | logger=migrator t=2025-06-18T15:20:29.089915296Z level=info msg="Executing migration" id="Add column region_id to annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.094189876Z level=info msg="Migration successfully executed" id="Add column region_id to annotation table" duration=4.27474ms grafana | logger=migrator t=2025-06-18T15:20:29.100787908Z level=info msg="Executing migration" id="Drop category_id index" grafana | logger=migrator t=2025-06-18T15:20:29.101759496Z level=info msg="Migration successfully executed" id="Drop category_id index" duration=970.778µs grafana | logger=migrator t=2025-06-18T15:20:29.111360026Z level=info msg="Executing migration" id="Add column tags to annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.118193299Z level=info msg="Migration successfully executed" id="Add column tags to annotation table" duration=6.835143ms grafana | logger=migrator t=2025-06-18T15:20:29.123925563Z level=info msg="Executing migration" id="Create annotation_tag table v2" grafana | logger=migrator t=2025-06-18T15:20:29.124906612Z level=info msg="Migration successfully executed" id="Create annotation_tag table v2" duration=983.649µs grafana | logger=migrator t=2025-06-18T15:20:29.133255119Z level=info msg="Executing migration" id="Add unique index annotation_tag.annotation_id_tag_id" grafana | logger=migrator t=2025-06-18T15:20:29.134242819Z level=info msg="Migration successfully executed" id="Add unique index annotation_tag.annotation_id_tag_id" duration=987.13µs grafana | logger=migrator t=2025-06-18T15:20:29.140936571Z level=info msg="Executing migration" id="drop index UQE_annotation_tag_annotation_id_tag_id - v2" grafana | logger=migrator t=2025-06-18T15:20:29.142006861Z level=info msg="Migration successfully executed" id="drop index UQE_annotation_tag_annotation_id_tag_id - v2" duration=1.06994ms grafana | logger=migrator t=2025-06-18T15:20:29.14613039Z level=info msg="Executing migration" id="Rename table annotation_tag to annotation_tag_v2 - v2" grafana | logger=migrator t=2025-06-18T15:20:29.163914285Z level=info msg="Migration successfully executed" id="Rename table annotation_tag to annotation_tag_v2 - v2" duration=17.782725ms grafana | logger=migrator t=2025-06-18T15:20:29.176989587Z level=info msg="Executing migration" id="Create annotation_tag table v3" grafana | logger=migrator t=2025-06-18T15:20:29.177982656Z level=info msg="Migration successfully executed" id="Create annotation_tag table v3" duration=993.069µs grafana | logger=migrator t=2025-06-18T15:20:29.183302305Z level=info msg="Executing migration" id="create index UQE_annotation_tag_annotation_id_tag_id - Add unique index annotation_tag.annotation_id_tag_id V3" grafana | logger=migrator t=2025-06-18T15:20:29.185010512Z level=info msg="Migration successfully executed" id="create index UQE_annotation_tag_annotation_id_tag_id - Add unique index annotation_tag.annotation_id_tag_id V3" duration=1.708867ms grafana | logger=migrator t=2025-06-18T15:20:29.188347823Z level=info msg="Executing migration" id="copy annotation_tag v2 to v3" grafana | logger=migrator t=2025-06-18T15:20:29.188657736Z level=info msg="Migration successfully executed" id="copy annotation_tag v2 to v3" duration=310.303µs grafana | logger=migrator t=2025-06-18T15:20:29.208787333Z level=info msg="Executing migration" id="drop table annotation_tag_v2" grafana | logger=migrator t=2025-06-18T15:20:29.210018835Z level=info msg="Migration successfully executed" id="drop table annotation_tag_v2" duration=1.232142ms grafana | logger=migrator t=2025-06-18T15:20:29.254337848Z level=info msg="Executing migration" id="Update alert annotations and set TEXT to empty" grafana | logger=migrator t=2025-06-18T15:20:29.254758022Z level=info msg="Migration successfully executed" id="Update alert annotations and set TEXT to empty" duration=422.134µs grafana | logger=migrator t=2025-06-18T15:20:29.258406735Z level=info msg="Executing migration" id="Add created time to annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.263594204Z level=info msg="Migration successfully executed" id="Add created time to annotation table" duration=5.187619ms grafana | logger=migrator t=2025-06-18T15:20:29.266726383Z level=info msg="Executing migration" id="Add updated time to annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.271627049Z level=info msg="Migration successfully executed" id="Add updated time to annotation table" duration=4.901215ms grafana | logger=migrator t=2025-06-18T15:20:29.280280209Z level=info msg="Executing migration" id="Add index for created in annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.281971714Z level=info msg="Migration successfully executed" id="Add index for created in annotation table" duration=1.690715ms grafana | logger=migrator t=2025-06-18T15:20:29.286199424Z level=info msg="Executing migration" id="Add index for updated in annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.287674037Z level=info msg="Migration successfully executed" id="Add index for updated in annotation table" duration=1.474093ms grafana | logger=migrator t=2025-06-18T15:20:29.292716635Z level=info msg="Executing migration" id="Convert existing annotations from seconds to milliseconds" grafana | logger=migrator t=2025-06-18T15:20:29.292996367Z level=info msg="Migration successfully executed" id="Convert existing annotations from seconds to milliseconds" duration=284.273µs grafana | logger=migrator t=2025-06-18T15:20:29.303137492Z level=info msg="Executing migration" id="Add epoch_end column" grafana | logger=migrator t=2025-06-18T15:20:29.310240688Z level=info msg="Migration successfully executed" id="Add epoch_end column" duration=7.097626ms grafana | logger=migrator t=2025-06-18T15:20:29.313247776Z level=info msg="Executing migration" id="Add index for epoch_end" grafana | logger=migrator t=2025-06-18T15:20:29.314203785Z level=info msg="Migration successfully executed" id="Add index for epoch_end" duration=955.249µs grafana | logger=migrator t=2025-06-18T15:20:29.31808985Z level=info msg="Executing migration" id="Make epoch_end the same as epoch" grafana | logger=migrator t=2025-06-18T15:20:29.318278042Z level=info msg="Migration successfully executed" id="Make epoch_end the same as epoch" duration=187.322µs grafana | logger=migrator t=2025-06-18T15:20:29.323937985Z level=info msg="Executing migration" id="Move region to single row" grafana | logger=migrator t=2025-06-18T15:20:29.32432243Z level=info msg="Migration successfully executed" id="Move region to single row" duration=384.415µs grafana | logger=migrator t=2025-06-18T15:20:29.331543446Z level=info msg="Executing migration" id="Remove index org_id_epoch from annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.33294967Z level=info msg="Migration successfully executed" id="Remove index org_id_epoch from annotation table" duration=1.404584ms grafana | logger=migrator t=2025-06-18T15:20:29.338212289Z level=info msg="Executing migration" id="Remove index org_id_dashboard_id_panel_id_epoch from annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.339122657Z level=info msg="Migration successfully executed" id="Remove index org_id_dashboard_id_panel_id_epoch from annotation table" duration=909.888µs grafana | logger=migrator t=2025-06-18T15:20:29.346478805Z level=info msg="Executing migration" id="Add index for org_id_dashboard_id_epoch_end_epoch on annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.347974009Z level=info msg="Migration successfully executed" id="Add index for org_id_dashboard_id_epoch_end_epoch on annotation table" duration=1.494654ms grafana | logger=migrator t=2025-06-18T15:20:29.35773891Z level=info msg="Executing migration" id="Add index for org_id_epoch_end_epoch on annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.358688249Z level=info msg="Migration successfully executed" id="Add index for org_id_epoch_end_epoch on annotation table" duration=948.579µs grafana | logger=migrator t=2025-06-18T15:20:29.362460704Z level=info msg="Executing migration" id="Remove index org_id_epoch_epoch_end from annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.364605304Z level=info msg="Migration successfully executed" id="Remove index org_id_epoch_epoch_end from annotation table" duration=2.14435ms grafana | logger=migrator t=2025-06-18T15:20:29.372056934Z level=info msg="Executing migration" id="Add index for alert_id on annotation table" grafana | logger=migrator t=2025-06-18T15:20:29.373053183Z level=info msg="Migration successfully executed" id="Add index for alert_id on annotation table" duration=996.419µs grafana | logger=migrator t=2025-06-18T15:20:29.376632546Z level=info msg="Executing migration" id="Increase tags column to length 4096" grafana | logger=migrator t=2025-06-18T15:20:29.376653126Z level=info msg="Migration successfully executed" id="Increase tags column to length 4096" duration=23.96µs grafana | logger=migrator t=2025-06-18T15:20:29.380574563Z level=info msg="Executing migration" id="Increase prev_state column to length 40 not null" grafana | logger=migrator t=2025-06-18T15:20:29.380601863Z level=info msg="Migration successfully executed" id="Increase prev_state column to length 40 not null" duration=28.08µs grafana | logger=migrator t=2025-06-18T15:20:29.448798288Z level=info msg="Executing migration" id="Increase new_state column to length 40 not null" grafana | logger=migrator t=2025-06-18T15:20:29.448823279Z level=info msg="Migration successfully executed" id="Increase new_state column to length 40 not null" duration=27.551µs grafana | logger=migrator t=2025-06-18T15:20:29.455163777Z level=info msg="Executing migration" id="create test_data table" grafana | logger=migrator t=2025-06-18T15:20:29.456031206Z level=info msg="Migration successfully executed" id="create test_data table" duration=867.369µs grafana | logger=migrator t=2025-06-18T15:20:29.459650009Z level=info msg="Executing migration" id="create dashboard_version table v1" grafana | logger=migrator t=2025-06-18T15:20:29.460435196Z level=info msg="Migration successfully executed" id="create dashboard_version table v1" duration=784.997µs grafana | logger=migrator t=2025-06-18T15:20:29.467687814Z level=info msg="Executing migration" id="add index dashboard_version.dashboard_id" grafana | logger=migrator t=2025-06-18T15:20:29.468521022Z level=info msg="Migration successfully executed" id="add index dashboard_version.dashboard_id" duration=832.958µs grafana | logger=migrator t=2025-06-18T15:20:29.47153908Z level=info msg="Executing migration" id="add unique index dashboard_version.dashboard_id and dashboard_version.version" grafana | logger=migrator t=2025-06-18T15:20:29.472862623Z level=info msg="Migration successfully executed" id="add unique index dashboard_version.dashboard_id and dashboard_version.version" duration=1.322783ms grafana | logger=migrator t=2025-06-18T15:20:29.476779699Z level=info msg="Executing migration" id="Set dashboard version to 1 where 0" grafana | logger=migrator t=2025-06-18T15:20:29.47698173Z level=info msg="Migration successfully executed" id="Set dashboard version to 1 where 0" duration=201.331µs grafana | logger=migrator t=2025-06-18T15:20:29.482346021Z level=info msg="Executing migration" id="save existing dashboard data in dashboard_version table v1" grafana | logger=migrator t=2025-06-18T15:20:29.482905446Z level=info msg="Migration successfully executed" id="save existing dashboard data in dashboard_version table v1" duration=559.515µs grafana | logger=migrator t=2025-06-18T15:20:29.490143784Z level=info msg="Executing migration" id="alter dashboard_version.data to mediumtext v1" grafana | logger=migrator t=2025-06-18T15:20:29.490166644Z level=info msg="Migration successfully executed" id="alter dashboard_version.data to mediumtext v1" duration=24.4µs grafana | logger=migrator t=2025-06-18T15:20:29.495669745Z level=info msg="Executing migration" id="Add apiVersion for dashboard_version" grafana | logger=migrator t=2025-06-18T15:20:29.503722Z level=info msg="Migration successfully executed" id="Add apiVersion for dashboard_version" duration=8.050185ms grafana | logger=migrator t=2025-06-18T15:20:29.510364741Z level=info msg="Executing migration" id="create team table" grafana | logger=migrator t=2025-06-18T15:20:29.511078358Z level=info msg="Migration successfully executed" id="create team table" duration=711.757µs grafana | logger=migrator t=2025-06-18T15:20:29.519608897Z level=info msg="Executing migration" id="add index team.org_id" grafana | logger=migrator t=2025-06-18T15:20:29.520981311Z level=info msg="Migration successfully executed" id="add index team.org_id" duration=1.372304ms grafana | logger=migrator t=2025-06-18T15:20:29.524799386Z level=info msg="Executing migration" id="add unique index team_org_id_name" grafana | logger=migrator t=2025-06-18T15:20:29.525698884Z level=info msg="Migration successfully executed" id="add unique index team_org_id_name" duration=899.048µs grafana | logger=migrator t=2025-06-18T15:20:29.528887904Z level=info msg="Executing migration" id="Add column uid in team" grafana | logger=migrator t=2025-06-18T15:20:29.533613238Z level=info msg="Migration successfully executed" id="Add column uid in team" duration=4.723194ms grafana | logger=migrator t=2025-06-18T15:20:29.541828034Z level=info msg="Executing migration" id="Update uid column values in team" grafana | logger=migrator t=2025-06-18T15:20:29.542015096Z level=info msg="Migration successfully executed" id="Update uid column values in team" duration=189.992µs grafana | logger=migrator t=2025-06-18T15:20:29.544712141Z level=info msg="Executing migration" id="Add unique index team_org_id_uid" grafana | logger=migrator t=2025-06-18T15:20:29.545636541Z level=info msg="Migration successfully executed" id="Add unique index team_org_id_uid" duration=923.92µs grafana | logger=migrator t=2025-06-18T15:20:29.568000339Z level=info msg="Executing migration" id="Add column external_uid in team" grafana | logger=migrator t=2025-06-18T15:20:29.573389448Z level=info msg="Migration successfully executed" id="Add column external_uid in team" duration=5.392859ms grafana | logger=migrator t=2025-06-18T15:20:29.577631188Z level=info msg="Executing migration" id="Add column is_provisioned in team" grafana | logger=migrator t=2025-06-18T15:20:29.582253291Z level=info msg="Migration successfully executed" id="Add column is_provisioned in team" duration=4.621833ms grafana | logger=migrator t=2025-06-18T15:20:29.611694075Z level=info msg="Executing migration" id="create team member table" grafana | logger=migrator t=2025-06-18T15:20:29.613786925Z level=info msg="Migration successfully executed" id="create team member table" duration=2.09222ms grafana | logger=migrator t=2025-06-18T15:20:29.617363228Z level=info msg="Executing migration" id="add index team_member.org_id" grafana | logger=migrator t=2025-06-18T15:20:29.618384948Z level=info msg="Migration successfully executed" id="add index team_member.org_id" duration=1.01985ms grafana | logger=migrator t=2025-06-18T15:20:29.624923928Z level=info msg="Executing migration" id="add unique index team_member_org_id_team_id_user_id" grafana | logger=migrator t=2025-06-18T15:20:29.626472902Z level=info msg="Migration successfully executed" id="add unique index team_member_org_id_team_id_user_id" duration=1.548324ms grafana | logger=migrator t=2025-06-18T15:20:29.630111417Z level=info msg="Executing migration" id="add index team_member.team_id" grafana | logger=migrator t=2025-06-18T15:20:29.63156908Z level=info msg="Migration successfully executed" id="add index team_member.team_id" duration=1.461013ms grafana | logger=migrator t=2025-06-18T15:20:29.63586371Z level=info msg="Executing migration" id="Add column email to team table" grafana | logger=migrator t=2025-06-18T15:20:29.6411604Z level=info msg="Migration successfully executed" id="Add column email to team table" duration=5.29469ms grafana | logger=migrator t=2025-06-18T15:20:29.644748783Z level=info msg="Executing migration" id="Add column external to team_member table" grafana | logger=migrator t=2025-06-18T15:20:29.649584218Z level=info msg="Migration successfully executed" id="Add column external to team_member table" duration=4.837415ms grafana | logger=migrator t=2025-06-18T15:20:29.656441572Z level=info msg="Executing migration" id="Add column permission to team_member table" grafana | logger=migrator t=2025-06-18T15:20:29.662014504Z level=info msg="Migration successfully executed" id="Add column permission to team_member table" duration=5.571872ms grafana | logger=migrator t=2025-06-18T15:20:29.665805158Z level=info msg="Executing migration" id="add unique index team_member_user_id_org_id" grafana | logger=migrator t=2025-06-18T15:20:29.666746948Z level=info msg="Migration successfully executed" id="add unique index team_member_user_id_org_id" duration=942.04µs grafana | logger=migrator t=2025-06-18T15:20:29.670712394Z level=info msg="Executing migration" id="create dashboard acl table" grafana | logger=migrator t=2025-06-18T15:20:29.671868796Z level=info msg="Migration successfully executed" id="create dashboard acl table" duration=1.155902ms grafana | logger=migrator t=2025-06-18T15:20:29.679585297Z level=info msg="Executing migration" id="add index dashboard_acl_dashboard_id" grafana | logger=migrator t=2025-06-18T15:20:29.681311134Z level=info msg="Migration successfully executed" id="add index dashboard_acl_dashboard_id" duration=1.726007ms grafana | logger=migrator t=2025-06-18T15:20:29.685845386Z level=info msg="Executing migration" id="add unique index dashboard_acl_dashboard_id_user_id" grafana | logger=migrator t=2025-06-18T15:20:29.686882185Z level=info msg="Migration successfully executed" id="add unique index dashboard_acl_dashboard_id_user_id" duration=1.036579ms grafana | logger=migrator t=2025-06-18T15:20:29.6905934Z level=info msg="Executing migration" id="add unique index dashboard_acl_dashboard_id_team_id" grafana | logger=migrator t=2025-06-18T15:20:29.691595529Z level=info msg="Migration successfully executed" id="add unique index dashboard_acl_dashboard_id_team_id" duration=1.001629ms grafana | logger=migrator t=2025-06-18T15:20:29.698064479Z level=info msg="Executing migration" id="add index dashboard_acl_user_id" grafana | logger=migrator t=2025-06-18T15:20:29.699721865Z level=info msg="Migration successfully executed" id="add index dashboard_acl_user_id" duration=1.656476ms grafana | logger=migrator t=2025-06-18T15:20:29.703379289Z level=info msg="Executing migration" id="add index dashboard_acl_team_id" grafana | logger=migrator t=2025-06-18T15:20:29.705118206Z level=info msg="Migration successfully executed" id="add index dashboard_acl_team_id" duration=1.738196ms grafana | logger=migrator t=2025-06-18T15:20:29.718723122Z level=info msg="Executing migration" id="add index dashboard_acl_org_id_role" grafana | logger=migrator t=2025-06-18T15:20:29.720545849Z level=info msg="Migration successfully executed" id="add index dashboard_acl_org_id_role" duration=1.821688ms grafana | logger=migrator t=2025-06-18T15:20:29.726883987Z level=info msg="Executing migration" id="add index dashboard_permission" grafana | logger=migrator t=2025-06-18T15:20:29.727933278Z level=info msg="Migration successfully executed" id="add index dashboard_permission" duration=1.049291ms grafana | logger=migrator t=2025-06-18T15:20:29.731776483Z level=info msg="Executing migration" id="save default acl rules in dashboard_acl table" grafana | logger=migrator t=2025-06-18T15:20:29.732659671Z level=info msg="Migration successfully executed" id="save default acl rules in dashboard_acl table" duration=882.248µs grafana | logger=migrator t=2025-06-18T15:20:29.738623437Z level=info msg="Executing migration" id="delete acl rules for deleted dashboards and folders" grafana | logger=migrator t=2025-06-18T15:20:29.739077122Z level=info msg="Migration successfully executed" id="delete acl rules for deleted dashboards and folders" duration=453.885µs grafana | logger=migrator t=2025-06-18T15:20:29.744235169Z level=info msg="Executing migration" id="create tag table" grafana | logger=migrator t=2025-06-18T15:20:29.74539561Z level=info msg="Migration successfully executed" id="create tag table" duration=1.160351ms grafana | logger=migrator t=2025-06-18T15:20:29.750000304Z level=info msg="Executing migration" id="add index tag.key_value" grafana | logger=migrator t=2025-06-18T15:20:29.750950702Z level=info msg="Migration successfully executed" id="add index tag.key_value" duration=950.318µs grafana | logger=migrator t=2025-06-18T15:20:29.772128439Z level=info msg="Executing migration" id="create login attempt table" grafana | logger=migrator t=2025-06-18T15:20:29.772990828Z level=info msg="Migration successfully executed" id="create login attempt table" duration=863.369µs grafana | logger=migrator t=2025-06-18T15:20:29.777003795Z level=info msg="Executing migration" id="add index login_attempt.username" grafana | logger=migrator t=2025-06-18T15:20:29.778289177Z level=info msg="Migration successfully executed" id="add index login_attempt.username" duration=1.284092ms grafana | logger=migrator t=2025-06-18T15:20:29.78399824Z level=info msg="Executing migration" id="drop index IDX_login_attempt_username - v1" grafana | logger=migrator t=2025-06-18T15:20:29.785472513Z level=info msg="Migration successfully executed" id="drop index IDX_login_attempt_username - v1" duration=1.473133ms grafana | logger=migrator t=2025-06-18T15:20:29.789760154Z level=info msg="Executing migration" id="Rename table login_attempt to login_attempt_tmp_qwerty - v1" grafana | logger=migrator t=2025-06-18T15:20:29.807332607Z level=info msg="Migration successfully executed" id="Rename table login_attempt to login_attempt_tmp_qwerty - v1" duration=17.569503ms grafana | logger=migrator t=2025-06-18T15:20:29.811774409Z level=info msg="Executing migration" id="create login_attempt v2" grafana | logger=migrator t=2025-06-18T15:20:29.812364124Z level=info msg="Migration successfully executed" id="create login_attempt v2" duration=589.815µs grafana | logger=migrator t=2025-06-18T15:20:29.815239461Z level=info msg="Executing migration" id="create index IDX_login_attempt_username - v2" grafana | logger=migrator t=2025-06-18T15:20:29.815936067Z level=info msg="Migration successfully executed" id="create index IDX_login_attempt_username - v2" duration=696.506µs grafana | logger=migrator t=2025-06-18T15:20:29.820079176Z level=info msg="Executing migration" id="copy login_attempt v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:29.82055859Z level=info msg="Migration successfully executed" id="copy login_attempt v1 to v2" duration=481.304µs grafana | logger=migrator t=2025-06-18T15:20:29.823930532Z level=info msg="Executing migration" id="drop login_attempt_tmp_qwerty" grafana | logger=migrator t=2025-06-18T15:20:29.824637798Z level=info msg="Migration successfully executed" id="drop login_attempt_tmp_qwerty" duration=706.486µs grafana | logger=migrator t=2025-06-18T15:20:29.829625414Z level=info msg="Executing migration" id="create user auth table" grafana | logger=migrator t=2025-06-18T15:20:29.830430142Z level=info msg="Migration successfully executed" id="create user auth table" duration=803.918µs grafana | logger=migrator t=2025-06-18T15:20:29.836950463Z level=info msg="Executing migration" id="create index IDX_user_auth_auth_module_auth_id - v1" grafana | logger=migrator t=2025-06-18T15:20:29.837976043Z level=info msg="Migration successfully executed" id="create index IDX_user_auth_auth_module_auth_id - v1" duration=1.02522ms grafana | logger=migrator t=2025-06-18T15:20:29.842901879Z level=info msg="Executing migration" id="alter user_auth.auth_id to length 190" grafana | logger=migrator t=2025-06-18T15:20:29.842933029Z level=info msg="Migration successfully executed" id="alter user_auth.auth_id to length 190" duration=32.4µs grafana | logger=migrator t=2025-06-18T15:20:29.846501322Z level=info msg="Executing migration" id="Add OAuth access token to user_auth" grafana | logger=migrator t=2025-06-18T15:20:29.85384532Z level=info msg="Migration successfully executed" id="Add OAuth access token to user_auth" duration=7.344398ms grafana | logger=migrator t=2025-06-18T15:20:29.870555856Z level=info msg="Executing migration" id="Add OAuth refresh token to user_auth" grafana | logger=migrator t=2025-06-18T15:20:29.880057194Z level=info msg="Migration successfully executed" id="Add OAuth refresh token to user_auth" duration=9.502158ms grafana | logger=migrator t=2025-06-18T15:20:29.884009781Z level=info msg="Executing migration" id="Add OAuth token type to user_auth" grafana | logger=migrator t=2025-06-18T15:20:29.888034039Z level=info msg="Migration successfully executed" id="Add OAuth token type to user_auth" duration=4.023378ms grafana | logger=migrator t=2025-06-18T15:20:29.892289588Z level=info msg="Executing migration" id="Add OAuth expiry to user_auth" grafana | logger=migrator t=2025-06-18T15:20:29.898070792Z level=info msg="Migration successfully executed" id="Add OAuth expiry to user_auth" duration=5.778564ms grafana | logger=migrator t=2025-06-18T15:20:29.903004038Z level=info msg="Executing migration" id="Add index to user_id column in user_auth" grafana | logger=migrator t=2025-06-18T15:20:29.904113439Z level=info msg="Migration successfully executed" id="Add index to user_id column in user_auth" duration=1.109441ms grafana | logger=migrator t=2025-06-18T15:20:29.907368619Z level=info msg="Executing migration" id="Add OAuth ID token to user_auth" grafana | logger=migrator t=2025-06-18T15:20:29.913277184Z level=info msg="Migration successfully executed" id="Add OAuth ID token to user_auth" duration=5.907555ms grafana | logger=migrator t=2025-06-18T15:20:29.938971703Z level=info msg="Executing migration" id="Add user_unique_id to user_auth" grafana | logger=migrator t=2025-06-18T15:20:29.948112208Z level=info msg="Migration successfully executed" id="Add user_unique_id to user_auth" duration=9.141765ms grafana | logger=migrator t=2025-06-18T15:20:29.951746082Z level=info msg="Executing migration" id="create server_lock table" grafana | logger=migrator t=2025-06-18T15:20:29.953065884Z level=info msg="Migration successfully executed" id="create server_lock table" duration=1.321452ms grafana | logger=migrator t=2025-06-18T15:20:29.957696877Z level=info msg="Executing migration" id="add index server_lock.operation_uid" grafana | logger=migrator t=2025-06-18T15:20:29.958906698Z level=info msg="Migration successfully executed" id="add index server_lock.operation_uid" duration=1.208941ms grafana | logger=migrator t=2025-06-18T15:20:29.965989965Z level=info msg="Executing migration" id="create user auth token table" grafana | logger=migrator t=2025-06-18T15:20:29.967420938Z level=info msg="Migration successfully executed" id="create user auth token table" duration=1.430963ms grafana | logger=migrator t=2025-06-18T15:20:29.9719622Z level=info msg="Executing migration" id="add unique index user_auth_token.auth_token" grafana | logger=migrator t=2025-06-18T15:20:29.973601885Z level=info msg="Migration successfully executed" id="add unique index user_auth_token.auth_token" duration=1.635965ms grafana | logger=migrator t=2025-06-18T15:20:29.9772857Z level=info msg="Executing migration" id="add unique index user_auth_token.prev_auth_token" grafana | logger=migrator t=2025-06-18T15:20:29.978269969Z level=info msg="Migration successfully executed" id="add unique index user_auth_token.prev_auth_token" duration=983.929µs grafana | logger=migrator t=2025-06-18T15:20:29.983148684Z level=info msg="Executing migration" id="add index user_auth_token.user_id" grafana | logger=migrator t=2025-06-18T15:20:29.984122713Z level=info msg="Migration successfully executed" id="add index user_auth_token.user_id" duration=973.579µs grafana | logger=migrator t=2025-06-18T15:20:29.990320791Z level=info msg="Executing migration" id="Add revoked_at to the user auth token" grafana | logger=migrator t=2025-06-18T15:20:29.997297506Z level=info msg="Migration successfully executed" id="Add revoked_at to the user auth token" duration=6.975885ms grafana | logger=migrator t=2025-06-18T15:20:30.001183412Z level=info msg="Executing migration" id="add index user_auth_token.revoked_at" grafana | logger=migrator t=2025-06-18T15:20:30.001843688Z level=info msg="Migration successfully executed" id="add index user_auth_token.revoked_at" duration=660.016µs grafana | logger=migrator t=2025-06-18T15:20:30.00741837Z level=info msg="Executing migration" id="add external_session_id to user_auth_token" grafana | logger=migrator t=2025-06-18T15:20:30.013330987Z level=info msg="Migration successfully executed" id="add external_session_id to user_auth_token" duration=5.912327ms grafana | logger=migrator t=2025-06-18T15:20:30.030945004Z level=info msg="Executing migration" id="create cache_data table" grafana | logger=migrator t=2025-06-18T15:20:30.031847792Z level=info msg="Migration successfully executed" id="create cache_data table" duration=902.588µs grafana | logger=migrator t=2025-06-18T15:20:30.040918308Z level=info msg="Executing migration" id="add unique index cache_data.cache_key" grafana | logger=migrator t=2025-06-18T15:20:30.041969639Z level=info msg="Migration successfully executed" id="add unique index cache_data.cache_key" duration=1.048671ms grafana | logger=migrator t=2025-06-18T15:20:30.050315649Z level=info msg="Executing migration" id="create short_url table v1" grafana | logger=migrator t=2025-06-18T15:20:30.0515988Z level=info msg="Migration successfully executed" id="create short_url table v1" duration=1.282971ms grafana | logger=migrator t=2025-06-18T15:20:30.057421115Z level=info msg="Executing migration" id="add index short_url.org_id-uid" grafana | logger=migrator t=2025-06-18T15:20:30.05893761Z level=info msg="Migration successfully executed" id="add index short_url.org_id-uid" duration=1.512865ms grafana | logger=migrator t=2025-06-18T15:20:30.064678195Z level=info msg="Executing migration" id="alter table short_url alter column created_by type to bigint" grafana | logger=migrator t=2025-06-18T15:20:30.064701995Z level=info msg="Migration successfully executed" id="alter table short_url alter column created_by type to bigint" duration=25.45µs grafana | logger=migrator t=2025-06-18T15:20:30.072154587Z level=info msg="Executing migration" id="delete alert_definition table" grafana | logger=migrator t=2025-06-18T15:20:30.072289908Z level=info msg="Migration successfully executed" id="delete alert_definition table" duration=135.681µs grafana | logger=migrator t=2025-06-18T15:20:30.075476648Z level=info msg="Executing migration" id="recreate alert_definition table" grafana | logger=migrator t=2025-06-18T15:20:30.076838701Z level=info msg="Migration successfully executed" id="recreate alert_definition table" duration=1.354133ms grafana | logger=migrator t=2025-06-18T15:20:30.082201482Z level=info msg="Executing migration" id="add index in alert_definition on org_id and title columns" grafana | logger=migrator t=2025-06-18T15:20:30.083214271Z level=info msg="Migration successfully executed" id="add index in alert_definition on org_id and title columns" duration=1.011399ms grafana | logger=migrator t=2025-06-18T15:20:30.098919951Z level=info msg="Executing migration" id="add index in alert_definition on org_id and uid columns" grafana | logger=migrator t=2025-06-18T15:20:30.100448746Z level=info msg="Migration successfully executed" id="add index in alert_definition on org_id and uid columns" duration=1.528175ms grafana | logger=migrator t=2025-06-18T15:20:30.114196757Z level=info msg="Executing migration" id="alter alert_definition table data column to mediumtext in mysql" grafana | logger=migrator t=2025-06-18T15:20:30.114224657Z level=info msg="Migration successfully executed" id="alter alert_definition table data column to mediumtext in mysql" duration=29.32µs grafana | logger=migrator t=2025-06-18T15:20:30.117898512Z level=info msg="Executing migration" id="drop index in alert_definition on org_id and title columns" grafana | logger=migrator t=2025-06-18T15:20:30.11984458Z level=info msg="Migration successfully executed" id="drop index in alert_definition on org_id and title columns" duration=1.944898ms grafana | logger=migrator t=2025-06-18T15:20:30.123639606Z level=info msg="Executing migration" id="drop index in alert_definition on org_id and uid columns" grafana | logger=migrator t=2025-06-18T15:20:30.124632175Z level=info msg="Migration successfully executed" id="drop index in alert_definition on org_id and uid columns" duration=992.989µs grafana | logger=migrator t=2025-06-18T15:20:30.129016688Z level=info msg="Executing migration" id="add unique index in alert_definition on org_id and title columns" grafana | logger=migrator t=2025-06-18T15:20:30.130081727Z level=info msg="Migration successfully executed" id="add unique index in alert_definition on org_id and title columns" duration=1.064749ms grafana | logger=migrator t=2025-06-18T15:20:30.133358369Z level=info msg="Executing migration" id="add unique index in alert_definition on org_id and uid columns" grafana | logger=migrator t=2025-06-18T15:20:30.134365899Z level=info msg="Migration successfully executed" id="add unique index in alert_definition on org_id and uid columns" duration=1.0068ms grafana | logger=migrator t=2025-06-18T15:20:30.139317275Z level=info msg="Executing migration" id="Add column paused in alert_definition" grafana | logger=migrator t=2025-06-18T15:20:30.146861617Z level=info msg="Migration successfully executed" id="Add column paused in alert_definition" duration=7.543332ms grafana | logger=migrator t=2025-06-18T15:20:30.150629944Z level=info msg="Executing migration" id="drop alert_definition table" grafana | logger=migrator t=2025-06-18T15:20:30.151639183Z level=info msg="Migration successfully executed" id="drop alert_definition table" duration=1.008669ms grafana | logger=migrator t=2025-06-18T15:20:30.154805463Z level=info msg="Executing migration" id="delete alert_definition_version table" grafana | logger=migrator t=2025-06-18T15:20:30.154888454Z level=info msg="Migration successfully executed" id="delete alert_definition_version table" duration=83.231µs grafana | logger=migrator t=2025-06-18T15:20:30.158880501Z level=info msg="Executing migration" id="recreate alert_definition_version table" grafana | logger=migrator t=2025-06-18T15:20:30.159819341Z level=info msg="Migration successfully executed" id="recreate alert_definition_version table" duration=938.66µs grafana | logger=migrator t=2025-06-18T15:20:30.164355134Z level=info msg="Executing migration" id="add index in alert_definition_version table on alert_definition_id and version columns" grafana | logger=migrator t=2025-06-18T15:20:30.165456825Z level=info msg="Migration successfully executed" id="add index in alert_definition_version table on alert_definition_id and version columns" duration=1.07643ms grafana | logger=migrator t=2025-06-18T15:20:30.183218793Z level=info msg="Executing migration" id="add index in alert_definition_version table on alert_definition_uid and version columns" grafana | logger=migrator t=2025-06-18T15:20:30.186276062Z level=info msg="Migration successfully executed" id="add index in alert_definition_version table on alert_definition_uid and version columns" duration=3.057459ms grafana | logger=migrator t=2025-06-18T15:20:30.191172499Z level=info msg="Executing migration" id="alter alert_definition_version table data column to mediumtext in mysql" grafana | logger=migrator t=2025-06-18T15:20:30.191195519Z level=info msg="Migration successfully executed" id="alter alert_definition_version table data column to mediumtext in mysql" duration=24.23µs grafana | logger=migrator t=2025-06-18T15:20:30.194218738Z level=info msg="Executing migration" id="drop alert_definition_version table" grafana | logger=migrator t=2025-06-18T15:20:30.19540183Z level=info msg="Migration successfully executed" id="drop alert_definition_version table" duration=1.179572ms grafana | logger=migrator t=2025-06-18T15:20:30.199762241Z level=info msg="Executing migration" id="create alert_instance table" grafana | logger=migrator t=2025-06-18T15:20:30.201276936Z level=info msg="Migration successfully executed" id="create alert_instance table" duration=1.513975ms grafana | logger=migrator t=2025-06-18T15:20:30.204594677Z level=info msg="Executing migration" id="add index in alert_instance table on def_org_id, def_uid and current_state columns" grafana | logger=migrator t=2025-06-18T15:20:30.205577856Z level=info msg="Migration successfully executed" id="add index in alert_instance table on def_org_id, def_uid and current_state columns" duration=982.629µs grafana | logger=migrator t=2025-06-18T15:20:30.209073539Z level=info msg="Executing migration" id="add index in alert_instance table on def_org_id, current_state columns" grafana | logger=migrator t=2025-06-18T15:20:30.210042139Z level=info msg="Migration successfully executed" id="add index in alert_instance table on def_org_id, current_state columns" duration=964.7µs grafana | logger=migrator t=2025-06-18T15:20:30.215684553Z level=info msg="Executing migration" id="add column current_state_end to alert_instance" grafana | logger=migrator t=2025-06-18T15:20:30.221599419Z level=info msg="Migration successfully executed" id="add column current_state_end to alert_instance" duration=5.914296ms grafana | logger=migrator t=2025-06-18T15:20:30.225250874Z level=info msg="Executing migration" id="remove index def_org_id, def_uid, current_state on alert_instance" grafana | logger=migrator t=2025-06-18T15:20:30.226240024Z level=info msg="Migration successfully executed" id="remove index def_org_id, def_uid, current_state on alert_instance" duration=989µs grafana | logger=migrator t=2025-06-18T15:20:30.23123226Z level=info msg="Executing migration" id="remove index def_org_id, current_state on alert_instance" grafana | logger=migrator t=2025-06-18T15:20:30.232970287Z level=info msg="Migration successfully executed" id="remove index def_org_id, current_state on alert_instance" duration=1.737617ms grafana | logger=migrator t=2025-06-18T15:20:30.238964834Z level=info msg="Executing migration" id="rename def_org_id to rule_org_id in alert_instance" grafana | logger=migrator t=2025-06-18T15:20:30.266372315Z level=info msg="Migration successfully executed" id="rename def_org_id to rule_org_id in alert_instance" duration=27.354991ms grafana | logger=migrator t=2025-06-18T15:20:30.277159098Z level=info msg="Executing migration" id="rename def_uid to rule_uid in alert_instance" grafana | logger=migrator t=2025-06-18T15:20:30.30782022Z level=info msg="Migration successfully executed" id="rename def_uid to rule_uid in alert_instance" duration=30.661632ms grafana | logger=migrator t=2025-06-18T15:20:30.311667006Z level=info msg="Executing migration" id="add index rule_org_id, rule_uid, current_state on alert_instance" grafana | logger=migrator t=2025-06-18T15:20:30.312491554Z level=info msg="Migration successfully executed" id="add index rule_org_id, rule_uid, current_state on alert_instance" duration=824.858µs grafana | logger=migrator t=2025-06-18T15:20:30.326626829Z level=info msg="Executing migration" id="add index rule_org_id, current_state on alert_instance" grafana | logger=migrator t=2025-06-18T15:20:30.328336585Z level=info msg="Migration successfully executed" id="add index rule_org_id, current_state on alert_instance" duration=1.709556ms grafana | logger=migrator t=2025-06-18T15:20:30.332868479Z level=info msg="Executing migration" id="add current_reason column related to current_state" grafana | logger=migrator t=2025-06-18T15:20:30.339827764Z level=info msg="Migration successfully executed" id="add current_reason column related to current_state" duration=6.958206ms grafana | logger=migrator t=2025-06-18T15:20:30.343498349Z level=info msg="Executing migration" id="add result_fingerprint column to alert_instance" grafana | logger=migrator t=2025-06-18T15:20:30.349695339Z level=info msg="Migration successfully executed" id="add result_fingerprint column to alert_instance" duration=6.19575ms grafana | logger=migrator t=2025-06-18T15:20:30.356547144Z level=info msg="Executing migration" id="create alert_rule table" grafana | logger=migrator t=2025-06-18T15:20:30.357664334Z level=info msg="Migration successfully executed" id="create alert_rule table" duration=1.11703ms grafana | logger=migrator t=2025-06-18T15:20:30.361188498Z level=info msg="Executing migration" id="add index in alert_rule on org_id and title columns" grafana | logger=migrator t=2025-06-18T15:20:30.362274578Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id and title columns" duration=1.08593ms grafana | logger=migrator t=2025-06-18T15:20:30.368345496Z level=info msg="Executing migration" id="add index in alert_rule on org_id and uid columns" grafana | logger=migrator t=2025-06-18T15:20:30.369475587Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id and uid columns" duration=1.129951ms grafana | logger=migrator t=2025-06-18T15:20:30.373629106Z level=info msg="Executing migration" id="add index in alert_rule on org_id, namespace_uid, group_uid columns" grafana | logger=migrator t=2025-06-18T15:20:30.374704507Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, namespace_uid, group_uid columns" duration=1.075141ms grafana | logger=migrator t=2025-06-18T15:20:30.377949798Z level=info msg="Executing migration" id="alter alert_rule table data column to mediumtext in mysql" grafana | logger=migrator t=2025-06-18T15:20:30.377969538Z level=info msg="Migration successfully executed" id="alter alert_rule table data column to mediumtext in mysql" duration=20.29µs grafana | logger=migrator t=2025-06-18T15:20:30.382750893Z level=info msg="Executing migration" id="add column for to alert_rule" grafana | logger=migrator t=2025-06-18T15:20:30.389677599Z level=info msg="Migration successfully executed" id="add column for to alert_rule" duration=6.925666ms grafana | logger=migrator t=2025-06-18T15:20:30.393765378Z level=info msg="Executing migration" id="add column annotations to alert_rule" grafana | logger=migrator t=2025-06-18T15:20:30.40022173Z level=info msg="Migration successfully executed" id="add column annotations to alert_rule" duration=6.456012ms grafana | logger=migrator t=2025-06-18T15:20:30.403539961Z level=info msg="Executing migration" id="add column labels to alert_rule" grafana | logger=migrator t=2025-06-18T15:20:30.410551378Z level=info msg="Migration successfully executed" id="add column labels to alert_rule" duration=7.010667ms grafana | logger=migrator t=2025-06-18T15:20:30.428384968Z level=info msg="Executing migration" id="remove unique index from alert_rule on org_id, title columns" grafana | logger=migrator t=2025-06-18T15:20:30.430050913Z level=info msg="Migration successfully executed" id="remove unique index from alert_rule on org_id, title columns" duration=1.668415ms grafana | logger=migrator t=2025-06-18T15:20:30.437711396Z level=info msg="Executing migration" id="add index in alert_rule on org_id, namespase_uid and title columns" grafana | logger=migrator t=2025-06-18T15:20:30.439301522Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, namespase_uid and title columns" duration=1.588316ms grafana | logger=migrator t=2025-06-18T15:20:30.443396891Z level=info msg="Executing migration" id="add dashboard_uid column to alert_rule" grafana | logger=migrator t=2025-06-18T15:20:30.452607698Z level=info msg="Migration successfully executed" id="add dashboard_uid column to alert_rule" duration=9.213487ms grafana | logger=migrator t=2025-06-18T15:20:30.457162221Z level=info msg="Executing migration" id="add panel_id column to alert_rule" grafana | logger=migrator t=2025-06-18T15:20:30.463854645Z level=info msg="Migration successfully executed" id="add panel_id column to alert_rule" duration=6.692124ms grafana | logger=migrator t=2025-06-18T15:20:30.467179617Z level=info msg="Executing migration" id="add index in alert_rule on org_id, dashboard_uid and panel_id columns" grafana | logger=migrator t=2025-06-18T15:20:30.468200267Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, dashboard_uid and panel_id columns" duration=1.02015ms grafana | logger=migrator t=2025-06-18T15:20:30.486297728Z level=info msg="Executing migration" id="add rule_group_idx column to alert_rule" grafana | logger=migrator t=2025-06-18T15:20:30.494574067Z level=info msg="Migration successfully executed" id="add rule_group_idx column to alert_rule" duration=8.276509ms grafana | logger=migrator t=2025-06-18T15:20:30.499941969Z level=info msg="Executing migration" id="add is_paused column to alert_rule table" grafana | logger=migrator t=2025-06-18T15:20:30.506135488Z level=info msg="Migration successfully executed" id="add is_paused column to alert_rule table" duration=6.192899ms grafana | logger=migrator t=2025-06-18T15:20:30.519046561Z level=info msg="Executing migration" id="fix is_paused column for alert_rule table" grafana | logger=migrator t=2025-06-18T15:20:30.519076481Z level=info msg="Migration successfully executed" id="fix is_paused column for alert_rule table" duration=31.35µs grafana | logger=migrator t=2025-06-18T15:20:30.524727115Z level=info msg="Executing migration" id="create alert_rule_version table" grafana | logger=migrator t=2025-06-18T15:20:30.526442461Z level=info msg="Migration successfully executed" id="create alert_rule_version table" duration=1.715256ms grafana | logger=migrator t=2025-06-18T15:20:30.529934065Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_org_id, rule_uid and version columns" grafana | logger=migrator t=2025-06-18T15:20:30.530988034Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_org_id, rule_uid and version columns" duration=1.053649ms grafana | logger=migrator t=2025-06-18T15:20:30.535538838Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_org_id, rule_namespace_uid and rule_group columns" grafana | logger=migrator t=2025-06-18T15:20:30.537359366Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_org_id, rule_namespace_uid and rule_group columns" duration=1.820278ms grafana | logger=migrator t=2025-06-18T15:20:30.541961508Z level=info msg="Executing migration" id="alter alert_rule_version table data column to mediumtext in mysql" grafana | logger=migrator t=2025-06-18T15:20:30.541989969Z level=info msg="Migration successfully executed" id="alter alert_rule_version table data column to mediumtext in mysql" duration=29.801µs grafana | logger=migrator t=2025-06-18T15:20:30.547088197Z level=info msg="Executing migration" id="add column for to alert_rule_version" grafana | logger=migrator t=2025-06-18T15:20:30.554034514Z level=info msg="Migration successfully executed" id="add column for to alert_rule_version" duration=6.945827ms grafana | logger=migrator t=2025-06-18T15:20:30.558793719Z level=info msg="Executing migration" id="add column annotations to alert_rule_version" grafana | logger=migrator t=2025-06-18T15:20:30.565301512Z level=info msg="Migration successfully executed" id="add column annotations to alert_rule_version" duration=6.507013ms grafana | logger=migrator t=2025-06-18T15:20:30.568945566Z level=info msg="Executing migration" id="add column labels to alert_rule_version" grafana | logger=migrator t=2025-06-18T15:20:30.575479858Z level=info msg="Migration successfully executed" id="add column labels to alert_rule_version" duration=6.533382ms grafana | logger=migrator t=2025-06-18T15:20:30.593248848Z level=info msg="Executing migration" id="add rule_group_idx column to alert_rule_version" grafana | logger=migrator t=2025-06-18T15:20:30.603455594Z level=info msg="Migration successfully executed" id="add rule_group_idx column to alert_rule_version" duration=10.209406ms grafana | logger=migrator t=2025-06-18T15:20:30.607793676Z level=info msg="Executing migration" id="add is_paused column to alert_rule_versions table" grafana | logger=migrator t=2025-06-18T15:20:30.612367859Z level=info msg="Migration successfully executed" id="add is_paused column to alert_rule_versions table" duration=4.574263ms grafana | logger=migrator t=2025-06-18T15:20:30.61558682Z level=info msg="Executing migration" id="fix is_paused column for alert_rule_version table" grafana | logger=migrator t=2025-06-18T15:20:30.61560537Z level=info msg="Migration successfully executed" id="fix is_paused column for alert_rule_version table" duration=19.28µs grafana | logger=migrator t=2025-06-18T15:20:30.619141434Z level=info msg="Executing migration" id=create_alert_configuration_table grafana | logger=migrator t=2025-06-18T15:20:30.619939341Z level=info msg="Migration successfully executed" id=create_alert_configuration_table duration=795.957µs grafana | logger=migrator t=2025-06-18T15:20:30.639170634Z level=info msg="Executing migration" id="Add column default in alert_configuration" grafana | logger=migrator t=2025-06-18T15:20:30.649003158Z level=info msg="Migration successfully executed" id="Add column default in alert_configuration" duration=9.831994ms grafana | logger=migrator t=2025-06-18T15:20:30.653077257Z level=info msg="Executing migration" id="alert alert_configuration alertmanager_configuration column from TEXT to MEDIUMTEXT if mysql" grafana | logger=migrator t=2025-06-18T15:20:30.653094587Z level=info msg="Migration successfully executed" id="alert alert_configuration alertmanager_configuration column from TEXT to MEDIUMTEXT if mysql" duration=18.12µs grafana | logger=migrator t=2025-06-18T15:20:30.656228787Z level=info msg="Executing migration" id="add column org_id in alert_configuration" grafana | logger=migrator t=2025-06-18T15:20:30.663000221Z level=info msg="Migration successfully executed" id="add column org_id in alert_configuration" duration=6.770514ms grafana | logger=migrator t=2025-06-18T15:20:30.667289812Z level=info msg="Executing migration" id="add index in alert_configuration table on org_id column" grafana | logger=migrator t=2025-06-18T15:20:30.66818098Z level=info msg="Migration successfully executed" id="add index in alert_configuration table on org_id column" duration=890.938µs grafana | logger=migrator t=2025-06-18T15:20:30.671452661Z level=info msg="Executing migration" id="add configuration_hash column to alert_configuration" grafana | logger=migrator t=2025-06-18T15:20:30.677900143Z level=info msg="Migration successfully executed" id="add configuration_hash column to alert_configuration" duration=6.446182ms grafana | logger=migrator t=2025-06-18T15:20:30.683046413Z level=info msg="Executing migration" id=create_ngalert_configuration_table grafana | logger=migrator t=2025-06-18T15:20:30.683942281Z level=info msg="Migration successfully executed" id=create_ngalert_configuration_table duration=895.598µs grafana | logger=migrator t=2025-06-18T15:20:30.687808787Z level=info msg="Executing migration" id="add index in ngalert_configuration on org_id column" grafana | logger=migrator t=2025-06-18T15:20:30.6891056Z level=info msg="Migration successfully executed" id="add index in ngalert_configuration on org_id column" duration=1.296503ms grafana | logger=migrator t=2025-06-18T15:20:30.694997126Z level=info msg="Executing migration" id="add column send_alerts_to in ngalert_configuration" grafana | logger=migrator t=2025-06-18T15:20:30.702451177Z level=info msg="Migration successfully executed" id="add column send_alerts_to in ngalert_configuration" duration=7.452921ms grafana | logger=migrator t=2025-06-18T15:20:30.705723258Z level=info msg="Executing migration" id="create provenance_type table" grafana | logger=migrator t=2025-06-18T15:20:30.706631977Z level=info msg="Migration successfully executed" id="create provenance_type table" duration=908.159µs grafana | logger=migrator t=2025-06-18T15:20:30.709914758Z level=info msg="Executing migration" id="add index to uniquify (record_key, record_type, org_id) columns" grafana | logger=migrator t=2025-06-18T15:20:30.711061479Z level=info msg="Migration successfully executed" id="add index to uniquify (record_key, record_type, org_id) columns" duration=1.146351ms grafana | logger=migrator t=2025-06-18T15:20:30.714403991Z level=info msg="Executing migration" id="create alert_image table" grafana | logger=migrator t=2025-06-18T15:20:30.715245349Z level=info msg="Migration successfully executed" id="create alert_image table" duration=840.998µs grafana | logger=migrator t=2025-06-18T15:20:30.721235446Z level=info msg="Executing migration" id="add unique index on token to alert_image table" grafana | logger=migrator t=2025-06-18T15:20:30.723101624Z level=info msg="Migration successfully executed" id="add unique index on token to alert_image table" duration=1.865798ms grafana | logger=migrator t=2025-06-18T15:20:30.727602476Z level=info msg="Executing migration" id="support longer URLs in alert_image table" grafana | logger=migrator t=2025-06-18T15:20:30.727662987Z level=info msg="Migration successfully executed" id="support longer URLs in alert_image table" duration=36.091µs grafana | logger=migrator t=2025-06-18T15:20:30.761003094Z level=info msg="Executing migration" id=create_alert_configuration_history_table grafana | logger=migrator t=2025-06-18T15:20:30.76265953Z level=info msg="Migration successfully executed" id=create_alert_configuration_history_table duration=1.656176ms grafana | logger=migrator t=2025-06-18T15:20:30.77001262Z level=info msg="Executing migration" id="drop non-unique orgID index on alert_configuration" grafana | logger=migrator t=2025-06-18T15:20:30.77109244Z level=info msg="Migration successfully executed" id="drop non-unique orgID index on alert_configuration" duration=1.07657ms grafana | logger=migrator t=2025-06-18T15:20:30.787454486Z level=info msg="Executing migration" id="drop unique orgID index on alert_configuration if exists" grafana | logger=migrator t=2025-06-18T15:20:30.788081382Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop unique orgID index on alert_configuration if exists" grafana | logger=migrator t=2025-06-18T15:20:30.793963408Z level=info msg="Executing migration" id="extract alertmanager configuration history to separate table" grafana | logger=migrator t=2025-06-18T15:20:30.794658194Z level=info msg="Migration successfully executed" id="extract alertmanager configuration history to separate table" duration=695.446µs grafana | logger=migrator t=2025-06-18T15:20:30.799515301Z level=info msg="Executing migration" id="add unique index on orgID to alert_configuration" grafana | logger=migrator t=2025-06-18T15:20:30.800696953Z level=info msg="Migration successfully executed" id="add unique index on orgID to alert_configuration" duration=1.181612ms grafana | logger=migrator t=2025-06-18T15:20:30.804456168Z level=info msg="Executing migration" id="add last_applied column to alert_configuration_history" grafana | logger=migrator t=2025-06-18T15:20:30.811867729Z level=info msg="Migration successfully executed" id="add last_applied column to alert_configuration_history" duration=7.410871ms grafana | logger=migrator t=2025-06-18T15:20:30.815451263Z level=info msg="Executing migration" id="create library_element table v1" grafana | logger=migrator t=2025-06-18T15:20:30.816528133Z level=info msg="Migration successfully executed" id="create library_element table v1" duration=1.075749ms grafana | logger=migrator t=2025-06-18T15:20:30.821823103Z level=info msg="Executing migration" id="add index library_element org_id-folder_id-name-kind" grafana | logger=migrator t=2025-06-18T15:20:30.823256587Z level=info msg="Migration successfully executed" id="add index library_element org_id-folder_id-name-kind" duration=1.432954ms grafana | logger=migrator t=2025-06-18T15:20:30.827334676Z level=info msg="Executing migration" id="create library_element_connection table v1" grafana | logger=migrator t=2025-06-18T15:20:30.828225644Z level=info msg="Migration successfully executed" id="create library_element_connection table v1" duration=891.008µs grafana | logger=migrator t=2025-06-18T15:20:30.833106401Z level=info msg="Executing migration" id="add index library_element_connection element_id-kind-connection_id" grafana | logger=migrator t=2025-06-18T15:20:30.834264122Z level=info msg="Migration successfully executed" id="add index library_element_connection element_id-kind-connection_id" duration=1.157591ms grafana | logger=migrator t=2025-06-18T15:20:30.839040217Z level=info msg="Executing migration" id="add unique index library_element org_id_uid" grafana | logger=migrator t=2025-06-18T15:20:30.840087408Z level=info msg="Migration successfully executed" id="add unique index library_element org_id_uid" duration=1.048121ms grafana | logger=migrator t=2025-06-18T15:20:30.843302428Z level=info msg="Executing migration" id="increase max description length to 2048" grafana | logger=migrator t=2025-06-18T15:20:30.843331148Z level=info msg="Migration successfully executed" id="increase max description length to 2048" duration=26.78µs grafana | logger=migrator t=2025-06-18T15:20:30.848639859Z level=info msg="Executing migration" id="alter library_element model to mediumtext" grafana | logger=migrator t=2025-06-18T15:20:30.848660519Z level=info msg="Migration successfully executed" id="alter library_element model to mediumtext" duration=21.5µs grafana | logger=migrator t=2025-06-18T15:20:30.853394543Z level=info msg="Executing migration" id="add library_element folder uid" grafana | logger=migrator t=2025-06-18T15:20:30.863113257Z level=info msg="Migration successfully executed" id="add library_element folder uid" duration=9.718944ms grafana | logger=migrator t=2025-06-18T15:20:30.866574249Z level=info msg="Executing migration" id="populate library_element folder_uid" grafana | logger=migrator t=2025-06-18T15:20:30.866954283Z level=info msg="Migration successfully executed" id="populate library_element folder_uid" duration=415.844µs grafana | logger=migrator t=2025-06-18T15:20:30.870547888Z level=info msg="Executing migration" id="add index library_element org_id-folder_uid-name-kind" grafana | logger=migrator t=2025-06-18T15:20:30.871659418Z level=info msg="Migration successfully executed" id="add index library_element org_id-folder_uid-name-kind" duration=1.11194ms grafana | logger=migrator t=2025-06-18T15:20:30.878187869Z level=info msg="Executing migration" id="clone move dashboard alerts to unified alerting" grafana | logger=migrator t=2025-06-18T15:20:30.878452822Z level=info msg="Migration successfully executed" id="clone move dashboard alerts to unified alerting" duration=264.533µs grafana | logger=migrator t=2025-06-18T15:20:30.882748303Z level=info msg="Executing migration" id="create data_keys table" grafana | logger=migrator t=2025-06-18T15:20:30.884277087Z level=info msg="Migration successfully executed" id="create data_keys table" duration=1.528444ms grafana | logger=migrator t=2025-06-18T15:20:30.888025784Z level=info msg="Executing migration" id="create secrets table" grafana | logger=migrator t=2025-06-18T15:20:30.888871361Z level=info msg="Migration successfully executed" id="create secrets table" duration=845.457µs grafana | logger=migrator t=2025-06-18T15:20:30.892345514Z level=info msg="Executing migration" id="rename data_keys name column to id" grafana | logger=migrator t=2025-06-18T15:20:30.925981405Z level=info msg="Migration successfully executed" id="rename data_keys name column to id" duration=33.631621ms grafana | logger=migrator t=2025-06-18T15:20:30.94641029Z level=info msg="Executing migration" id="add name column into data_keys" grafana | logger=migrator t=2025-06-18T15:20:30.951827681Z level=info msg="Migration successfully executed" id="add name column into data_keys" duration=5.416391ms grafana | logger=migrator t=2025-06-18T15:20:30.955468876Z level=info msg="Executing migration" id="copy data_keys id column values into name" grafana | logger=migrator t=2025-06-18T15:20:30.955594177Z level=info msg="Migration successfully executed" id="copy data_keys id column values into name" duration=125.351µs grafana | logger=migrator t=2025-06-18T15:20:30.95798665Z level=info msg="Executing migration" id="rename data_keys name column to label" grafana | logger=migrator t=2025-06-18T15:20:30.991140086Z level=info msg="Migration successfully executed" id="rename data_keys name column to label" duration=33.151376ms grafana | logger=migrator t=2025-06-18T15:20:30.997299724Z level=info msg="Executing migration" id="rename data_keys id column back to name" grafana | logger=migrator t=2025-06-18T15:20:31.029978971Z level=info msg="Migration successfully executed" id="rename data_keys id column back to name" duration=32.677417ms grafana | logger=migrator t=2025-06-18T15:20:31.033252234Z level=info msg="Executing migration" id="create kv_store table v1" grafana | logger=migrator t=2025-06-18T15:20:31.033975771Z level=info msg="Migration successfully executed" id="create kv_store table v1" duration=722.517µs grafana | logger=migrator t=2025-06-18T15:20:31.03700155Z level=info msg="Executing migration" id="add index kv_store.org_id-namespace-key" grafana | logger=migrator t=2025-06-18T15:20:31.037807138Z level=info msg="Migration successfully executed" id="add index kv_store.org_id-namespace-key" duration=805.168µs grafana | logger=migrator t=2025-06-18T15:20:31.042191341Z level=info msg="Executing migration" id="update dashboard_uid and panel_id from existing annotations" grafana | logger=migrator t=2025-06-18T15:20:31.042424343Z level=info msg="Migration successfully executed" id="update dashboard_uid and panel_id from existing annotations" duration=233.032µs grafana | logger=migrator t=2025-06-18T15:20:31.046270041Z level=info msg="Executing migration" id="create permission table" grafana | logger=migrator t=2025-06-18T15:20:31.04710616Z level=info msg="Migration successfully executed" id="create permission table" duration=836.339µs grafana | logger=migrator t=2025-06-18T15:20:31.051458002Z level=info msg="Executing migration" id="add unique index permission.role_id" grafana | logger=migrator t=2025-06-18T15:20:31.052808906Z level=info msg="Migration successfully executed" id="add unique index permission.role_id" duration=1.350284ms grafana | logger=migrator t=2025-06-18T15:20:31.056995456Z level=info msg="Executing migration" id="add unique index role_id_action_scope" grafana | logger=migrator t=2025-06-18T15:20:31.058546852Z level=info msg="Migration successfully executed" id="add unique index role_id_action_scope" duration=1.550456ms grafana | logger=migrator t=2025-06-18T15:20:31.063658871Z level=info msg="Executing migration" id="create role table" grafana | logger=migrator t=2025-06-18T15:20:31.064632461Z level=info msg="Migration successfully executed" id="create role table" duration=973.87µs grafana | logger=migrator t=2025-06-18T15:20:31.06871751Z level=info msg="Executing migration" id="add column display_name" grafana | logger=migrator t=2025-06-18T15:20:31.076711429Z level=info msg="Migration successfully executed" id="add column display_name" duration=7.993139ms grafana | logger=migrator t=2025-06-18T15:20:31.105766133Z level=info msg="Executing migration" id="add column group_name" grafana | logger=migrator t=2025-06-18T15:20:31.113768282Z level=info msg="Migration successfully executed" id="add column group_name" duration=7.999409ms grafana | logger=migrator t=2025-06-18T15:20:31.117851202Z level=info msg="Executing migration" id="add index role.org_id" grafana | logger=migrator t=2025-06-18T15:20:31.118919162Z level=info msg="Migration successfully executed" id="add index role.org_id" duration=1.07355ms grafana | logger=migrator t=2025-06-18T15:20:31.122097253Z level=info msg="Executing migration" id="add unique index role_org_id_name" grafana | logger=migrator t=2025-06-18T15:20:31.123058493Z level=info msg="Migration successfully executed" id="add unique index role_org_id_name" duration=965.15µs grafana | logger=migrator t=2025-06-18T15:20:31.126665279Z level=info msg="Executing migration" id="add index role_org_id_uid" grafana | logger=migrator t=2025-06-18T15:20:31.128719929Z level=info msg="Migration successfully executed" id="add index role_org_id_uid" duration=2.05317ms grafana | logger=migrator t=2025-06-18T15:20:31.135645656Z level=info msg="Executing migration" id="create team role table" grafana | logger=migrator t=2025-06-18T15:20:31.136801568Z level=info msg="Migration successfully executed" id="create team role table" duration=1.154792ms grafana | logger=migrator t=2025-06-18T15:20:31.143115109Z level=info msg="Executing migration" id="add index team_role.org_id" grafana | logger=migrator t=2025-06-18T15:20:31.144324251Z level=info msg="Migration successfully executed" id="add index team_role.org_id" duration=1.209092ms grafana | logger=migrator t=2025-06-18T15:20:31.149540142Z level=info msg="Executing migration" id="add unique index team_role_org_id_team_id_role_id" grafana | logger=migrator t=2025-06-18T15:20:31.150797725Z level=info msg="Migration successfully executed" id="add unique index team_role_org_id_team_id_role_id" duration=1.253693ms grafana | logger=migrator t=2025-06-18T15:20:31.154517581Z level=info msg="Executing migration" id="add index team_role.team_id" grafana | logger=migrator t=2025-06-18T15:20:31.155727543Z level=info msg="Migration successfully executed" id="add index team_role.team_id" duration=1.209972ms grafana | logger=migrator t=2025-06-18T15:20:31.159643582Z level=info msg="Executing migration" id="create user role table" grafana | logger=migrator t=2025-06-18T15:20:31.16051621Z level=info msg="Migration successfully executed" id="create user role table" duration=872.968µs grafana | logger=migrator t=2025-06-18T15:20:31.165265407Z level=info msg="Executing migration" id="add index user_role.org_id" grafana | logger=migrator t=2025-06-18T15:20:31.166384067Z level=info msg="Migration successfully executed" id="add index user_role.org_id" duration=1.11826ms grafana | logger=migrator t=2025-06-18T15:20:31.171126344Z level=info msg="Executing migration" id="add unique index user_role_org_id_user_id_role_id" grafana | logger=migrator t=2025-06-18T15:20:31.172257674Z level=info msg="Migration successfully executed" id="add unique index user_role_org_id_user_id_role_id" duration=1.13101ms grafana | logger=migrator t=2025-06-18T15:20:31.175767698Z level=info msg="Executing migration" id="add index user_role.user_id" grafana | logger=migrator t=2025-06-18T15:20:31.17689007Z level=info msg="Migration successfully executed" id="add index user_role.user_id" duration=1.122452ms grafana | logger=migrator t=2025-06-18T15:20:31.181183342Z level=info msg="Executing migration" id="create builtin role table" grafana | logger=migrator t=2025-06-18T15:20:31.182096512Z level=info msg="Migration successfully executed" id="create builtin role table" duration=913.17µs grafana | logger=migrator t=2025-06-18T15:20:31.187948039Z level=info msg="Executing migration" id="add index builtin_role.role_id" grafana | logger=migrator t=2025-06-18T15:20:31.189950619Z level=info msg="Migration successfully executed" id="add index builtin_role.role_id" duration=2.001999ms grafana | logger=migrator t=2025-06-18T15:20:31.194158819Z level=info msg="Executing migration" id="add index builtin_role.name" grafana | logger=migrator t=2025-06-18T15:20:31.195171339Z level=info msg="Migration successfully executed" id="add index builtin_role.name" duration=1.01215ms grafana | logger=migrator t=2025-06-18T15:20:31.206694613Z level=info msg="Executing migration" id="Add column org_id to builtin_role table" grafana | logger=migrator t=2025-06-18T15:20:31.214893312Z level=info msg="Migration successfully executed" id="Add column org_id to builtin_role table" duration=8.200069ms grafana | logger=migrator t=2025-06-18T15:20:31.224375825Z level=info msg="Executing migration" id="add index builtin_role.org_id" grafana | logger=migrator t=2025-06-18T15:20:31.225519426Z level=info msg="Migration successfully executed" id="add index builtin_role.org_id" duration=1.141701ms grafana | logger=migrator t=2025-06-18T15:20:31.230452085Z level=info msg="Executing migration" id="add unique index builtin_role_org_id_role_id_role" grafana | logger=migrator t=2025-06-18T15:20:31.231495424Z level=info msg="Migration successfully executed" id="add unique index builtin_role_org_id_role_id_role" duration=1.042859ms grafana | logger=migrator t=2025-06-18T15:20:31.235143851Z level=info msg="Executing migration" id="Remove unique index role_org_id_uid" grafana | logger=migrator t=2025-06-18T15:20:31.236328082Z level=info msg="Migration successfully executed" id="Remove unique index role_org_id_uid" duration=1.183731ms grafana | logger=migrator t=2025-06-18T15:20:31.275619927Z level=info msg="Executing migration" id="add unique index role.uid" grafana | logger=migrator t=2025-06-18T15:20:31.27802756Z level=info msg="Migration successfully executed" id="add unique index role.uid" duration=2.410503ms grafana | logger=migrator t=2025-06-18T15:20:31.282191651Z level=info msg="Executing migration" id="create seed assignment table" grafana | logger=migrator t=2025-06-18T15:20:31.282990309Z level=info msg="Migration successfully executed" id="create seed assignment table" duration=798.678µs grafana | logger=migrator t=2025-06-18T15:20:31.286551384Z level=info msg="Executing migration" id="add unique index builtin_role_role_name" grafana | logger=migrator t=2025-06-18T15:20:31.287596324Z level=info msg="Migration successfully executed" id="add unique index builtin_role_role_name" duration=1.04467ms grafana | logger=migrator t=2025-06-18T15:20:31.292336541Z level=info msg="Executing migration" id="add column hidden to role table" grafana | logger=migrator t=2025-06-18T15:20:31.300379089Z level=info msg="Migration successfully executed" id="add column hidden to role table" duration=8.041788ms grafana | logger=migrator t=2025-06-18T15:20:31.304100115Z level=info msg="Executing migration" id="permission kind migration" grafana | logger=migrator t=2025-06-18T15:20:31.312487878Z level=info msg="Migration successfully executed" id="permission kind migration" duration=8.386593ms grafana | logger=migrator t=2025-06-18T15:20:31.317061213Z level=info msg="Executing migration" id="permission attribute migration" grafana | logger=migrator t=2025-06-18T15:20:31.325522466Z level=info msg="Migration successfully executed" id="permission attribute migration" duration=8.460093ms grafana | logger=migrator t=2025-06-18T15:20:31.329805167Z level=info msg="Executing migration" id="permission identifier migration" grafana | logger=migrator t=2025-06-18T15:20:31.338091268Z level=info msg="Migration successfully executed" id="permission identifier migration" duration=8.285631ms grafana | logger=migrator t=2025-06-18T15:20:31.342051508Z level=info msg="Executing migration" id="add permission identifier index" grafana | logger=migrator t=2025-06-18T15:20:31.343095768Z level=info msg="Migration successfully executed" id="add permission identifier index" duration=1.04399ms grafana | logger=migrator t=2025-06-18T15:20:31.346496501Z level=info msg="Executing migration" id="add permission action scope role_id index" grafana | logger=migrator t=2025-06-18T15:20:31.347542291Z level=info msg="Migration successfully executed" id="add permission action scope role_id index" duration=1.04807ms grafana | logger=migrator t=2025-06-18T15:20:31.35658775Z level=info msg="Executing migration" id="remove permission role_id action scope index" grafana | logger=migrator t=2025-06-18T15:20:31.358108285Z level=info msg="Migration successfully executed" id="remove permission role_id action scope index" duration=1.520125ms grafana | logger=migrator t=2025-06-18T15:20:31.361994203Z level=info msg="Executing migration" id="add group mapping UID column to user_role table" grafana | logger=migrator t=2025-06-18T15:20:31.373735498Z level=info msg="Migration successfully executed" id="add group mapping UID column to user_role table" duration=11.741405ms grafana | logger=migrator t=2025-06-18T15:20:31.382912247Z level=info msg="Executing migration" id="add user_role org ID, user ID, role ID, group mapping UID index" grafana | logger=migrator t=2025-06-18T15:20:31.38524485Z level=info msg="Migration successfully executed" id="add user_role org ID, user ID, role ID, group mapping UID index" duration=2.332013ms grafana | logger=migrator t=2025-06-18T15:20:31.390944496Z level=info msg="Executing migration" id="remove user_role org ID, user ID, role ID index" grafana | logger=migrator t=2025-06-18T15:20:31.392041847Z level=info msg="Migration successfully executed" id="remove user_role org ID, user ID, role ID index" duration=1.097171ms grafana | logger=migrator t=2025-06-18T15:20:31.404267537Z level=info msg="Executing migration" id="create query_history table v1" grafana | logger=migrator t=2025-06-18T15:20:31.405892743Z level=info msg="Migration successfully executed" id="create query_history table v1" duration=1.623746ms grafana | logger=migrator t=2025-06-18T15:20:31.410712639Z level=info msg="Executing migration" id="add index query_history.org_id-created_by-datasource_uid" grafana | logger=migrator t=2025-06-18T15:20:31.412481477Z level=info msg="Migration successfully executed" id="add index query_history.org_id-created_by-datasource_uid" duration=1.767188ms grafana | logger=migrator t=2025-06-18T15:20:31.417513207Z level=info msg="Executing migration" id="alter table query_history alter column created_by type to bigint" grafana | logger=migrator t=2025-06-18T15:20:31.417539557Z level=info msg="Migration successfully executed" id="alter table query_history alter column created_by type to bigint" duration=26.65µs grafana | logger=migrator t=2025-06-18T15:20:31.421237303Z level=info msg="Executing migration" id="create query_history_details table v1" grafana | logger=migrator t=2025-06-18T15:20:31.422314053Z level=info msg="Migration successfully executed" id="create query_history_details table v1" duration=1.07642ms grafana | logger=migrator t=2025-06-18T15:20:31.44141476Z level=info msg="Executing migration" id="rbac disabled migrator" grafana | logger=migrator t=2025-06-18T15:20:31.441509071Z level=info msg="Migration successfully executed" id="rbac disabled migrator" duration=98.661µs grafana | logger=migrator t=2025-06-18T15:20:31.448028226Z level=info msg="Executing migration" id="teams permissions migration" grafana | logger=migrator t=2025-06-18T15:20:31.448660472Z level=info msg="Migration successfully executed" id="teams permissions migration" duration=632.076µs grafana | logger=migrator t=2025-06-18T15:20:31.454204316Z level=info msg="Executing migration" id="dashboard permissions" grafana | logger=migrator t=2025-06-18T15:20:31.455067354Z level=info msg="Migration successfully executed" id="dashboard permissions" duration=864.168µs grafana | logger=migrator t=2025-06-18T15:20:31.459227635Z level=info msg="Executing migration" id="dashboard permissions uid scopes" grafana | logger=migrator t=2025-06-18T15:20:31.460197174Z level=info msg="Migration successfully executed" id="dashboard permissions uid scopes" duration=969.099µs grafana | logger=migrator t=2025-06-18T15:20:31.465287164Z level=info msg="Executing migration" id="drop managed folder create actions" grafana | logger=migrator t=2025-06-18T15:20:31.465527557Z level=info msg="Migration successfully executed" id="drop managed folder create actions" duration=237.662µs grafana | logger=migrator t=2025-06-18T15:20:31.471599946Z level=info msg="Executing migration" id="alerting notification permissions" grafana | logger=migrator t=2025-06-18T15:20:31.472127341Z level=info msg="Migration successfully executed" id="alerting notification permissions" duration=529.345µs grafana | logger=migrator t=2025-06-18T15:20:31.477373403Z level=info msg="Executing migration" id="create query_history_star table v1" grafana | logger=migrator t=2025-06-18T15:20:31.478363673Z level=info msg="Migration successfully executed" id="create query_history_star table v1" duration=990.079µs grafana | logger=migrator t=2025-06-18T15:20:31.481830797Z level=info msg="Executing migration" id="add index query_history.user_id-query_uid" grafana | logger=migrator t=2025-06-18T15:20:31.483516192Z level=info msg="Migration successfully executed" id="add index query_history.user_id-query_uid" duration=1.683765ms grafana | logger=migrator t=2025-06-18T15:20:31.489028206Z level=info msg="Executing migration" id="add column org_id in query_history_star" grafana | logger=migrator t=2025-06-18T15:20:31.499727342Z level=info msg="Migration successfully executed" id="add column org_id in query_history_star" duration=10.673455ms grafana | logger=migrator t=2025-06-18T15:20:31.506024083Z level=info msg="Executing migration" id="alter table query_history_star_mig column user_id type to bigint" grafana | logger=migrator t=2025-06-18T15:20:31.506042343Z level=info msg="Migration successfully executed" id="alter table query_history_star_mig column user_id type to bigint" duration=20.42µs grafana | logger=migrator t=2025-06-18T15:20:31.510529977Z level=info msg="Executing migration" id="create correlation table v1" grafana | logger=migrator t=2025-06-18T15:20:31.512253655Z level=info msg="Migration successfully executed" id="create correlation table v1" duration=1.723107ms grafana | logger=migrator t=2025-06-18T15:20:31.51592123Z level=info msg="Executing migration" id="add index correlations.uid" grafana | logger=migrator t=2025-06-18T15:20:31.519713197Z level=info msg="Migration successfully executed" id="add index correlations.uid" duration=3.792617ms grafana | logger=migrator t=2025-06-18T15:20:31.524363573Z level=info msg="Executing migration" id="add index correlations.source_uid" grafana | logger=migrator t=2025-06-18T15:20:31.525909948Z level=info msg="Migration successfully executed" id="add index correlations.source_uid" duration=1.545745ms grafana | logger=migrator t=2025-06-18T15:20:31.529610374Z level=info msg="Executing migration" id="add correlation config column" grafana | logger=migrator t=2025-06-18T15:20:31.539098637Z level=info msg="Migration successfully executed" id="add correlation config column" duration=9.490623ms grafana | logger=migrator t=2025-06-18T15:20:31.54457874Z level=info msg="Executing migration" id="drop index IDX_correlation_uid - v1" grafana | logger=migrator t=2025-06-18T15:20:31.545800023Z level=info msg="Migration successfully executed" id="drop index IDX_correlation_uid - v1" duration=1.221433ms grafana | logger=migrator t=2025-06-18T15:20:31.572822657Z level=info msg="Executing migration" id="drop index IDX_correlation_source_uid - v1" grafana | logger=migrator t=2025-06-18T15:20:31.574964158Z level=info msg="Migration successfully executed" id="drop index IDX_correlation_source_uid - v1" duration=2.142511ms grafana | logger=migrator t=2025-06-18T15:20:31.580010568Z level=info msg="Executing migration" id="Rename table correlation to correlation_tmp_qwerty - v1" grafana | logger=migrator t=2025-06-18T15:20:31.604289795Z level=info msg="Migration successfully executed" id="Rename table correlation to correlation_tmp_qwerty - v1" duration=24.277397ms grafana | logger=migrator t=2025-06-18T15:20:31.629109048Z level=info msg="Executing migration" id="create correlation v2" grafana | logger=migrator t=2025-06-18T15:20:31.630710824Z level=info msg="Migration successfully executed" id="create correlation v2" duration=1.603016ms grafana | logger=migrator t=2025-06-18T15:20:31.634699463Z level=info msg="Executing migration" id="create index IDX_correlation_uid - v2" grafana | logger=migrator t=2025-06-18T15:20:31.635830795Z level=info msg="Migration successfully executed" id="create index IDX_correlation_uid - v2" duration=1.131282ms grafana | logger=migrator t=2025-06-18T15:20:31.639701762Z level=info msg="Executing migration" id="create index IDX_correlation_source_uid - v2" grafana | logger=migrator t=2025-06-18T15:20:31.640922445Z level=info msg="Migration successfully executed" id="create index IDX_correlation_source_uid - v2" duration=1.220713ms grafana | logger=migrator t=2025-06-18T15:20:31.649006663Z level=info msg="Executing migration" id="create index IDX_correlation_org_id - v2" grafana | logger=migrator t=2025-06-18T15:20:31.651400466Z level=info msg="Migration successfully executed" id="create index IDX_correlation_org_id - v2" duration=2.389713ms grafana | logger=migrator t=2025-06-18T15:20:31.657386225Z level=info msg="Executing migration" id="copy correlation v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:31.657712149Z level=info msg="Migration successfully executed" id="copy correlation v1 to v2" duration=325.484µs grafana | logger=migrator t=2025-06-18T15:20:31.662235223Z level=info msg="Executing migration" id="drop correlation_tmp_qwerty" grafana | logger=migrator t=2025-06-18T15:20:31.66395315Z level=info msg="Migration successfully executed" id="drop correlation_tmp_qwerty" duration=1.719347ms grafana | logger=migrator t=2025-06-18T15:20:31.671008608Z level=info msg="Executing migration" id="add provisioning column" grafana | logger=migrator t=2025-06-18T15:20:31.680946576Z level=info msg="Migration successfully executed" id="add provisioning column" duration=9.939058ms grafana | logger=migrator t=2025-06-18T15:20:31.684944925Z level=info msg="Executing migration" id="add type column" grafana | logger=migrator t=2025-06-18T15:20:31.695112085Z level=info msg="Migration successfully executed" id="add type column" duration=10.16535ms grafana | logger=migrator t=2025-06-18T15:20:31.700202045Z level=info msg="Executing migration" id="create entity_events table" grafana | logger=migrator t=2025-06-18T15:20:31.701380726Z level=info msg="Migration successfully executed" id="create entity_events table" duration=1.181391ms grafana | logger=migrator t=2025-06-18T15:20:31.717393643Z level=info msg="Executing migration" id="create dashboard public config v1" grafana | logger=migrator t=2025-06-18T15:20:31.719279072Z level=info msg="Migration successfully executed" id="create dashboard public config v1" duration=1.886338ms grafana | logger=migrator t=2025-06-18T15:20:31.723375322Z level=info msg="Executing migration" id="drop index UQE_dashboard_public_config_uid - v1" grafana | logger=migrator t=2025-06-18T15:20:31.724376781Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop index UQE_dashboard_public_config_uid - v1" grafana | logger=migrator t=2025-06-18T15:20:31.72826143Z level=info msg="Executing migration" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v1" grafana | logger=migrator t=2025-06-18T15:20:31.728948386Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v1" grafana | logger=migrator t=2025-06-18T15:20:31.73339511Z level=info msg="Executing migration" id="Drop old dashboard public config table" grafana | logger=migrator t=2025-06-18T15:20:31.734481301Z level=info msg="Migration successfully executed" id="Drop old dashboard public config table" duration=1.087241ms grafana | logger=migrator t=2025-06-18T15:20:31.742091145Z level=info msg="Executing migration" id="recreate dashboard public config v1" grafana | logger=migrator t=2025-06-18T15:20:31.743555489Z level=info msg="Migration successfully executed" id="recreate dashboard public config v1" duration=1.465424ms grafana | logger=migrator t=2025-06-18T15:20:31.748648479Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_uid - v1" grafana | logger=migrator t=2025-06-18T15:20:31.749966411Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_uid - v1" duration=1.317492ms grafana | logger=migrator t=2025-06-18T15:20:31.776368271Z level=info msg="Executing migration" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v1" grafana | logger=migrator t=2025-06-18T15:20:31.779516571Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v1" duration=3.14451ms grafana | logger=migrator t=2025-06-18T15:20:31.789770532Z level=info msg="Executing migration" id="drop index UQE_dashboard_public_config_uid - v2" grafana | logger=migrator t=2025-06-18T15:20:31.79160689Z level=info msg="Migration successfully executed" id="drop index UQE_dashboard_public_config_uid - v2" duration=1.835588ms grafana | logger=migrator t=2025-06-18T15:20:31.79566232Z level=info msg="Executing migration" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v2" grafana | logger=migrator t=2025-06-18T15:20:31.796847751Z level=info msg="Migration successfully executed" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v2" duration=1.185201ms grafana | logger=migrator t=2025-06-18T15:20:31.800849491Z level=info msg="Executing migration" id="Drop public config table" grafana | logger=migrator t=2025-06-18T15:20:31.801675708Z level=info msg="Migration successfully executed" id="Drop public config table" duration=825.867µs grafana | logger=migrator t=2025-06-18T15:20:31.807107522Z level=info msg="Executing migration" id="Recreate dashboard public config v2" grafana | logger=migrator t=2025-06-18T15:20:31.808488705Z level=info msg="Migration successfully executed" id="Recreate dashboard public config v2" duration=1.380733ms grafana | logger=migrator t=2025-06-18T15:20:31.815698306Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_uid - v2" grafana | logger=migrator t=2025-06-18T15:20:31.817037509Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_uid - v2" duration=1.336852ms grafana | logger=migrator t=2025-06-18T15:20:31.823786095Z level=info msg="Executing migration" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v2" grafana | logger=migrator t=2025-06-18T15:20:31.825843115Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v2" duration=2.05573ms grafana | logger=migrator t=2025-06-18T15:20:31.831669872Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_access_token - v2" grafana | logger=migrator t=2025-06-18T15:20:31.832839853Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_access_token - v2" duration=1.169301ms grafana | logger=migrator t=2025-06-18T15:20:31.839446948Z level=info msg="Executing migration" id="Rename table dashboard_public_config to dashboard_public - v2" grafana | logger=migrator t=2025-06-18T15:20:31.863952288Z level=info msg="Migration successfully executed" id="Rename table dashboard_public_config to dashboard_public - v2" duration=24.4944ms grafana | logger=migrator t=2025-06-18T15:20:31.872816675Z level=info msg="Executing migration" id="add annotations_enabled column" grafana | logger=migrator t=2025-06-18T15:20:31.882665672Z level=info msg="Migration successfully executed" id="add annotations_enabled column" duration=9.850516ms grafana | logger=migrator t=2025-06-18T15:20:31.890447378Z level=info msg="Executing migration" id="add time_selection_enabled column" grafana | logger=migrator t=2025-06-18T15:20:31.900235393Z level=info msg="Migration successfully executed" id="add time_selection_enabled column" duration=9.786875ms grafana | logger=migrator t=2025-06-18T15:20:31.921524732Z level=info msg="Executing migration" id="delete orphaned public dashboards" grafana | logger=migrator t=2025-06-18T15:20:31.922016797Z level=info msg="Migration successfully executed" id="delete orphaned public dashboards" duration=493.295µs grafana | logger=migrator t=2025-06-18T15:20:31.962319861Z level=info msg="Executing migration" id="add share column" grafana | logger=migrator t=2025-06-18T15:20:31.975186428Z level=info msg="Migration successfully executed" id="add share column" duration=12.869537ms grafana | logger=migrator t=2025-06-18T15:20:32.000043661Z level=info msg="Executing migration" id="backfill empty share column fields with default of public" grafana | logger=migrator t=2025-06-18T15:20:32.000365774Z level=info msg="Migration successfully executed" id="backfill empty share column fields with default of public" duration=322.703µs grafana | logger=migrator t=2025-06-18T15:20:32.018505537Z level=info msg="Executing migration" id="create file table" grafana | logger=migrator t=2025-06-18T15:20:32.019915821Z level=info msg="Migration successfully executed" id="create file table" duration=1.415514ms grafana | logger=migrator t=2025-06-18T15:20:32.033813422Z level=info msg="Executing migration" id="file table idx: path natural pk" grafana | logger=migrator t=2025-06-18T15:20:32.035671459Z level=info msg="Migration successfully executed" id="file table idx: path natural pk" duration=1.859577ms grafana | logger=migrator t=2025-06-18T15:20:32.046092458Z level=info msg="Executing migration" id="file table idx: parent_folder_path_hash fast folder retrieval" grafana | logger=migrator t=2025-06-18T15:20:32.047520082Z level=info msg="Migration successfully executed" id="file table idx: parent_folder_path_hash fast folder retrieval" duration=1.427484ms grafana | logger=migrator t=2025-06-18T15:20:32.095569545Z level=info msg="Executing migration" id="create file_meta table" grafana | logger=migrator t=2025-06-18T15:20:32.097213489Z level=info msg="Migration successfully executed" id="create file_meta table" duration=1.644914ms grafana | logger=migrator t=2025-06-18T15:20:32.165347003Z level=info msg="Executing migration" id="file table idx: path key" grafana | logger=migrator t=2025-06-18T15:20:32.166700205Z level=info msg="Migration successfully executed" id="file table idx: path key" duration=1.354922ms grafana | logger=migrator t=2025-06-18T15:20:32.197924039Z level=info msg="Executing migration" id="set path collation in file table" grafana | logger=migrator t=2025-06-18T15:20:32.19795777Z level=info msg="Migration successfully executed" id="set path collation in file table" duration=35.881µs grafana | logger=migrator t=2025-06-18T15:20:32.212480107Z level=info msg="Executing migration" id="migrate contents column to mediumblob for MySQL" grafana | logger=migrator t=2025-06-18T15:20:32.212507077Z level=info msg="Migration successfully executed" id="migrate contents column to mediumblob for MySQL" duration=28.44µs grafana | logger=migrator t=2025-06-18T15:20:32.220531573Z level=info msg="Executing migration" id="managed permissions migration" grafana | logger=migrator t=2025-06-18T15:20:32.22118367Z level=info msg="Migration successfully executed" id="managed permissions migration" duration=652.667µs grafana | logger=migrator t=2025-06-18T15:20:32.225846493Z level=info msg="Executing migration" id="managed folder permissions alert actions migration" grafana | logger=migrator t=2025-06-18T15:20:32.226208656Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions migration" duration=361.683µs grafana | logger=migrator t=2025-06-18T15:20:32.230964311Z level=info msg="Executing migration" id="RBAC action name migrator" grafana | logger=migrator t=2025-06-18T15:20:32.232595547Z level=info msg="Migration successfully executed" id="RBAC action name migrator" duration=1.631196ms grafana | logger=migrator t=2025-06-18T15:20:32.236861027Z level=info msg="Executing migration" id="Add UID column to playlist" grafana | logger=migrator t=2025-06-18T15:20:32.250005351Z level=info msg="Migration successfully executed" id="Add UID column to playlist" duration=13.139594ms grafana | logger=migrator t=2025-06-18T15:20:32.262046934Z level=info msg="Executing migration" id="Update uid column values in playlist" grafana | logger=migrator t=2025-06-18T15:20:32.262326747Z level=info msg="Migration successfully executed" id="Update uid column values in playlist" duration=279.483µs grafana | logger=migrator t=2025-06-18T15:20:32.271823307Z level=info msg="Executing migration" id="Add index for uid in playlist" grafana | logger=migrator t=2025-06-18T15:20:32.272887066Z level=info msg="Migration successfully executed" id="Add index for uid in playlist" duration=1.064619ms grafana | logger=migrator t=2025-06-18T15:20:32.323373723Z level=info msg="Executing migration" id="update group index for alert rules" grafana | logger=migrator t=2025-06-18T15:20:32.32422919Z level=info msg="Migration successfully executed" id="update group index for alert rules" duration=858.308µs grafana | logger=migrator t=2025-06-18T15:20:32.360810826Z level=info msg="Executing migration" id="managed folder permissions alert actions repeated migration" grafana | logger=migrator t=2025-06-18T15:20:32.361317141Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions repeated migration" duration=508.585µs grafana | logger=migrator t=2025-06-18T15:20:32.368796341Z level=info msg="Executing migration" id="admin only folder/dashboard permission" grafana | logger=migrator t=2025-06-18T15:20:32.369884142Z level=info msg="Migration successfully executed" id="admin only folder/dashboard permission" duration=1.086601ms grafana | logger=migrator t=2025-06-18T15:20:32.37613504Z level=info msg="Executing migration" id="add action column to seed_assignment" grafana | logger=migrator t=2025-06-18T15:20:32.38664966Z level=info msg="Migration successfully executed" id="add action column to seed_assignment" duration=10.513939ms grafana | logger=migrator t=2025-06-18T15:20:32.399033706Z level=info msg="Executing migration" id="add scope column to seed_assignment" grafana | logger=migrator t=2025-06-18T15:20:32.408136452Z level=info msg="Migration successfully executed" id="add scope column to seed_assignment" duration=9.102756ms grafana | logger=migrator t=2025-06-18T15:20:32.414875716Z level=info msg="Executing migration" id="remove unique index builtin_role_role_name before nullable update" grafana | logger=migrator t=2025-06-18T15:20:32.415853945Z level=info msg="Migration successfully executed" id="remove unique index builtin_role_role_name before nullable update" duration=977.859µs grafana | logger=migrator t=2025-06-18T15:20:32.423119333Z level=info msg="Executing migration" id="update seed_assignment role_name column to nullable" grafana | logger=migrator t=2025-06-18T15:20:32.500664974Z level=info msg="Migration successfully executed" id="update seed_assignment role_name column to nullable" duration=77.541591ms grafana | logger=migrator t=2025-06-18T15:20:32.512414226Z level=info msg="Executing migration" id="add unique index builtin_role_name back" grafana | logger=migrator t=2025-06-18T15:20:32.514734307Z level=info msg="Migration successfully executed" id="add unique index builtin_role_name back" duration=2.313961ms grafana | logger=migrator t=2025-06-18T15:20:32.54574109Z level=info msg="Executing migration" id="add unique index builtin_role_action_scope" grafana | logger=migrator t=2025-06-18T15:20:32.548047661Z level=info msg="Migration successfully executed" id="add unique index builtin_role_action_scope" duration=2.306751ms grafana | logger=migrator t=2025-06-18T15:20:32.563351196Z level=info msg="Executing migration" id="add primary key to seed_assigment" grafana | logger=migrator t=2025-06-18T15:20:32.596269657Z level=info msg="Migration successfully executed" id="add primary key to seed_assigment" duration=32.92665ms grafana | logger=migrator t=2025-06-18T15:20:32.607033238Z level=info msg="Executing migration" id="add origin column to seed_assignment" grafana | logger=migrator t=2025-06-18T15:20:32.618730018Z level=info msg="Migration successfully executed" id="add origin column to seed_assignment" duration=11.695421ms grafana | logger=migrator t=2025-06-18T15:20:32.645643242Z level=info msg="Executing migration" id="add origin to plugin seed_assignment" grafana | logger=migrator t=2025-06-18T15:20:32.646053255Z level=info msg="Migration successfully executed" id="add origin to plugin seed_assignment" duration=411.893µs grafana | logger=migrator t=2025-06-18T15:20:32.654310973Z level=info msg="Executing migration" id="prevent seeding OnCall access" grafana | logger=migrator t=2025-06-18T15:20:32.654537885Z level=info msg="Migration successfully executed" id="prevent seeding OnCall access" duration=226.252µs grafana | logger=migrator t=2025-06-18T15:20:32.658340782Z level=info msg="Executing migration" id="managed folder permissions alert actions repeated fixed migration" grafana | logger=migrator t=2025-06-18T15:20:32.658642245Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions repeated fixed migration" duration=300.893µs grafana | logger=migrator t=2025-06-18T15:20:32.662857514Z level=info msg="Executing migration" id="managed folder permissions library panel actions migration" grafana | logger=migrator t=2025-06-18T15:20:32.663027775Z level=info msg="Migration successfully executed" id="managed folder permissions library panel actions migration" duration=169.951µs grafana | logger=migrator t=2025-06-18T15:20:32.681655202Z level=info msg="Executing migration" id="migrate external alertmanagers to datsourcse" grafana | logger=migrator t=2025-06-18T15:20:32.682061375Z level=info msg="Migration successfully executed" id="migrate external alertmanagers to datsourcse" duration=407.794µs grafana | logger=migrator t=2025-06-18T15:20:32.758407215Z level=info msg="Executing migration" id="create folder table" grafana | logger=migrator t=2025-06-18T15:20:32.760114281Z level=info msg="Migration successfully executed" id="create folder table" duration=1.730516ms grafana | logger=migrator t=2025-06-18T15:20:32.8680643Z level=info msg="Executing migration" id="Add index for parent_uid" grafana | logger=migrator t=2025-06-18T15:20:32.869485573Z level=info msg="Migration successfully executed" id="Add index for parent_uid" duration=1.422793ms grafana | logger=migrator t=2025-06-18T15:20:32.918474135Z level=info msg="Executing migration" id="Add unique index for folder.uid and folder.org_id" grafana | logger=migrator t=2025-06-18T15:20:32.919819007Z level=info msg="Migration successfully executed" id="Add unique index for folder.uid and folder.org_id" duration=1.345952ms grafana | logger=migrator t=2025-06-18T15:20:32.963587641Z level=info msg="Executing migration" id="Update folder title length" grafana | logger=migrator t=2025-06-18T15:20:32.963644091Z level=info msg="Migration successfully executed" id="Update folder title length" duration=58.92µs grafana | logger=migrator t=2025-06-18T15:20:33.066293678Z level=info msg="Executing migration" id="Add unique index for folder.title and folder.parent_uid" grafana | logger=migrator t=2025-06-18T15:20:33.068391508Z level=info msg="Migration successfully executed" id="Add unique index for folder.title and folder.parent_uid" duration=2.09805ms grafana | logger=migrator t=2025-06-18T15:20:33.099246499Z level=info msg="Executing migration" id="Remove unique index for folder.title and folder.parent_uid" grafana | logger=migrator t=2025-06-18T15:20:33.101358489Z level=info msg="Migration successfully executed" id="Remove unique index for folder.title and folder.parent_uid" duration=2.11277ms grafana | logger=migrator t=2025-06-18T15:20:33.140775301Z level=info msg="Executing migration" id="Add unique index for title, parent_uid, and org_id" grafana | logger=migrator t=2025-06-18T15:20:33.143242854Z level=info msg="Migration successfully executed" id="Add unique index for title, parent_uid, and org_id" duration=2.466423ms grafana | logger=migrator t=2025-06-18T15:20:33.18003366Z level=info msg="Executing migration" id="Sync dashboard and folder table" grafana | logger=migrator t=2025-06-18T15:20:33.180820218Z level=info msg="Migration successfully executed" id="Sync dashboard and folder table" duration=785.987µs grafana | logger=migrator t=2025-06-18T15:20:33.188580571Z level=info msg="Executing migration" id="Remove ghost folders from the folder table" grafana | logger=migrator t=2025-06-18T15:20:33.189167347Z level=info msg="Migration successfully executed" id="Remove ghost folders from the folder table" duration=595.606µs grafana | logger=migrator t=2025-06-18T15:20:33.197498855Z level=info msg="Executing migration" id="Remove unique index UQE_folder_uid_org_id" grafana | logger=migrator t=2025-06-18T15:20:33.199673716Z level=info msg="Migration successfully executed" id="Remove unique index UQE_folder_uid_org_id" duration=2.175361ms grafana | logger=migrator t=2025-06-18T15:20:33.208934213Z level=info msg="Executing migration" id="Add unique index UQE_folder_org_id_uid" grafana | logger=migrator t=2025-06-18T15:20:33.21068055Z level=info msg="Migration successfully executed" id="Add unique index UQE_folder_org_id_uid" duration=1.747057ms grafana | logger=migrator t=2025-06-18T15:20:33.226912273Z level=info msg="Executing migration" id="Remove unique index UQE_folder_title_parent_uid_org_id" grafana | logger=migrator t=2025-06-18T15:20:33.228918071Z level=info msg="Migration successfully executed" id="Remove unique index UQE_folder_title_parent_uid_org_id" duration=2.003408ms grafana | logger=migrator t=2025-06-18T15:20:33.278490868Z level=info msg="Executing migration" id="Add unique index UQE_folder_org_id_parent_uid_title" grafana | logger=migrator t=2025-06-18T15:20:33.280476637Z level=info msg="Migration successfully executed" id="Add unique index UQE_folder_org_id_parent_uid_title" duration=1.987619ms grafana | logger=migrator t=2025-06-18T15:20:33.343090818Z level=info msg="Executing migration" id="Remove index IDX_folder_parent_uid_org_id" grafana | logger=migrator t=2025-06-18T15:20:33.344628933Z level=info msg="Migration successfully executed" id="Remove index IDX_folder_parent_uid_org_id" duration=1.538684ms grafana | logger=migrator t=2025-06-18T15:20:33.40907079Z level=info msg="Executing migration" id="Remove unique index UQE_folder_org_id_parent_uid_title" grafana | logger=migrator t=2025-06-18T15:20:33.410908186Z level=info msg="Migration successfully executed" id="Remove unique index UQE_folder_org_id_parent_uid_title" duration=1.837746ms grafana | logger=migrator t=2025-06-18T15:20:33.464491951Z level=info msg="Executing migration" id="create anon_device table" grafana | logger=migrator t=2025-06-18T15:20:33.46535049Z level=info msg="Migration successfully executed" id="create anon_device table" duration=860.469µs grafana | logger=migrator t=2025-06-18T15:20:33.625090236Z level=info msg="Executing migration" id="add unique index anon_device.device_id" grafana | logger=migrator t=2025-06-18T15:20:33.627216626Z level=info msg="Migration successfully executed" id="add unique index anon_device.device_id" duration=2.12791ms grafana | logger=migrator t=2025-06-18T15:20:33.64568764Z level=info msg="Executing migration" id="add index anon_device.updated_at" grafana | logger=migrator t=2025-06-18T15:20:33.646786191Z level=info msg="Migration successfully executed" id="add index anon_device.updated_at" duration=1.100361ms grafana | logger=migrator t=2025-06-18T15:20:33.777657673Z level=info msg="Executing migration" id="create signing_key table" grafana | logger=migrator t=2025-06-18T15:20:33.778834415Z level=info msg="Migration successfully executed" id="create signing_key table" duration=1.178192ms grafana | logger=migrator t=2025-06-18T15:20:33.818849782Z level=info msg="Executing migration" id="add unique index signing_key.key_id" grafana | logger=migrator t=2025-06-18T15:20:33.820214865Z level=info msg="Migration successfully executed" id="add unique index signing_key.key_id" duration=1.366253ms grafana | logger=migrator t=2025-06-18T15:20:33.864913116Z level=info msg="Executing migration" id="set legacy alert migration status in kvstore" grafana | logger=migrator t=2025-06-18T15:20:33.866051697Z level=info msg="Migration successfully executed" id="set legacy alert migration status in kvstore" duration=1.150611ms grafana | logger=migrator t=2025-06-18T15:20:33.915522813Z level=info msg="Executing migration" id="migrate record of created folders during legacy migration to kvstore" grafana | logger=migrator t=2025-06-18T15:20:33.916095079Z level=info msg="Migration successfully executed" id="migrate record of created folders during legacy migration to kvstore" duration=572.576µs grafana | logger=migrator t=2025-06-18T15:20:33.944719249Z level=info msg="Executing migration" id="Add folder_uid for dashboard" grafana | logger=migrator t=2025-06-18T15:20:33.956477789Z level=info msg="Migration successfully executed" id="Add folder_uid for dashboard" duration=11.75714ms grafana | logger=migrator t=2025-06-18T15:20:33.963678227Z level=info msg="Executing migration" id="Populate dashboard folder_uid column" grafana | logger=migrator t=2025-06-18T15:20:33.965241962Z level=info msg="Migration successfully executed" id="Populate dashboard folder_uid column" duration=1.565895ms grafana | logger=migrator t=2025-06-18T15:20:33.977905622Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_folder_uid_title" grafana | logger=migrator t=2025-06-18T15:20:33.977957512Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_folder_uid_title" duration=54.82µs grafana | logger=migrator t=2025-06-18T15:20:33.987468951Z level=info msg="Executing migration" id="Delete unique index for dashboard_org_id_folder_id_title" grafana | logger=migrator t=2025-06-18T15:20:33.988721434Z level=info msg="Migration successfully executed" id="Delete unique index for dashboard_org_id_folder_id_title" duration=1.252933ms grafana | logger=migrator t=2025-06-18T15:20:34.007121217Z level=info msg="Executing migration" id="Delete unique index for dashboard_org_id_folder_uid_title" grafana | logger=migrator t=2025-06-18T15:20:34.007161887Z level=info msg="Migration successfully executed" id="Delete unique index for dashboard_org_id_folder_uid_title" duration=44.451µs grafana | logger=migrator t=2025-06-18T15:20:34.023824284Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_folder_uid_title_is_folder" grafana | logger=migrator t=2025-06-18T15:20:34.025356129Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_folder_uid_title_is_folder" duration=1.533435ms grafana | logger=migrator t=2025-06-18T15:20:34.035117471Z level=info msg="Executing migration" id="Restore index for dashboard_org_id_folder_id_title" grafana | logger=migrator t=2025-06-18T15:20:34.037966477Z level=info msg="Migration successfully executed" id="Restore index for dashboard_org_id_folder_id_title" duration=2.848056ms grafana | logger=migrator t=2025-06-18T15:20:34.068774588Z level=info msg="Executing migration" id="Remove unique index for dashboard_org_id_folder_uid_title_is_folder" grafana | logger=migrator t=2025-06-18T15:20:34.070962058Z level=info msg="Migration successfully executed" id="Remove unique index for dashboard_org_id_folder_uid_title_is_folder" duration=2.18343ms grafana | logger=migrator t=2025-06-18T15:20:34.118078622Z level=info msg="Executing migration" id="create sso_setting table" grafana | logger=migrator t=2025-06-18T15:20:34.119324494Z level=info msg="Migration successfully executed" id="create sso_setting table" duration=1.247472ms grafana | logger=migrator t=2025-06-18T15:20:34.16235547Z level=info msg="Executing migration" id="copy kvstore migration status to each org" grafana | logger=migrator t=2025-06-18T15:20:34.163680592Z level=info msg="Migration successfully executed" id="copy kvstore migration status to each org" duration=1.327442ms grafana | logger=migrator t=2025-06-18T15:20:34.25053211Z level=info msg="Executing migration" id="add back entry for orgid=0 migrated status" grafana | logger=migrator t=2025-06-18T15:20:34.251220936Z level=info msg="Migration successfully executed" id="add back entry for orgid=0 migrated status" duration=689.966µs grafana | logger=migrator t=2025-06-18T15:20:34.318650492Z level=info msg="Executing migration" id="managed dashboard permissions annotation actions migration" grafana | logger=migrator t=2025-06-18T15:20:34.319958175Z level=info msg="Migration successfully executed" id="managed dashboard permissions annotation actions migration" duration=1.309963ms grafana | logger=migrator t=2025-06-18T15:20:34.349554003Z level=info msg="Executing migration" id="create cloud_migration table v1" grafana | logger=migrator t=2025-06-18T15:20:34.350707394Z level=info msg="Migration successfully executed" id="create cloud_migration table v1" duration=1.154221ms grafana | logger=migrator t=2025-06-18T15:20:34.455198618Z level=info msg="Executing migration" id="create cloud_migration_run table v1" grafana | logger=migrator t=2025-06-18T15:20:34.456437161Z level=info msg="Migration successfully executed" id="create cloud_migration_run table v1" duration=1.239773ms grafana | logger=migrator t=2025-06-18T15:20:34.479893591Z level=info msg="Executing migration" id="add stack_id column" grafana | logger=migrator t=2025-06-18T15:20:34.490160158Z level=info msg="Migration successfully executed" id="add stack_id column" duration=10.266197ms grafana | logger=migrator t=2025-06-18T15:20:34.51257445Z level=info msg="Executing migration" id="add region_slug column" grafana | logger=migrator t=2025-06-18T15:20:34.524706614Z level=info msg="Migration successfully executed" id="add region_slug column" duration=12.128884ms grafana | logger=migrator t=2025-06-18T15:20:34.604698128Z level=info msg="Executing migration" id="add cluster_slug column" grafana | logger=migrator t=2025-06-18T15:20:34.616286197Z level=info msg="Migration successfully executed" id="add cluster_slug column" duration=11.589079ms grafana | logger=migrator t=2025-06-18T15:20:34.640448564Z level=info msg="Executing migration" id="add migration uid column" grafana | logger=migrator t=2025-06-18T15:20:34.657168251Z level=info msg="Migration successfully executed" id="add migration uid column" duration=16.720507ms grafana | logger=migrator t=2025-06-18T15:20:34.743203182Z level=info msg="Executing migration" id="Update uid column values for migration" grafana | logger=migrator t=2025-06-18T15:20:34.743611396Z level=info msg="Migration successfully executed" id="Update uid column values for migration" duration=408.074µs grafana | logger=migrator t=2025-06-18T15:20:34.833856376Z level=info msg="Executing migration" id="Add unique index migration_uid" grafana | logger=migrator t=2025-06-18T15:20:34.83639268Z level=info msg="Migration successfully executed" id="Add unique index migration_uid" duration=2.537324ms grafana | logger=migrator t=2025-06-18T15:20:34.910554989Z level=info msg="Executing migration" id="add migration run uid column" grafana | logger=migrator t=2025-06-18T15:20:34.920045318Z level=info msg="Migration successfully executed" id="add migration run uid column" duration=9.490319ms grafana | logger=migrator t=2025-06-18T15:20:35.069368925Z level=info msg="Executing migration" id="Update uid column values for migration run" grafana | logger=migrator t=2025-06-18T15:20:35.06986727Z level=info msg="Migration successfully executed" id="Update uid column values for migration run" duration=500.646µs grafana | logger=migrator t=2025-06-18T15:20:35.086796639Z level=info msg="Executing migration" id="Add unique index migration_run_uid" grafana | logger=migrator t=2025-06-18T15:20:35.088171132Z level=info msg="Migration successfully executed" id="Add unique index migration_run_uid" duration=1.375213ms grafana | logger=migrator t=2025-06-18T15:20:35.114794163Z level=info msg="Executing migration" id="Rename table cloud_migration to cloud_migration_session_tmp_qwerty - v1" grafana | logger=migrator t=2025-06-18T15:20:35.143438953Z level=info msg="Migration successfully executed" id="Rename table cloud_migration to cloud_migration_session_tmp_qwerty - v1" duration=28.64589ms grafana | logger=migrator t=2025-06-18T15:20:35.15062162Z level=info msg="Executing migration" id="create cloud_migration_session v2" grafana | logger=migrator t=2025-06-18T15:20:35.152110335Z level=info msg="Migration successfully executed" id="create cloud_migration_session v2" duration=1.486615ms grafana | logger=migrator t=2025-06-18T15:20:35.158495684Z level=info msg="Executing migration" id="create index UQE_cloud_migration_session_uid - v2" grafana | logger=migrator t=2025-06-18T15:20:35.160429323Z level=info msg="Migration successfully executed" id="create index UQE_cloud_migration_session_uid - v2" duration=1.933169ms grafana | logger=migrator t=2025-06-18T15:20:35.16534326Z level=info msg="Executing migration" id="copy cloud_migration_session v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:35.165743903Z level=info msg="Migration successfully executed" id="copy cloud_migration_session v1 to v2" duration=400.203µs grafana | logger=migrator t=2025-06-18T15:20:35.170423447Z level=info msg="Executing migration" id="drop cloud_migration_session_tmp_qwerty" grafana | logger=migrator t=2025-06-18T15:20:35.17184423Z level=info msg="Migration successfully executed" id="drop cloud_migration_session_tmp_qwerty" duration=1.420213ms grafana | logger=migrator t=2025-06-18T15:20:35.177580864Z level=info msg="Executing migration" id="Rename table cloud_migration_run to cloud_migration_snapshot_tmp_qwerty - v1" grafana | logger=migrator t=2025-06-18T15:20:35.204626799Z level=info msg="Migration successfully executed" id="Rename table cloud_migration_run to cloud_migration_snapshot_tmp_qwerty - v1" duration=27.043465ms grafana | logger=migrator t=2025-06-18T15:20:35.214959536Z level=info msg="Executing migration" id="create cloud_migration_snapshot v2" grafana | logger=migrator t=2025-06-18T15:20:35.216756753Z level=info msg="Migration successfully executed" id="create cloud_migration_snapshot v2" duration=1.797587ms grafana | logger=migrator t=2025-06-18T15:20:35.235813063Z level=info msg="Executing migration" id="create index UQE_cloud_migration_snapshot_uid - v2" grafana | logger=migrator t=2025-06-18T15:20:35.237871233Z level=info msg="Migration successfully executed" id="create index UQE_cloud_migration_snapshot_uid - v2" duration=2.05786ms grafana | logger=migrator t=2025-06-18T15:20:35.256434667Z level=info msg="Executing migration" id="copy cloud_migration_snapshot v1 to v2" grafana | logger=migrator t=2025-06-18T15:20:35.257136263Z level=info msg="Migration successfully executed" id="copy cloud_migration_snapshot v1 to v2" duration=702.126µs grafana | logger=migrator t=2025-06-18T15:20:35.268659813Z level=info msg="Executing migration" id="drop cloud_migration_snapshot_tmp_qwerty" grafana | logger=migrator t=2025-06-18T15:20:35.270490249Z level=info msg="Migration successfully executed" id="drop cloud_migration_snapshot_tmp_qwerty" duration=1.825616ms grafana | logger=migrator t=2025-06-18T15:20:35.335598222Z level=info msg="Executing migration" id="add snapshot upload_url column" grafana | logger=migrator t=2025-06-18T15:20:35.349984068Z level=info msg="Migration successfully executed" id="add snapshot upload_url column" duration=14.387916ms grafana | logger=migrator t=2025-06-18T15:20:35.376265705Z level=info msg="Executing migration" id="add snapshot status column" grafana | logger=migrator t=2025-06-18T15:20:35.388201168Z level=info msg="Migration successfully executed" id="add snapshot status column" duration=11.936523ms grafana | logger=migrator t=2025-06-18T15:20:35.401232811Z level=info msg="Executing migration" id="add snapshot local_directory column" grafana | logger=migrator t=2025-06-18T15:20:35.415975929Z level=info msg="Migration successfully executed" id="add snapshot local_directory column" duration=14.743198ms grafana | logger=migrator t=2025-06-18T15:20:35.482025182Z level=info msg="Executing migration" id="add snapshot gms_snapshot_uid column" grafana | logger=migrator t=2025-06-18T15:20:35.495901022Z level=info msg="Migration successfully executed" id="add snapshot gms_snapshot_uid column" duration=13.8775ms grafana | logger=migrator t=2025-06-18T15:20:35.521918577Z level=info msg="Executing migration" id="add snapshot encryption_key column" grafana | logger=migrator t=2025-06-18T15:20:35.533639887Z level=info msg="Migration successfully executed" id="add snapshot encryption_key column" duration=11.72175ms grafana | logger=migrator t=2025-06-18T15:20:35.562149386Z level=info msg="Executing migration" id="add snapshot error_string column" grafana | logger=migrator t=2025-06-18T15:20:35.576680602Z level=info msg="Migration successfully executed" id="add snapshot error_string column" duration=14.529317ms grafana | logger=migrator t=2025-06-18T15:20:35.580394558Z level=info msg="Executing migration" id="create cloud_migration_resource table v1" grafana | logger=migrator t=2025-06-18T15:20:35.581148095Z level=info msg="Migration successfully executed" id="create cloud_migration_resource table v1" duration=753.597µs grafana | logger=migrator t=2025-06-18T15:20:35.587279862Z level=info msg="Executing migration" id="delete cloud_migration_snapshot.result column" grafana | logger=migrator t=2025-06-18T15:20:35.626615143Z level=info msg="Migration successfully executed" id="delete cloud_migration_snapshot.result column" duration=39.334771ms grafana | logger=migrator t=2025-06-18T15:20:35.651413327Z level=info msg="Executing migration" id="add cloud_migration_resource.name column" grafana | logger=migrator t=2025-06-18T15:20:35.664205558Z level=info msg="Migration successfully executed" id="add cloud_migration_resource.name column" duration=12.792631ms grafana | logger=migrator t=2025-06-18T15:20:35.679285349Z level=info msg="Executing migration" id="add cloud_migration_resource.parent_name column" grafana | logger=migrator t=2025-06-18T15:20:35.68784922Z level=info msg="Migration successfully executed" id="add cloud_migration_resource.parent_name column" duration=8.562701ms grafana | logger=migrator t=2025-06-18T15:20:35.699969844Z level=info msg="Executing migration" id="add cloud_migration_session.org_id column" grafana | logger=migrator t=2025-06-18T15:20:35.710552004Z level=info msg="Migration successfully executed" id="add cloud_migration_session.org_id column" duration=10.585589ms grafana | logger=migrator t=2025-06-18T15:20:35.722009331Z level=info msg="Executing migration" id="add cloud_migration_resource.error_code column" grafana | logger=migrator t=2025-06-18T15:20:35.732261548Z level=info msg="Migration successfully executed" id="add cloud_migration_resource.error_code column" duration=10.251567ms grafana | logger=migrator t=2025-06-18T15:20:35.737654209Z level=info msg="Executing migration" id="increase resource_uid column length" grafana | logger=migrator t=2025-06-18T15:20:35.737868041Z level=info msg="Migration successfully executed" id="increase resource_uid column length" duration=216.072µs grafana | logger=migrator t=2025-06-18T15:20:35.74305898Z level=info msg="Executing migration" id="alter kv_store.value to longtext" grafana | logger=migrator t=2025-06-18T15:20:35.743214661Z level=info msg="Migration successfully executed" id="alter kv_store.value to longtext" duration=156.311µs grafana | logger=migrator t=2025-06-18T15:20:35.74837025Z level=info msg="Executing migration" id="add notification_settings column to alert_rule table" grafana | logger=migrator t=2025-06-18T15:20:35.75890687Z level=info msg="Migration successfully executed" id="add notification_settings column to alert_rule table" duration=10.53576ms grafana | logger=migrator t=2025-06-18T15:20:35.764888736Z level=info msg="Executing migration" id="add notification_settings column to alert_rule_version table" grafana | logger=migrator t=2025-06-18T15:20:35.775344924Z level=info msg="Migration successfully executed" id="add notification_settings column to alert_rule_version table" duration=10.455237ms grafana | logger=migrator t=2025-06-18T15:20:35.780776755Z level=info msg="Executing migration" id="removing scope from alert.instances:read action migration" grafana | logger=migrator t=2025-06-18T15:20:35.78128747Z level=info msg="Migration successfully executed" id="removing scope from alert.instances:read action migration" duration=510.295µs grafana | logger=migrator t=2025-06-18T15:20:35.807940231Z level=info msg="Executing migration" id="managed folder permissions alerting silences actions migration" grafana | logger=migrator t=2025-06-18T15:20:35.808376565Z level=info msg="Migration successfully executed" id="managed folder permissions alerting silences actions migration" duration=437.784µs grafana | logger=migrator t=2025-06-18T15:20:35.813945637Z level=info msg="Executing migration" id="add record column to alert_rule table" grafana | logger=migrator t=2025-06-18T15:20:35.825063992Z level=info msg="Migration successfully executed" id="add record column to alert_rule table" duration=11.118235ms grafana | logger=migrator t=2025-06-18T15:20:35.845629345Z level=info msg="Executing migration" id="add record column to alert_rule_version table" grafana | logger=migrator t=2025-06-18T15:20:35.859398715Z level=info msg="Migration successfully executed" id="add record column to alert_rule_version table" duration=13.77088ms grafana | logger=migrator t=2025-06-18T15:20:35.865309962Z level=info msg="Executing migration" id="add resolved_at column to alert_instance table" grafana | logger=migrator t=2025-06-18T15:20:35.876243934Z level=info msg="Migration successfully executed" id="add resolved_at column to alert_instance table" duration=10.933342ms grafana | logger=migrator t=2025-06-18T15:20:35.88432983Z level=info msg="Executing migration" id="add last_sent_at column to alert_instance table" grafana | logger=migrator t=2025-06-18T15:20:35.89602444Z level=info msg="Migration successfully executed" id="add last_sent_at column to alert_instance table" duration=11.6912ms grafana | logger=migrator t=2025-06-18T15:20:35.919466271Z level=info msg="Executing migration" id="Add scope to alert.notifications.receivers:read and alert.notifications.receivers.secrets:read" grafana | logger=migrator t=2025-06-18T15:20:35.920681972Z level=info msg="Migration successfully executed" id="Add scope to alert.notifications.receivers:read and alert.notifications.receivers.secrets:read" duration=1.214981ms grafana | logger=migrator t=2025-06-18T15:20:35.932603995Z level=info msg="Executing migration" id="add metadata column to alert_rule table" grafana | logger=migrator t=2025-06-18T15:20:35.943051113Z level=info msg="Migration successfully executed" id="add metadata column to alert_rule table" duration=10.448778ms grafana | logger=migrator t=2025-06-18T15:20:35.989021366Z level=info msg="Executing migration" id="add metadata column to alert_rule_version table" grafana | logger=migrator t=2025-06-18T15:20:36.003112139Z level=info msg="Migration successfully executed" id="add metadata column to alert_rule_version table" duration=14.092763ms grafana | logger=migrator t=2025-06-18T15:20:36.014560266Z level=info msg="Executing migration" id="delete orphaned service account permissions" grafana | logger=migrator t=2025-06-18T15:20:36.01494708Z level=info msg="Migration successfully executed" id="delete orphaned service account permissions" duration=386.454µs grafana | logger=migrator t=2025-06-18T15:20:36.032070111Z level=info msg="Executing migration" id="adding action set permissions" grafana | logger=migrator t=2025-06-18T15:20:36.032875198Z level=info msg="Migration successfully executed" id="adding action set permissions" duration=805.407µs grafana | logger=migrator t=2025-06-18T15:20:36.047137782Z level=info msg="Executing migration" id="create user_external_session table" grafana | logger=migrator t=2025-06-18T15:20:36.049342233Z level=info msg="Migration successfully executed" id="create user_external_session table" duration=2.207981ms grafana | logger=migrator t=2025-06-18T15:20:36.059312548Z level=info msg="Executing migration" id="increase name_id column length to 1024" grafana | logger=migrator t=2025-06-18T15:20:36.059331688Z level=info msg="Migration successfully executed" id="increase name_id column length to 1024" duration=20.09µs grafana | logger=migrator t=2025-06-18T15:20:36.094523249Z level=info msg="Executing migration" id="increase session_id column length to 1024" grafana | logger=migrator t=2025-06-18T15:20:36.094548169Z level=info msg="Migration successfully executed" id="increase session_id column length to 1024" duration=26.77µs grafana | logger=migrator t=2025-06-18T15:20:36.110255637Z level=info msg="Executing migration" id="remove scope from alert.notifications.receivers:create" grafana | logger=migrator t=2025-06-18T15:20:36.113042904Z level=info msg="Migration successfully executed" id="remove scope from alert.notifications.receivers:create" duration=2.796246ms grafana | logger=migrator t=2025-06-18T15:20:36.118177712Z level=info msg="Executing migration" id="add created_by column to alert_rule_version table" grafana | logger=migrator t=2025-06-18T15:20:36.132020412Z level=info msg="Migration successfully executed" id="add created_by column to alert_rule_version table" duration=13.84213ms grafana | logger=migrator t=2025-06-18T15:20:36.15198119Z level=info msg="Executing migration" id="add updated_by column to alert_rule table" grafana | logger=migrator t=2025-06-18T15:20:36.164691169Z level=info msg="Migration successfully executed" id="add updated_by column to alert_rule table" duration=12.708239ms grafana | logger=migrator t=2025-06-18T15:20:36.172674664Z level=info msg="Executing migration" id="add alert_rule_state table" grafana | logger=migrator t=2025-06-18T15:20:36.173759755Z level=info msg="Migration successfully executed" id="add alert_rule_state table" duration=1.084681ms grafana | logger=migrator t=2025-06-18T15:20:36.194819163Z level=info msg="Executing migration" id="add index to alert_rule_state on org_id and rule_uid columns" grafana | logger=migrator t=2025-06-18T15:20:36.197231006Z level=info msg="Migration successfully executed" id="add index to alert_rule_state on org_id and rule_uid columns" duration=2.410393ms grafana | logger=migrator t=2025-06-18T15:20:36.206019888Z level=info msg="Executing migration" id="add guid column to alert_rule table" grafana | logger=migrator t=2025-06-18T15:20:36.216173994Z level=info msg="Migration successfully executed" id="add guid column to alert_rule table" duration=10.154816ms grafana | logger=migrator t=2025-06-18T15:20:36.222465023Z level=info msg="Executing migration" id="add rule_guid column to alert_rule_version table" grafana | logger=migrator t=2025-06-18T15:20:36.230271117Z level=info msg="Migration successfully executed" id="add rule_guid column to alert_rule_version table" duration=7.804854ms grafana | logger=migrator t=2025-06-18T15:20:36.234230094Z level=info msg="Executing migration" id="cleanup alert_rule_version table" grafana | logger=migrator t=2025-06-18T15:20:36.234253074Z level=info msg="Rule version record limit is not set, fallback to 100" limit=0 grafana | logger=migrator t=2025-06-18T15:20:36.234438756Z level=info msg="Cleaning up table `alert_rule_version`" batchSize=50 batches=0 keepVersions=100 grafana | logger=migrator t=2025-06-18T15:20:36.234455986Z level=info msg="Migration successfully executed" id="cleanup alert_rule_version table" duration=226.152µs grafana | logger=migrator t=2025-06-18T15:20:36.241852286Z level=info msg="Executing migration" id="populate rule guid in alert rule table" grafana | logger=migrator t=2025-06-18T15:20:36.242420021Z level=info msg="Migration successfully executed" id="populate rule guid in alert rule table" duration=567.555µs grafana | logger=migrator t=2025-06-18T15:20:36.246042005Z level=info msg="Executing migration" id="drop index in alert_rule_version table on rule_org_id, rule_uid and version columns" grafana | logger=migrator t=2025-06-18T15:20:36.247211827Z level=info msg="Migration successfully executed" id="drop index in alert_rule_version table on rule_org_id, rule_uid and version columns" duration=1.169072ms grafana | logger=migrator t=2025-06-18T15:20:36.265014994Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_org_id, rule_uid, rule_guid and version columns" grafana | logger=migrator t=2025-06-18T15:20:36.266572809Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_org_id, rule_uid, rule_guid and version columns" duration=1.557175ms grafana | logger=migrator t=2025-06-18T15:20:36.284496277Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_guid and version columns" grafana | logger=migrator t=2025-06-18T15:20:36.28572374Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_guid and version columns" duration=1.229222ms grafana | logger=migrator t=2025-06-18T15:20:36.298722191Z level=info msg="Executing migration" id="add index in alert_rule table on guid columns" grafana | logger=migrator t=2025-06-18T15:20:36.300795411Z level=info msg="Migration successfully executed" id="add index in alert_rule table on guid columns" duration=2.07283ms grafana | logger=migrator t=2025-06-18T15:20:36.370077933Z level=info msg="Executing migration" id="add keep_firing_for column to alert_rule" grafana | logger=migrator t=2025-06-18T15:20:36.383124875Z level=info msg="Migration successfully executed" id="add keep_firing_for column to alert_rule" duration=13.049552ms grafana | logger=migrator t=2025-06-18T15:20:36.407578106Z level=info msg="Executing migration" id="add keep_firing_for column to alert_rule_version" grafana | logger=migrator t=2025-06-18T15:20:36.41545673Z level=info msg="Migration successfully executed" id="add keep_firing_for column to alert_rule_version" duration=7.879544ms grafana | logger=migrator t=2025-06-18T15:20:36.45046976Z level=info msg="Executing migration" id="add missing_series_evals_to_resolve column to alert_rule" grafana | logger=migrator t=2025-06-18T15:20:36.45789367Z level=info msg="Migration successfully executed" id="add missing_series_evals_to_resolve column to alert_rule" duration=7.42634ms grafana | logger=migrator t=2025-06-18T15:20:36.533709513Z level=info msg="Executing migration" id="add missing_series_evals_to_resolve column to alert_rule_version" grafana | logger=migrator t=2025-06-18T15:20:36.547384442Z level=info msg="Migration successfully executed" id="add missing_series_evals_to_resolve column to alert_rule_version" duration=13.674639ms grafana | logger=migrator t=2025-06-18T15:20:36.571141185Z level=info msg="Executing migration" id="remove the datasources:drilldown action" grafana | logger=migrator t=2025-06-18T15:20:36.571665821Z level=info msg="Removed 0 datasources:drilldown permissions" grafana | logger=migrator t=2025-06-18T15:20:36.571697051Z level=info msg="Migration successfully executed" id="remove the datasources:drilldown action" duration=557.246µs grafana | logger=migrator t=2025-06-18T15:20:36.593271353Z level=info msg="Executing migration" id="remove title in folder unique index" grafana | logger=migrator t=2025-06-18T15:20:36.595592195Z level=info msg="Migration successfully executed" id="remove title in folder unique index" duration=2.321072ms grafana | logger=migrator t=2025-06-18T15:20:36.633015039Z level=info msg="migrations completed" performed=654 skipped=0 duration=10.245881899s grafana | logger=migrator t=2025-06-18T15:20:36.633744595Z level=info msg="Unlocking database" grafana | logger=sqlstore t=2025-06-18T15:20:36.652083387Z level=info msg="Created default admin" user=admin grafana | logger=sqlstore t=2025-06-18T15:20:36.65234355Z level=info msg="Created default organization" grafana | logger=secrets t=2025-06-18T15:20:36.717917148Z level=info msg="Envelope encryption state" enabled=true currentprovider=secretKey.v1 grafana | logger=plugin.angulardetectorsprovider.dynamic t=2025-06-18T15:20:36.794782281Z level=info msg="Restored cache from database" duration=455.805µs grafana | logger=resource-migrator t=2025-06-18T15:20:36.803122349Z level=info msg="Locking database" grafana | logger=resource-migrator t=2025-06-18T15:20:36.803136689Z level=info msg="Starting DB migrations" grafana | logger=resource-migrator t=2025-06-18T15:20:36.810905653Z level=info msg="Executing migration" id="create resource_migration_log table" grafana | logger=resource-migrator t=2025-06-18T15:20:36.811637759Z level=info msg="Migration successfully executed" id="create resource_migration_log table" duration=731.976µs grafana | logger=resource-migrator t=2025-06-18T15:20:36.873680493Z level=info msg="Executing migration" id="Initialize resource tables" grafana | logger=resource-migrator t=2025-06-18T15:20:36.873724834Z level=info msg="Migration successfully executed" id="Initialize resource tables" duration=47.431µs grafana | logger=resource-migrator t=2025-06-18T15:20:36.881373696Z level=info msg="Executing migration" id="drop table resource" grafana | logger=resource-migrator t=2025-06-18T15:20:36.881706529Z level=info msg="Migration successfully executed" id="drop table resource" duration=332.943µs grafana | logger=resource-migrator t=2025-06-18T15:20:36.92532226Z level=info msg="Executing migration" id="create table resource" grafana | logger=resource-migrator t=2025-06-18T15:20:36.92746581Z level=info msg="Migration successfully executed" id="create table resource" duration=2.13892ms grafana | logger=resource-migrator t=2025-06-18T15:20:36.964715191Z level=info msg="Executing migration" id="create table resource, index: 0" grafana | logger=resource-migrator t=2025-06-18T15:20:36.966816611Z level=info msg="Migration successfully executed" id="create table resource, index: 0" duration=2.10216ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.071866279Z level=info msg="Executing migration" id="drop table resource_history" grafana | logger=resource-migrator t=2025-06-18T15:20:37.07198515Z level=info msg="Migration successfully executed" id="drop table resource_history" duration=121.671µs grafana | logger=resource-migrator t=2025-06-18T15:20:37.099736191Z level=info msg="Executing migration" id="create table resource_history" grafana | logger=resource-migrator t=2025-06-18T15:20:37.100726371Z level=info msg="Migration successfully executed" id="create table resource_history" duration=990.62µs grafana | logger=resource-migrator t=2025-06-18T15:20:37.129397361Z level=info msg="Executing migration" id="create table resource_history, index: 0" grafana | logger=resource-migrator t=2025-06-18T15:20:37.13035684Z level=info msg="Migration successfully executed" id="create table resource_history, index: 0" duration=959.089µs grafana | logger=resource-migrator t=2025-06-18T15:20:37.145864476Z level=info msg="Executing migration" id="create table resource_history, index: 1" grafana | logger=resource-migrator t=2025-06-18T15:20:37.147747123Z level=info msg="Migration successfully executed" id="create table resource_history, index: 1" duration=1.881347ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.19414094Z level=info msg="Executing migration" id="drop table resource_version" grafana | logger=resource-migrator t=2025-06-18T15:20:37.194397882Z level=info msg="Migration successfully executed" id="drop table resource_version" duration=257.872µs grafana | logger=resource-migrator t=2025-06-18T15:20:37.20589009Z level=info msg="Executing migration" id="create table resource_version" grafana | logger=resource-migrator t=2025-06-18T15:20:37.207431775Z level=info msg="Migration successfully executed" id="create table resource_version" duration=1.541335ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.215432689Z level=info msg="Executing migration" id="create table resource_version, index: 0" grafana | logger=resource-migrator t=2025-06-18T15:20:37.216617281Z level=info msg="Migration successfully executed" id="create table resource_version, index: 0" duration=1.184262ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.228949457Z level=info msg="Executing migration" id="drop table resource_blob" grafana | logger=resource-migrator t=2025-06-18T15:20:37.229053058Z level=info msg="Migration successfully executed" id="drop table resource_blob" duration=102.621µs grafana | logger=resource-migrator t=2025-06-18T15:20:37.254083524Z level=info msg="Executing migration" id="create table resource_blob" grafana | logger=resource-migrator t=2025-06-18T15:20:37.256176564Z level=info msg="Migration successfully executed" id="create table resource_blob" duration=2.09332ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.287121064Z level=info msg="Executing migration" id="create table resource_blob, index: 0" grafana | logger=resource-migrator t=2025-06-18T15:20:37.289147764Z level=info msg="Migration successfully executed" id="create table resource_blob, index: 0" duration=2.02614ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.310608925Z level=info msg="Executing migration" id="create table resource_blob, index: 1" grafana | logger=resource-migrator t=2025-06-18T15:20:37.312686685Z level=info msg="Migration successfully executed" id="create table resource_blob, index: 1" duration=2.07666ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.374021752Z level=info msg="Executing migration" id="Add column previous_resource_version in resource_history" grafana | logger=resource-migrator t=2025-06-18T15:20:37.386967215Z level=info msg="Migration successfully executed" id="Add column previous_resource_version in resource_history" duration=12.947223ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.402585211Z level=info msg="Executing migration" id="Add column previous_resource_version in resource" grafana | logger=resource-migrator t=2025-06-18T15:20:37.416709264Z level=info msg="Migration successfully executed" id="Add column previous_resource_version in resource" duration=14.124813ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.440298606Z level=info msg="Executing migration" id="Add index to resource_history for polling" grafana | logger=resource-migrator t=2025-06-18T15:20:37.442643869Z level=info msg="Migration successfully executed" id="Add index to resource_history for polling" duration=2.345453ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.456888163Z level=info msg="Executing migration" id="Add index to resource for loading" grafana | logger=resource-migrator t=2025-06-18T15:20:37.459048442Z level=info msg="Migration successfully executed" id="Add index to resource for loading" duration=2.152939ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.498814157Z level=info msg="Executing migration" id="Add column folder in resource_history" grafana | logger=resource-migrator t=2025-06-18T15:20:37.512121962Z level=info msg="Migration successfully executed" id="Add column folder in resource_history" duration=13.309845ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.574788102Z level=info msg="Executing migration" id="Add column folder in resource" grafana | logger=resource-migrator t=2025-06-18T15:20:37.586871895Z level=info msg="Migration successfully executed" id="Add column folder in resource" duration=12.084603ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.642468578Z level=info msg="Executing migration" id="Migrate DeletionMarkers to real Resource objects" grafana | logger=deletion-marker-migrator t=2025-06-18T15:20:37.642516609Z level=info msg="finding any deletion markers" grafana | logger=resource-migrator t=2025-06-18T15:20:37.643267846Z level=info msg="Migration successfully executed" id="Migrate DeletionMarkers to real Resource objects" duration=800.588µs grafana | logger=resource-migrator t=2025-06-18T15:20:37.696197634Z level=info msg="Executing migration" id="Add index to resource_history for get trash" grafana | logger=resource-migrator t=2025-06-18T15:20:37.698022511Z level=info msg="Migration successfully executed" id="Add index to resource_history for get trash" duration=1.823507ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.777510979Z level=info msg="Executing migration" id="Add generation to resource history" grafana | logger=resource-migrator t=2025-06-18T15:20:37.791994225Z level=info msg="Migration successfully executed" id="Add generation to resource history" duration=14.476826ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.854023639Z level=info msg="Executing migration" id="Add generation index to resource history" grafana | logger=resource-migrator t=2025-06-18T15:20:37.856374961Z level=info msg="Migration successfully executed" id="Add generation index to resource history" duration=2.352172ms grafana | logger=resource-migrator t=2025-06-18T15:20:37.903146451Z level=info msg="migrations completed" performed=26 skipped=0 duration=1.092277979s grafana | logger=resource-migrator t=2025-06-18T15:20:37.904025469Z level=info msg="Unlocking database" grafana | t=2025-06-18T15:20:37.904528984Z level=info caller=logger.go:214 time=2025-06-18T15:20:37.904508604Z msg="Using channel notifier" logger=sql-resource-server grafana | logger=plugin.store t=2025-06-18T15:20:37.920599515Z level=info msg="Loading plugins..." grafana | logger=plugins.registration t=2025-06-18T15:20:37.953928319Z level=error msg="Could not register plugin" pluginId=table error="plugin table is already registered" grafana | logger=plugins.initialization t=2025-06-18T15:20:37.953957639Z level=error msg="Could not initialize plugin" pluginId=table error="plugin table is already registered" grafana | logger=plugin.store t=2025-06-18T15:20:37.95407058Z level=info msg="Plugins loaded" count=53 duration=33.469655ms grafana | logger=query_data t=2025-06-18T15:20:37.959959916Z level=info msg="Query Service initialization" grafana | logger=live.push_http t=2025-06-18T15:20:37.967719189Z level=info msg="Live Push Gateway initialization" grafana | logger=ngalert.notifier.alertmanager org=1 t=2025-06-18T15:20:37.980683731Z level=info msg="Applying new configuration to Alertmanager" configHash=d2c56faca6af2a5772ff4253222f7386 grafana | logger=ngalert t=2025-06-18T15:20:38.029829563Z level=info msg="Using simple database alert instance store" grafana | logger=ngalert.state.manager.persist t=2025-06-18T15:20:38.029867173Z level=info msg="Using sync state persister" grafana | logger=infra.usagestats.collector t=2025-06-18T15:20:38.03366171Z level=info msg="registering usage stat providers" usageStatsProvidersLen=2 grafana | logger=plugin.backgroundinstaller t=2025-06-18T15:20:38.034407367Z level=info msg="Installing plugin" pluginId=grafana-lokiexplore-app version= grafana | logger=ngalert.state.manager t=2025-06-18T15:20:38.034463197Z level=info msg="Warming state cache for startup" grafana | logger=grafanaStorageLogger t=2025-06-18T15:20:38.034956242Z level=info msg="Storage starting" grafana | logger=http.server t=2025-06-18T15:20:38.036720168Z level=info msg="HTTP Server Listen" address=[::]:3000 protocol=http subUrl= socket= grafana | logger=ngalert.multiorg.alertmanager t=2025-06-18T15:20:38.037204273Z level=info msg="Starting MultiOrg Alertmanager" grafana | logger=sqlstore.transactions t=2025-06-18T15:20:38.049906991Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=0 grafana | logger=plugins.update.checker t=2025-06-18T15:20:38.134480908Z level=info msg="Update check succeeded" duration=96.68507ms grafana | logger=grafana.update.checker t=2025-06-18T15:20:38.13469959Z level=info msg="Update check succeeded" duration=97.270995ms grafana | logger=ngalert.state.manager t=2025-06-18T15:20:38.688672139Z level=info msg="State cache has been initialized" states=0 duration=654.207512ms grafana | logger=ngalert.scheduler t=2025-06-18T15:20:38.6887423Z level=info msg="Starting scheduler" tickInterval=10s maxAttempts=3 grafana | logger=ticker t=2025-06-18T15:20:38.688824101Z level=info msg=starting first_tick=2025-06-18T15:20:40Z grafana | logger=provisioning.datasources t=2025-06-18T15:20:38.692064392Z level=info msg="inserting datasource from configuration" name=PolicyPrometheus uid=dkSf71fnz grafana | logger=provisioning.alerting t=2025-06-18T15:20:38.793791589Z level=info msg="starting to provision alerting" grafana | logger=provisioning.alerting t=2025-06-18T15:20:38.793828499Z level=info msg="finished to provision alerting" grafana | logger=provisioning.dashboard t=2025-06-18T15:20:38.796519235Z level=info msg="starting to provision dashboards" grafana | logger=plugin.angulardetectorsprovider.dynamic t=2025-06-18T15:20:38.897471884Z level=info msg="Patterns update finished" duration=209.073706ms grafana | logger=plugin.installer t=2025-06-18T15:20:39.05875065Z level=info msg="Installing plugin" pluginId=grafana-lokiexplore-app version= grafana | logger=installer.fs t=2025-06-18T15:20:39.191115815Z level=info msg="Downloaded and extracted grafana-lokiexplore-app v1.0.18 zip successfully to /var/lib/grafana/plugins/grafana-lokiexplore-app" grafana | logger=grafana-apiserver t=2025-06-18T15:20:39.216799756Z level=info msg="Adding GroupVersion playlist.grafana.app v0alpha1 to ResourceManager" grafana | logger=grafana-apiserver t=2025-06-18T15:20:39.2204303Z level=info msg="Adding GroupVersion dashboard.grafana.app v1beta1 to ResourceManager" grafana | logger=grafana-apiserver t=2025-06-18T15:20:39.221124637Z level=info msg="Adding GroupVersion dashboard.grafana.app v0alpha1 to ResourceManager" grafana | logger=grafana-apiserver t=2025-06-18T15:20:39.221729353Z level=info msg="Adding GroupVersion dashboard.grafana.app v2alpha1 to ResourceManager" grafana | logger=grafana-apiserver t=2025-06-18T15:20:39.222344249Z level=info msg="Adding GroupVersion featuretoggle.grafana.app v0alpha1 to ResourceManager" grafana | logger=grafana-apiserver t=2025-06-18T15:20:39.223418208Z level=info msg="Adding GroupVersion folder.grafana.app v1beta1 to ResourceManager" grafana | logger=grafana-apiserver t=2025-06-18T15:20:39.224990923Z level=info msg="Adding GroupVersion iam.grafana.app v0alpha1 to ResourceManager" grafana | logger=grafana-apiserver t=2025-06-18T15:20:39.227938491Z level=info msg="Adding GroupVersion notifications.alerting.grafana.app v0alpha1 to ResourceManager" grafana | logger=grafana-apiserver t=2025-06-18T15:20:39.228793199Z level=info msg="Adding GroupVersion userstorage.grafana.app v0alpha1 to ResourceManager" grafana | logger=plugins.registration t=2025-06-18T15:20:39.244717859Z level=info msg="Plugin registered" pluginId=grafana-lokiexplore-app grafana | logger=plugin.backgroundinstaller t=2025-06-18T15:20:39.244752279Z level=info msg="Plugin successfully installed" pluginId=grafana-lokiexplore-app version= duration=1.210313242s grafana | logger=plugin.backgroundinstaller t=2025-06-18T15:20:39.244775959Z level=info msg="Installing plugin" pluginId=grafana-pyroscope-app version= grafana | logger=app-registry t=2025-06-18T15:20:39.277320646Z level=info msg="app registry initialized" grafana | logger=plugin.installer t=2025-06-18T15:20:39.747222303Z level=info msg="Installing plugin" pluginId=grafana-pyroscope-app version= grafana | logger=installer.fs t=2025-06-18T15:20:39.816819537Z level=info msg="Downloaded and extracted grafana-pyroscope-app v1.4.1 zip successfully to /var/lib/grafana/plugins/grafana-pyroscope-app" grafana | logger=plugins.registration t=2025-06-18T15:20:39.875548069Z level=info msg="Plugin registered" pluginId=grafana-pyroscope-app grafana | logger=plugin.backgroundinstaller t=2025-06-18T15:20:39.87558406Z level=info msg="Plugin successfully installed" pluginId=grafana-pyroscope-app version= duration=630.80061ms grafana | logger=plugin.backgroundinstaller t=2025-06-18T15:20:39.87561165Z level=info msg="Installing plugin" pluginId=grafana-exploretraces-app version= grafana | logger=plugin.installer t=2025-06-18T15:20:40.134602575Z level=info msg="Installing plugin" pluginId=grafana-exploretraces-app version= grafana | logger=installer.fs t=2025-06-18T15:20:40.251631304Z level=info msg="Downloaded and extracted grafana-exploretraces-app v1.0.0 zip successfully to /var/lib/grafana/plugins/grafana-exploretraces-app" grafana | logger=plugins.registration t=2025-06-18T15:20:40.268817465Z level=info msg="Plugin registered" pluginId=grafana-exploretraces-app grafana | logger=plugin.backgroundinstaller t=2025-06-18T15:20:40.268843585Z level=info msg="Plugin successfully installed" pluginId=grafana-exploretraces-app version= duration=393.227455ms grafana | logger=plugin.backgroundinstaller t=2025-06-18T15:20:40.268868045Z level=info msg="Installing plugin" pluginId=grafana-metricsdrilldown-app version= grafana | logger=provisioning.dashboard t=2025-06-18T15:20:40.384464313Z level=info msg="finished to provision dashboards" grafana | logger=plugin.installer t=2025-06-18T15:20:40.638472739Z level=info msg="Installing plugin" pluginId=grafana-metricsdrilldown-app version= grafana | logger=installer.fs t=2025-06-18T15:20:40.700244269Z level=info msg="Downloaded and extracted grafana-metricsdrilldown-app v1.0.3 zip successfully to /var/lib/grafana/plugins/grafana-metricsdrilldown-app" grafana | logger=plugins.registration t=2025-06-18T15:20:40.720370039Z level=info msg="Plugin registered" pluginId=grafana-metricsdrilldown-app grafana | logger=plugin.backgroundinstaller t=2025-06-18T15:20:40.720390379Z level=info msg="Plugin successfully installed" pluginId=grafana-metricsdrilldown-app version= duration=451.517774ms grafana | logger=infra.usagestats t=2025-06-18T15:21:44.042841272Z level=info msg="Usage stats are ready to report" kafka | ===> User kafka | uid=1000(appuser) gid=1000(appuser) groups=1000(appuser) kafka | ===> Configuring ... kafka | Running in Zookeeper mode... kafka | ===> Running preflight checks ... kafka | ===> Check if /var/lib/kafka/data is writable ... kafka | ===> Check if Zookeeper is healthy ... kafka | [2025-06-18 15:20:31,459] INFO Client environment:zookeeper.version=3.8.4-9316c2a7a97e1666d8f4593f34dd6fc36ecc436c, built on 2024-02-12 22:16 UTC (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,460] INFO Client environment:host.name=kafka (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,460] INFO Client environment:java.version=11.0.26 (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,460] INFO Client environment:java.vendor=Azul Systems, Inc. (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,460] INFO Client environment:java.home=/usr/lib/jvm/java-11-zulu-openjdk-ca (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,460] INFO Client environment:java.class.path=/usr/share/java/cp-base-new/kafka-storage-api-7.4.9-ccs.jar:/usr/share/java/cp-base-new/scala-logging_2.13-3.9.4.jar:/usr/share/java/cp-base-new/jackson-datatype-jdk8-2.14.2.jar:/usr/share/java/cp-base-new/logredactor-1.0.12.jar:/usr/share/java/cp-base-new/jolokia-core-1.7.1.jar:/usr/share/java/cp-base-new/re2j-1.6.jar:/usr/share/java/cp-base-new/kafka-server-common-7.4.9-ccs.jar:/usr/share/java/cp-base-new/scala-library-2.13.10.jar:/usr/share/java/cp-base-new/commons-cli-1.4.jar:/usr/share/java/cp-base-new/slf4j-reload4j-1.7.36.jar:/usr/share/java/cp-base-new/jackson-annotations-2.14.2.jar:/usr/share/java/cp-base-new/json-simple-1.1.1.jar:/usr/share/java/cp-base-new/kafka-clients-7.4.9-ccs.jar:/usr/share/java/cp-base-new/jackson-module-scala_2.13-2.14.2.jar:/usr/share/java/cp-base-new/kafka-storage-7.4.9-ccs.jar:/usr/share/java/cp-base-new/scala-java8-compat_2.13-1.0.2.jar:/usr/share/java/cp-base-new/kafka-raft-7.4.9-ccs.jar:/usr/share/java/cp-base-new/zookeeper-jute-3.8.4.jar:/usr/share/java/cp-base-new/zstd-jni-1.5.2-1.jar:/usr/share/java/cp-base-new/minimal-json-0.9.5.jar:/usr/share/java/cp-base-new/kafka-group-coordinator-7.4.9-ccs.jar:/usr/share/java/cp-base-new/common-utils-7.4.9.jar:/usr/share/java/cp-base-new/jackson-dataformat-yaml-2.14.2.jar:/usr/share/java/cp-base-new/kafka-metadata-7.4.9-ccs.jar:/usr/share/java/cp-base-new/slf4j-api-1.7.36.jar:/usr/share/java/cp-base-new/paranamer-2.8.jar:/usr/share/java/cp-base-new/jmx_prometheus_javaagent-0.18.0.jar:/usr/share/java/cp-base-new/reload4j-1.2.25.jar:/usr/share/java/cp-base-new/jackson-core-2.14.2.jar:/usr/share/java/cp-base-new/commons-io-2.16.0.jar:/usr/share/java/cp-base-new/argparse4j-0.7.0.jar:/usr/share/java/cp-base-new/audience-annotations-0.12.0.jar:/usr/share/java/cp-base-new/gson-2.9.0.jar:/usr/share/java/cp-base-new/snakeyaml-2.0.jar:/usr/share/java/cp-base-new/zookeeper-3.8.4.jar:/usr/share/java/cp-base-new/kafka_2.13-7.4.9-ccs.jar:/usr/share/java/cp-base-new/jopt-simple-5.0.4.jar:/usr/share/java/cp-base-new/lz4-java-1.8.0.jar:/usr/share/java/cp-base-new/logredactor-metrics-1.0.12.jar:/usr/share/java/cp-base-new/utility-belt-7.4.9-53.jar:/usr/share/java/cp-base-new/scala-reflect-2.13.10.jar:/usr/share/java/cp-base-new/scala-collection-compat_2.13-2.10.0.jar:/usr/share/java/cp-base-new/metrics-core-2.2.0.jar:/usr/share/java/cp-base-new/jackson-dataformat-csv-2.14.2.jar:/usr/share/java/cp-base-new/jolokia-jvm-1.7.1.jar:/usr/share/java/cp-base-new/metrics-core-4.1.12.1.jar:/usr/share/java/cp-base-new/jackson-databind-2.14.2.jar:/usr/share/java/cp-base-new/snappy-java-1.1.10.5.jar:/usr/share/java/cp-base-new/disk-usage-agent-7.4.9.jar:/usr/share/java/cp-base-new/jose4j-0.9.5.jar (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,460] INFO Client environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,460] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,460] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,461] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,461] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,461] INFO Client environment:os.version=4.15.0-192-generic (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,461] INFO Client environment:user.name=appuser (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,461] INFO Client environment:user.home=/home/appuser (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,461] INFO Client environment:user.dir=/home/appuser (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,461] INFO Client environment:os.memory.free=494MB (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,461] INFO Client environment:os.memory.max=8042MB (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,461] INFO Client environment:os.memory.total=504MB (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,464] INFO Initiating client connection, connectString=zookeeper:2181 sessionTimeout=40000 watcher=io.confluent.admin.utils.ZookeeperConnectionWatcher@221af3c0 (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,468] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) kafka | [2025-06-18 15:20:31,473] INFO jute.maxbuffer value is 1048575 Bytes (org.apache.zookeeper.ClientCnxnSocket) kafka | [2025-06-18 15:20:31,480] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) kafka | [2025-06-18 15:20:31,509] INFO Opening socket connection to server zookeeper/172.17.0.2:2181. (org.apache.zookeeper.ClientCnxn) kafka | [2025-06-18 15:20:31,510] INFO SASL config status: Will not attempt to authenticate using SASL (unknown error) (org.apache.zookeeper.ClientCnxn) kafka | [2025-06-18 15:20:31,517] INFO Socket connection established, initiating session, client: /172.17.0.7:34876, server: zookeeper/172.17.0.2:2181 (org.apache.zookeeper.ClientCnxn) kafka | [2025-06-18 15:20:31,546] INFO Session establishment complete on server zookeeper/172.17.0.2:2181, session id = 0x1000002fced0000, negotiated timeout = 40000 (org.apache.zookeeper.ClientCnxn) kafka | [2025-06-18 15:20:31,667] INFO Session: 0x1000002fced0000 closed (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:31,667] INFO EventThread shut down for session: 0x1000002fced0000 (org.apache.zookeeper.ClientCnxn) kafka | Using log4j config /etc/kafka/log4j.properties kafka | ===> Launching ... kafka | ===> Launching kafka ... kafka | [2025-06-18 15:20:32,369] INFO Registered kafka:type=kafka.Log4jController MBean (kafka.utils.Log4jControllerRegistration$) kafka | [2025-06-18 15:20:32,670] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) kafka | [2025-06-18 15:20:32,740] INFO Registered signal handlers for TERM, INT, HUP (org.apache.kafka.common.utils.LoggingSignalHandler) kafka | [2025-06-18 15:20:32,741] INFO starting (kafka.server.KafkaServer) kafka | [2025-06-18 15:20:32,741] INFO Connecting to zookeeper on zookeeper:2181 (kafka.server.KafkaServer) kafka | [2025-06-18 15:20:32,753] INFO [ZooKeeperClient Kafka server] Initializing a new session to zookeeper:2181. (kafka.zookeeper.ZooKeeperClient) kafka | [2025-06-18 15:20:32,757] INFO Client environment:zookeeper.version=3.8.4-9316c2a7a97e1666d8f4593f34dd6fc36ecc436c, built on 2024-02-12 22:16 UTC (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:host.name=kafka (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:java.version=11.0.26 (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:java.vendor=Azul Systems, Inc. (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:java.home=/usr/lib/jvm/java-11-zulu-openjdk-ca (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:java.class.path=/usr/bin/../share/java/kafka/kafka-storage-api-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/scala-logging_2.13-3.9.4.jar:/usr/bin/../share/java/kafka/jersey-common-2.39.1.jar:/usr/bin/../share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/bin/../share/java/kafka/netty-common-4.1.115.Final.jar:/usr/bin/../share/java/kafka/aopalliance-repackaged-2.6.1.jar:/usr/bin/../share/java/kafka/connect-mirror-client-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/connect-json-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/kafka-server-common-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/scala-library-2.13.10.jar:/usr/bin/../share/java/kafka/jackson-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/javax.activation-api-1.2.0.jar:/usr/bin/../share/java/kafka/jetty-util-ajax-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/commons-cli-1.4.jar:/usr/bin/../share/java/kafka/slf4j-reload4j-1.7.36.jar:/usr/bin/../share/java/kafka/kafka-shell-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/reflections-0.9.12.jar:/usr/bin/../share/java/kafka/jline-3.22.0.jar:/usr/bin/../share/java/kafka/jakarta.ws.rs-api-2.1.6.jar:/usr/bin/../share/java/kafka/kafka-clients-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/jetty-servlet-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/kafka-streams-examples-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/jakarta.annotation-api-1.3.5.jar:/usr/bin/../share/java/kafka/kafka-storage-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/scala-java8-compat_2.13-1.0.2.jar:/usr/bin/../share/java/kafka/kafka-raft-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/javax.ws.rs-api-2.1.1.jar:/usr/bin/../share/java/kafka/kafka-streams-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/zookeeper-jute-3.8.4.jar:/usr/bin/../share/java/kafka/plexus-utils-3.3.0.jar:/usr/bin/../share/java/kafka/zstd-jni-1.5.2-1.jar:/usr/bin/../share/java/kafka/connect-runtime-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/hk2-api-2.6.1.jar:/usr/bin/../share/java/kafka/jackson-dataformat-csv-2.13.5.jar:/usr/bin/../share/java/kafka/netty-transport-native-unix-common-4.1.115.Final.jar:/usr/bin/../share/java/kafka/kafka.jar:/usr/bin/../share/java/kafka/kafka-group-coordinator-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/jakarta.inject-2.6.1.jar:/usr/bin/../share/java/kafka/connect-api-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/netty-resolver-4.1.115.Final.jar:/usr/bin/../share/java/kafka/netty-transport-4.1.115.Final.jar:/usr/bin/../share/java/kafka/rocksdbjni-7.1.2.jar:/usr/bin/../share/java/kafka/jakarta.xml.bind-api-2.3.3.jar:/usr/bin/../share/java/kafka/jose4j-0.9.4.jar:/usr/bin/../share/java/kafka/hk2-locator-2.6.1.jar:/usr/bin/../share/java/kafka/kafka-metadata-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/netty-codec-4.1.115.Final.jar:/usr/bin/../share/java/kafka/connect-basic-auth-extension-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/slf4j-api-1.7.36.jar:/usr/bin/../share/java/kafka/jetty-continuation-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/kafka-log4j-appender-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/paranamer-2.8.jar:/usr/bin/../share/java/kafka/jaxb-api-2.3.1.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-2.39.1.jar:/usr/bin/../share/java/kafka/hk2-utils-2.6.1.jar:/usr/bin/../share/java/kafka/jackson-module-scala_2.13-2.13.5.jar:/usr/bin/../share/java/kafka/reload4j-1.2.25.jar:/usr/bin/../share/java/kafka/jackson-core-2.13.5.jar:/usr/bin/../share/java/kafka/netty-handler-4.1.115.Final.jar:/usr/bin/../share/java/kafka/jersey-hk2-2.39.1.jar:/usr/bin/../share/java/kafka/trogdor-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/jackson-databind-2.13.5.jar:/usr/bin/../share/java/kafka/jersey-client-2.39.1.jar:/usr/bin/../share/java/kafka/commons-io-2.16.0.jar:/usr/bin/../share/java/kafka/osgi-resource-locator-1.0.3.jar:/usr/bin/../share/java/kafka/jetty-util-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/connect-transforms-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/argparse4j-0.7.0.jar:/usr/bin/../share/java/kafka/jackson-datatype-jdk8-2.13.5.jar:/usr/bin/../share/java/kafka/audience-annotations-0.12.0.jar:/usr/bin/../share/java/kafka/jackson-module-jaxb-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/connect-mirror-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/javax.annotation-api-1.3.2.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-json-provider-2.13.5.jar:/usr/bin/../share/java/kafka/jakarta.validation-api-2.0.2.jar:/usr/bin/../share/java/kafka/jetty-client-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/zookeeper-3.8.4.jar:/usr/bin/../share/java/kafka/kafka-tools-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/jersey-server-2.39.1.jar:/usr/bin/../share/java/kafka/jetty-servlets-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/kafka_2.13-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/commons-lang3-3.8.1.jar:/usr/bin/../share/java/kafka/jopt-simple-5.0.4.jar:/usr/bin/../share/java/kafka/swagger-annotations-2.2.0.jar:/usr/bin/../share/java/kafka/kafka-streams-test-utils-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/lz4-java-1.8.0.jar:/usr/bin/../share/java/kafka/jakarta.activation-api-1.2.2.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-core-2.39.1.jar:/usr/bin/../share/java/kafka/jetty-server-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-base-2.13.5.jar:/usr/bin/../share/java/kafka/jetty-security-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/scala-reflect-2.13.10.jar:/usr/bin/../share/java/kafka/jetty-http-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/scala-collection-compat_2.13-2.10.0.jar:/usr/bin/../share/java/kafka/metrics-core-2.2.0.jar:/usr/bin/../share/java/kafka/netty-buffer-4.1.115.Final.jar:/usr/bin/../share/java/kafka/javassist-3.29.2-GA.jar:/usr/bin/../share/java/kafka/maven-artifact-3.8.4.jar:/usr/bin/../share/java/kafka/kafka-streams-scala_2.13-7.4.9-ccs.jar:/usr/bin/../share/java/kafka/netty-transport-classes-epoll-4.1.115.Final.jar:/usr/bin/../share/java/kafka/activation-1.1.1.jar:/usr/bin/../share/java/kafka/metrics-core-4.1.12.1.jar:/usr/bin/../share/java/kafka/snappy-java-1.1.10.5.jar:/usr/bin/../share/java/kafka/netty-transport-native-epoll-4.1.115.Final.jar:/usr/bin/../share/java/kafka/jetty-io-9.4.57.v20241219.jar:/usr/bin/../share/java/confluent-telemetry/* (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:os.version=4.15.0-192-generic (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:user.name=appuser (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:user.home=/home/appuser (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,757] INFO Client environment:user.dir=/home/appuser (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,758] INFO Client environment:os.memory.free=1009MB (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,758] INFO Client environment:os.memory.max=1024MB (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,758] INFO Client environment:os.memory.total=1024MB (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,759] INFO Initiating client connection, connectString=zookeeper:2181 sessionTimeout=18000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@52851b44 (org.apache.zookeeper.ZooKeeper) kafka | [2025-06-18 15:20:32,763] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket) kafka | [2025-06-18 15:20:32,769] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) kafka | [2025-06-18 15:20:32,773] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient) kafka | [2025-06-18 15:20:32,775] INFO Opening socket connection to server zookeeper/172.17.0.2:2181. (org.apache.zookeeper.ClientCnxn) kafka | [2025-06-18 15:20:32,782] INFO Socket connection established, initiating session, client: /172.17.0.7:34878, server: zookeeper/172.17.0.2:2181 (org.apache.zookeeper.ClientCnxn) kafka | [2025-06-18 15:20:32,861] INFO Session establishment complete on server zookeeper/172.17.0.2:2181, session id = 0x1000002fced0001, negotiated timeout = 18000 (org.apache.zookeeper.ClientCnxn) kafka | [2025-06-18 15:20:32,867] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient) kafka | [2025-06-18 15:20:33,414] INFO Cluster ID = 7hUyEsUsSwOpxuI0HNr3Gw (kafka.server.KafkaServer) kafka | [2025-06-18 15:20:33,418] WARN No meta.properties file under dir /var/lib/kafka/data/meta.properties (kafka.server.BrokerMetadataCheckpoint) kafka | [2025-06-18 15:20:33,462] INFO KafkaConfig values: kafka | advertised.listeners = PLAINTEXT://kafka:9092,PLAINTEXT_HOST://localhost:29092 kafka | alter.config.policy.class.name = null kafka | alter.log.dirs.replication.quota.window.num = 11 kafka | alter.log.dirs.replication.quota.window.size.seconds = 1 kafka | authorizer.class.name = kafka | auto.create.topics.enable = true kafka | auto.include.jmx.reporter = true kafka | auto.leader.rebalance.enable = true kafka | background.threads = 10 kafka | broker.heartbeat.interval.ms = 2000 kafka | broker.id = 1 kafka | broker.id.generation.enable = true kafka | broker.rack = null kafka | broker.session.timeout.ms = 9000 kafka | client.quota.callback.class = null kafka | compression.type = producer kafka | connection.failed.authentication.delay.ms = 100 kafka | connections.max.idle.ms = 600000 kafka | connections.max.reauth.ms = 0 kafka | control.plane.listener.name = null kafka | controlled.shutdown.enable = true kafka | controlled.shutdown.max.retries = 3 kafka | controlled.shutdown.retry.backoff.ms = 5000 kafka | controller.listener.names = null kafka | controller.quorum.append.linger.ms = 25 kafka | controller.quorum.election.backoff.max.ms = 1000 kafka | controller.quorum.election.timeout.ms = 1000 kafka | controller.quorum.fetch.timeout.ms = 2000 kafka | controller.quorum.request.timeout.ms = 2000 kafka | controller.quorum.retry.backoff.ms = 20 kafka | controller.quorum.voters = [] kafka | controller.quota.window.num = 11 kafka | controller.quota.window.size.seconds = 1 kafka | controller.socket.timeout.ms = 30000 kafka | create.topic.policy.class.name = null kafka | default.replication.factor = 1 kafka | delegation.token.expiry.check.interval.ms = 3600000 kafka | delegation.token.expiry.time.ms = 86400000 kafka | delegation.token.master.key = null kafka | delegation.token.max.lifetime.ms = 604800000 kafka | delegation.token.secret.key = null kafka | delete.records.purgatory.purge.interval.requests = 1 kafka | delete.topic.enable = true kafka | early.start.listeners = null kafka | fetch.max.bytes = 57671680 kafka | fetch.purgatory.purge.interval.requests = 1000 kafka | group.initial.rebalance.delay.ms = 3000 kafka | group.max.session.timeout.ms = 1800000 kafka | group.max.size = 2147483647 kafka | group.min.session.timeout.ms = 6000 kafka | initial.broker.registration.timeout.ms = 60000 kafka | inter.broker.listener.name = PLAINTEXT kafka | inter.broker.protocol.version = 3.4-IV0 kafka | kafka.metrics.polling.interval.secs = 10 kafka | kafka.metrics.reporters = [] kafka | leader.imbalance.check.interval.seconds = 300 kafka | leader.imbalance.per.broker.percentage = 10 kafka | listener.security.protocol.map = PLAINTEXT:PLAINTEXT,PLAINTEXT_HOST:PLAINTEXT kafka | listeners = PLAINTEXT://0.0.0.0:9092,PLAINTEXT_HOST://0.0.0.0:29092 kafka | log.cleaner.backoff.ms = 15000 kafka | log.cleaner.dedupe.buffer.size = 134217728 kafka | log.cleaner.delete.retention.ms = 86400000 kafka | log.cleaner.enable = true kafka | log.cleaner.io.buffer.load.factor = 0.9 kafka | log.cleaner.io.buffer.size = 524288 kafka | log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308 kafka | log.cleaner.max.compaction.lag.ms = 9223372036854775807 kafka | log.cleaner.min.cleanable.ratio = 0.5 kafka | log.cleaner.min.compaction.lag.ms = 0 kafka | log.cleaner.threads = 1 kafka | log.cleanup.policy = [delete] kafka | log.dir = /tmp/kafka-logs kafka | log.dirs = /var/lib/kafka/data kafka | log.flush.interval.messages = 9223372036854775807 kafka | log.flush.interval.ms = null kafka | log.flush.offset.checkpoint.interval.ms = 60000 kafka | log.flush.scheduler.interval.ms = 9223372036854775807 kafka | log.flush.start.offset.checkpoint.interval.ms = 60000 kafka | log.index.interval.bytes = 4096 kafka | log.index.size.max.bytes = 10485760 kafka | log.message.downconversion.enable = true kafka | log.message.format.version = 3.0-IV1 kafka | log.message.timestamp.difference.max.ms = 9223372036854775807 kafka | log.message.timestamp.type = CreateTime kafka | log.preallocate = false kafka | log.retention.bytes = -1 kafka | log.retention.check.interval.ms = 300000 kafka | log.retention.hours = 168 kafka | log.retention.minutes = null kafka | log.retention.ms = null kafka | log.roll.hours = 168 kafka | log.roll.jitter.hours = 0 kafka | log.roll.jitter.ms = null kafka | log.roll.ms = null kafka | log.segment.bytes = 1073741824 kafka | log.segment.delete.delay.ms = 60000 kafka | max.connection.creation.rate = 2147483647 kafka | max.connections = 2147483647 kafka | max.connections.per.ip = 2147483647 kafka | max.connections.per.ip.overrides = kafka | max.incremental.fetch.session.cache.slots = 1000 kafka | message.max.bytes = 1048588 kafka | metadata.log.dir = null kafka | metadata.log.max.record.bytes.between.snapshots = 20971520 kafka | metadata.log.max.snapshot.interval.ms = 3600000 kafka | metadata.log.segment.bytes = 1073741824 kafka | metadata.log.segment.min.bytes = 8388608 kafka | metadata.log.segment.ms = 604800000 kafka | metadata.max.idle.interval.ms = 500 kafka | metadata.max.retention.bytes = 104857600 kafka | metadata.max.retention.ms = 604800000 kafka | metric.reporters = [] kafka | metrics.num.samples = 2 kafka | metrics.recording.level = INFO kafka | metrics.sample.window.ms = 30000 kafka | min.insync.replicas = 1 kafka | node.id = 1 kafka | num.io.threads = 8 kafka | num.network.threads = 3 kafka | num.partitions = 1 kafka | num.recovery.threads.per.data.dir = 1 kafka | num.replica.alter.log.dirs.threads = null kafka | num.replica.fetchers = 1 kafka | offset.metadata.max.bytes = 4096 kafka | offsets.commit.required.acks = -1 kafka | offsets.commit.timeout.ms = 5000 kafka | offsets.load.buffer.size = 5242880 kafka | offsets.retention.check.interval.ms = 600000 kafka | offsets.retention.minutes = 10080 kafka | offsets.topic.compression.codec = 0 kafka | offsets.topic.num.partitions = 50 kafka | offsets.topic.replication.factor = 1 kafka | offsets.topic.segment.bytes = 104857600 kafka | password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding kafka | password.encoder.iterations = 4096 kafka | password.encoder.key.length = 128 kafka | password.encoder.keyfactory.algorithm = null kafka | password.encoder.old.secret = null kafka | password.encoder.secret = null kafka | principal.builder.class = class org.apache.kafka.common.security.authenticator.DefaultKafkaPrincipalBuilder kafka | process.roles = [] kafka | producer.id.expiration.check.interval.ms = 600000 kafka | producer.id.expiration.ms = 86400000 kafka | producer.purgatory.purge.interval.requests = 1000 kafka | queued.max.request.bytes = -1 kafka | queued.max.requests = 500 kafka | quota.window.num = 11 kafka | quota.window.size.seconds = 1 kafka | remote.log.index.file.cache.total.size.bytes = 1073741824 kafka | remote.log.manager.task.interval.ms = 30000 kafka | remote.log.manager.task.retry.backoff.max.ms = 30000 kafka | remote.log.manager.task.retry.backoff.ms = 500 kafka | remote.log.manager.task.retry.jitter = 0.2 kafka | remote.log.manager.thread.pool.size = 10 kafka | remote.log.metadata.manager.class.name = null kafka | remote.log.metadata.manager.class.path = null kafka | remote.log.metadata.manager.impl.prefix = null kafka | remote.log.metadata.manager.listener.name = null kafka | remote.log.reader.max.pending.tasks = 100 kafka | remote.log.reader.threads = 10 kafka | remote.log.storage.manager.class.name = null kafka | remote.log.storage.manager.class.path = null kafka | remote.log.storage.manager.impl.prefix = null kafka | remote.log.storage.system.enable = false kafka | replica.fetch.backoff.ms = 1000 kafka | replica.fetch.max.bytes = 1048576 kafka | replica.fetch.min.bytes = 1 kafka | replica.fetch.response.max.bytes = 10485760 kafka | replica.fetch.wait.max.ms = 500 kafka | replica.high.watermark.checkpoint.interval.ms = 5000 kafka | replica.lag.time.max.ms = 30000 kafka | replica.selector.class = null kafka | replica.socket.receive.buffer.bytes = 65536 kafka | replica.socket.timeout.ms = 30000 kafka | replication.quota.window.num = 11 kafka | replication.quota.window.size.seconds = 1 kafka | request.timeout.ms = 30000 kafka | reserved.broker.max.id = 1000 kafka | sasl.client.callback.handler.class = null kafka | sasl.enabled.mechanisms = [GSSAPI] kafka | sasl.jaas.config = null kafka | sasl.kerberos.kinit.cmd = /usr/bin/kinit kafka | sasl.kerberos.min.time.before.relogin = 60000 kafka | sasl.kerberos.principal.to.local.rules = [DEFAULT] kafka | sasl.kerberos.service.name = null kafka | sasl.kerberos.ticket.renew.jitter = 0.05 kafka | sasl.kerberos.ticket.renew.window.factor = 0.8 kafka | sasl.login.callback.handler.class = null kafka | sasl.login.class = null kafka | sasl.login.connect.timeout.ms = null kafka | sasl.login.read.timeout.ms = null kafka | sasl.login.refresh.buffer.seconds = 300 kafka | sasl.login.refresh.min.period.seconds = 60 kafka | sasl.login.refresh.window.factor = 0.8 kafka | sasl.login.refresh.window.jitter = 0.05 kafka | sasl.login.retry.backoff.max.ms = 10000 kafka | sasl.login.retry.backoff.ms = 100 kafka | sasl.mechanism.controller.protocol = GSSAPI kafka | sasl.mechanism.inter.broker.protocol = GSSAPI kafka | sasl.oauthbearer.clock.skew.seconds = 30 kafka | sasl.oauthbearer.expected.audience = null kafka | sasl.oauthbearer.expected.issuer = null kafka | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 kafka | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 kafka | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 kafka | sasl.oauthbearer.jwks.endpoint.url = null kafka | sasl.oauthbearer.scope.claim.name = scope kafka | sasl.oauthbearer.sub.claim.name = sub kafka | sasl.oauthbearer.token.endpoint.url = null kafka | sasl.server.callback.handler.class = null kafka | sasl.server.max.receive.size = 524288 kafka | security.inter.broker.protocol = PLAINTEXT kafka | security.providers = null kafka | socket.connection.setup.timeout.max.ms = 30000 kafka | socket.connection.setup.timeout.ms = 10000 kafka | socket.listen.backlog.size = 50 kafka | socket.receive.buffer.bytes = 102400 kafka | socket.request.max.bytes = 104857600 kafka | socket.send.buffer.bytes = 102400 kafka | ssl.cipher.suites = [] kafka | ssl.client.auth = none kafka | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] kafka | ssl.endpoint.identification.algorithm = https kafka | ssl.engine.factory.class = null kafka | ssl.key.password = null kafka | ssl.keymanager.algorithm = SunX509 kafka | ssl.keystore.certificate.chain = null kafka | ssl.keystore.key = null kafka | ssl.keystore.location = null kafka | ssl.keystore.password = null kafka | ssl.keystore.type = JKS kafka | ssl.principal.mapping.rules = DEFAULT kafka | ssl.protocol = TLSv1.3 kafka | ssl.provider = null kafka | ssl.secure.random.implementation = null kafka | ssl.trustmanager.algorithm = PKIX kafka | ssl.truststore.certificates = null kafka | ssl.truststore.location = null kafka | ssl.truststore.password = null kafka | ssl.truststore.type = JKS kafka | transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000 kafka | transaction.max.timeout.ms = 900000 kafka | transaction.remove.expired.transaction.cleanup.interval.ms = 3600000 kafka | transaction.state.log.load.buffer.size = 5242880 kafka | transaction.state.log.min.isr = 2 kafka | transaction.state.log.num.partitions = 50 kafka | transaction.state.log.replication.factor = 3 kafka | transaction.state.log.segment.bytes = 104857600 kafka | transactional.id.expiration.ms = 604800000 kafka | unclean.leader.election.enable = false kafka | zookeeper.clientCnxnSocket = null kafka | zookeeper.connect = zookeeper:2181 kafka | zookeeper.connection.timeout.ms = null kafka | zookeeper.max.in.flight.requests = 10 kafka | zookeeper.metadata.migration.enable = false kafka | zookeeper.session.timeout.ms = 18000 kafka | zookeeper.set.acl = false kafka | zookeeper.ssl.cipher.suites = null kafka | zookeeper.ssl.client.enable = false kafka | zookeeper.ssl.crl.enable = false kafka | zookeeper.ssl.enabled.protocols = null kafka | zookeeper.ssl.endpoint.identification.algorithm = HTTPS kafka | zookeeper.ssl.keystore.location = null kafka | zookeeper.ssl.keystore.password = null kafka | zookeeper.ssl.keystore.type = null kafka | zookeeper.ssl.ocsp.enable = false kafka | zookeeper.ssl.protocol = TLSv1.2 kafka | zookeeper.ssl.truststore.location = null kafka | zookeeper.ssl.truststore.password = null kafka | zookeeper.ssl.truststore.type = null kafka | (kafka.server.KafkaConfig) kafka | [2025-06-18 15:20:33,494] INFO [ThrottledChannelReaper-Fetch]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) kafka | [2025-06-18 15:20:33,495] INFO [ThrottledChannelReaper-Produce]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) kafka | [2025-06-18 15:20:33,497] INFO [ThrottledChannelReaper-Request]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) kafka | [2025-06-18 15:20:33,499] INFO [ThrottledChannelReaper-ControllerMutation]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) kafka | [2025-06-18 15:20:33,530] INFO Loading logs from log dirs ArraySeq(/var/lib/kafka/data) (kafka.log.LogManager) kafka | [2025-06-18 15:20:33,535] INFO Attempting recovery for all logs in /var/lib/kafka/data since no clean shutdown file was found (kafka.log.LogManager) kafka | [2025-06-18 15:20:33,547] INFO Loaded 0 logs in 16ms. (kafka.log.LogManager) kafka | [2025-06-18 15:20:33,547] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager) kafka | [2025-06-18 15:20:33,549] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager) kafka | [2025-06-18 15:20:33,559] INFO Starting the log cleaner (kafka.log.LogCleaner) kafka | [2025-06-18 15:20:33,603] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner) kafka | [2025-06-18 15:20:33,618] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) kafka | [2025-06-18 15:20:33,634] INFO Feature ZK node at path: /feature does not exist (kafka.server.FinalizedFeatureChangeListener) kafka | [2025-06-18 15:20:33,669] INFO [BrokerToControllerChannelManager broker=1 name=forwarding]: Starting (kafka.server.BrokerToControllerRequestThread) kafka | [2025-06-18 15:20:34,021] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) kafka | [2025-06-18 15:20:34,025] INFO Awaiting socket connections on 0.0.0.0:9092. (kafka.network.DataPlaneAcceptor) kafka | [2025-06-18 15:20:34,048] INFO [SocketServer listenerType=ZK_BROKER, nodeId=1] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer) kafka | [2025-06-18 15:20:34,049] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) kafka | [2025-06-18 15:20:34,049] INFO Awaiting socket connections on 0.0.0.0:29092. (kafka.network.DataPlaneAcceptor) kafka | [2025-06-18 15:20:34,054] INFO [SocketServer listenerType=ZK_BROKER, nodeId=1] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT_HOST) (kafka.network.SocketServer) kafka | [2025-06-18 15:20:34,058] INFO [BrokerToControllerChannelManager broker=1 name=alterPartition]: Starting (kafka.server.BrokerToControllerRequestThread) kafka | [2025-06-18 15:20:34,075] INFO [ExpirationReaper-1-Produce]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2025-06-18 15:20:34,077] INFO [ExpirationReaper-1-Fetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2025-06-18 15:20:34,081] INFO [ExpirationReaper-1-DeleteRecords]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2025-06-18 15:20:34,083] INFO [ExpirationReaper-1-ElectLeader]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2025-06-18 15:20:34,119] INFO [LogDirFailureHandler]: Starting (kafka.server.ReplicaManager$LogDirFailureHandler) kafka | [2025-06-18 15:20:34,144] INFO Creating /brokers/ids/1 (is it secure? false) (kafka.zk.KafkaZkClient) kafka | [2025-06-18 15:20:34,171] INFO Stat of the created znode at /brokers/ids/1 is: 27,27,1750260034159,1750260034159,1,0,0,72057606871252993,258,0,27 kafka | (kafka.zk.KafkaZkClient) kafka | [2025-06-18 15:20:34,171] INFO Registered broker 1 at path /brokers/ids/1 with addresses: PLAINTEXT://kafka:9092,PLAINTEXT_HOST://localhost:29092, czxid (broker epoch): 27 (kafka.zk.KafkaZkClient) kafka | [2025-06-18 15:20:34,289] INFO [ControllerEventThread controllerId=1] Starting (kafka.controller.ControllerEventManager$ControllerEventThread) kafka | [2025-06-18 15:20:34,300] INFO [ExpirationReaper-1-topic]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2025-06-18 15:20:34,313] INFO [ExpirationReaper-1-Heartbeat]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2025-06-18 15:20:34,314] INFO [ExpirationReaper-1-Rebalance]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2025-06-18 15:20:34,316] INFO Successfully created /controller_epoch with initial epoch 0 (kafka.zk.KafkaZkClient) kafka | [2025-06-18 15:20:34,333] INFO [Controller id=1] 1 successfully elected as the controller. Epoch incremented to 1 and epoch zk version is now 1 (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,334] INFO [GroupCoordinator 1]: Starting up. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:20:34,338] INFO [GroupCoordinator 1]: Startup complete. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:20:34,339] INFO [Controller id=1] Creating FeatureZNode at path: /feature with contents: FeatureZNode(2,Enabled,Map()) (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,351] INFO [TransactionCoordinator id=1] Starting up. (kafka.coordinator.transaction.TransactionCoordinator) kafka | [2025-06-18 15:20:34,353] INFO Feature ZK node created at path: /feature (kafka.server.FinalizedFeatureChangeListener) kafka | [2025-06-18 15:20:34,354] INFO [TransactionCoordinator id=1] Startup complete. (kafka.coordinator.transaction.TransactionCoordinator) kafka | [2025-06-18 15:20:34,355] INFO [Transaction Marker Channel Manager 1]: Starting (kafka.coordinator.transaction.TransactionMarkerChannelManager) kafka | [2025-06-18 15:20:34,382] INFO [ExpirationReaper-1-AlterAcls]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2025-06-18 15:20:34,383] INFO [MetadataCache brokerId=1] Updated cache from existing to latest FinalizedFeaturesAndEpoch(features=Map(), epoch=0). (kafka.server.metadata.ZkMetadataCache) kafka | [2025-06-18 15:20:34,383] INFO [Controller id=1] Registering handlers (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,388] INFO [Controller id=1] Deleting log dir event notifications (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,392] INFO [Controller id=1] Deleting isr change notifications (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,395] INFO [Controller id=1] Initializing controller context (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,407] INFO [/config/changes-event-process-thread]: Starting (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) kafka | [2025-06-18 15:20:34,415] INFO [Controller id=1] Initialized broker epochs cache: HashMap(1 -> 27) (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,418] INFO [SocketServer listenerType=ZK_BROKER, nodeId=1] Enabling request processing. (kafka.network.SocketServer) kafka | [2025-06-18 15:20:34,421] DEBUG [Controller id=1] Register BrokerModifications handler for Set(1) (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,427] DEBUG [Channel manager on controller 1]: Controller 1 trying to connect to broker 1 (kafka.controller.ControllerChannelManager) kafka | [2025-06-18 15:20:34,427] INFO Kafka version: 7.4.9-ccs (org.apache.kafka.common.utils.AppInfoParser) kafka | [2025-06-18 15:20:34,427] INFO Kafka commitId: 07d888cfc0d14765fe5557324f1fdb4ada6698a5 (org.apache.kafka.common.utils.AppInfoParser) kafka | [2025-06-18 15:20:34,427] INFO Kafka startTimeMs: 1750260034422 (org.apache.kafka.common.utils.AppInfoParser) kafka | [2025-06-18 15:20:34,429] INFO [KafkaServer id=1] started (kafka.server.KafkaServer) kafka | [2025-06-18 15:20:34,437] INFO [RequestSendThread controllerId=1] Starting (kafka.controller.RequestSendThread) kafka | [2025-06-18 15:20:34,438] INFO [Controller id=1] Currently active brokers in the cluster: Set(1) (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,438] INFO [Controller id=1] Currently shutting brokers in the cluster: HashSet() (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,438] INFO [Controller id=1] Current list of topics in the cluster: HashSet() (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,438] INFO [Controller id=1] Fetching topic deletions in progress (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,442] INFO [Controller id=1] List of topics to be deleted: (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,443] INFO [Controller id=1] List of topics ineligible for deletion: (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,443] INFO [Controller id=1] Initializing topic deletion manager (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,443] INFO [Topic Deletion Manager 1] Initializing manager with initial deletions: Set(), initial ineligible deletions: HashSet() (kafka.controller.TopicDeletionManager) kafka | [2025-06-18 15:20:34,445] INFO [Controller id=1] Sending update metadata request (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,450] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet(1) for 0 partitions (state.change.logger) kafka | [2025-06-18 15:20:34,461] INFO [ReplicaStateMachine controllerId=1] Initializing replica state (kafka.controller.ZkReplicaStateMachine) kafka | [2025-06-18 15:20:34,462] INFO [ReplicaStateMachine controllerId=1] Triggering online replica state changes (kafka.controller.ZkReplicaStateMachine) kafka | [2025-06-18 15:20:34,466] INFO [RequestSendThread controllerId=1] Controller 1 connected to kafka:9092 (id: 1 rack: null) for sending state change requests (kafka.controller.RequestSendThread) kafka | [2025-06-18 15:20:34,468] INFO [ReplicaStateMachine controllerId=1] Triggering offline replica state changes (kafka.controller.ZkReplicaStateMachine) kafka | [2025-06-18 15:20:34,468] DEBUG [ReplicaStateMachine controllerId=1] Started replica state machine with initial state -> HashMap() (kafka.controller.ZkReplicaStateMachine) kafka | [2025-06-18 15:20:34,469] INFO [PartitionStateMachine controllerId=1] Initializing partition state (kafka.controller.ZkPartitionStateMachine) kafka | [2025-06-18 15:20:34,471] INFO [PartitionStateMachine controllerId=1] Triggering online partition state changes (kafka.controller.ZkPartitionStateMachine) kafka | [2025-06-18 15:20:34,475] DEBUG [PartitionStateMachine controllerId=1] Started partition state machine with initial state -> HashMap() (kafka.controller.ZkPartitionStateMachine) kafka | [2025-06-18 15:20:34,476] INFO [Controller id=1] Ready to serve as the new controller with epoch 1 (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,487] INFO [Controller id=1] Partitions undergoing preferred replica election: (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,487] INFO [Controller id=1] Partitions that completed preferred replica election: (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,488] INFO [Controller id=1] Skipping preferred replica election for partitions due to topic deletion: (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,488] INFO [Controller id=1] Resuming preferred replica election for partitions: (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,490] INFO [Controller id=1] Starting replica leader election (PREFERRED) for partitions triggered by ZkTriggered (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,513] INFO [Controller id=1] Starting the controller scheduler (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:34,523] TRACE [Controller id=1 epoch=1] Received response UpdateMetadataResponseData(errorCode=0) for request UPDATE_METADATA with correlation id 0 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) kafka | [2025-06-18 15:20:34,565] INFO [BrokerToControllerChannelManager broker=1 name=alterPartition]: Recorded new controller, from now on will use node kafka:9092 (id: 1 rack: null) (kafka.server.BrokerToControllerRequestThread) kafka | [2025-06-18 15:20:34,596] INFO [BrokerToControllerChannelManager broker=1 name=forwarding]: Recorded new controller, from now on will use node kafka:9092 (id: 1 rack: null) (kafka.server.BrokerToControllerRequestThread) kafka | [2025-06-18 15:20:39,515] INFO [Controller id=1] Processing automatic preferred replica leader election (kafka.controller.KafkaController) kafka | [2025-06-18 15:20:39,515] TRACE [Controller id=1] Checking need to trigger auto leader balancing (kafka.controller.KafkaController) kafka | [2025-06-18 15:21:04,600] INFO Creating topic __consumer_offsets with configuration {compression.type=producer, cleanup.policy=compact, segment.bytes=104857600} and initial partition assignment HashMap(0 -> ArrayBuffer(1), 1 -> ArrayBuffer(1), 2 -> ArrayBuffer(1), 3 -> ArrayBuffer(1), 4 -> ArrayBuffer(1), 5 -> ArrayBuffer(1), 6 -> ArrayBuffer(1), 7 -> ArrayBuffer(1), 8 -> ArrayBuffer(1), 9 -> ArrayBuffer(1), 10 -> ArrayBuffer(1), 11 -> ArrayBuffer(1), 12 -> ArrayBuffer(1), 13 -> ArrayBuffer(1), 14 -> ArrayBuffer(1), 15 -> ArrayBuffer(1), 16 -> ArrayBuffer(1), 17 -> ArrayBuffer(1), 18 -> ArrayBuffer(1), 19 -> ArrayBuffer(1), 20 -> ArrayBuffer(1), 21 -> ArrayBuffer(1), 22 -> ArrayBuffer(1), 23 -> ArrayBuffer(1), 24 -> ArrayBuffer(1), 25 -> ArrayBuffer(1), 26 -> ArrayBuffer(1), 27 -> ArrayBuffer(1), 28 -> ArrayBuffer(1), 29 -> ArrayBuffer(1), 30 -> ArrayBuffer(1), 31 -> ArrayBuffer(1), 32 -> ArrayBuffer(1), 33 -> ArrayBuffer(1), 34 -> ArrayBuffer(1), 35 -> ArrayBuffer(1), 36 -> ArrayBuffer(1), 37 -> ArrayBuffer(1), 38 -> ArrayBuffer(1), 39 -> ArrayBuffer(1), 40 -> ArrayBuffer(1), 41 -> ArrayBuffer(1), 42 -> ArrayBuffer(1), 43 -> ArrayBuffer(1), 44 -> ArrayBuffer(1), 45 -> ArrayBuffer(1), 46 -> ArrayBuffer(1), 47 -> ArrayBuffer(1), 48 -> ArrayBuffer(1), 49 -> ArrayBuffer(1)) (kafka.zk.AdminZkClient) kafka | [2025-06-18 15:21:04,601] INFO Creating topic policy-pdp-pap with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(1)) (kafka.zk.AdminZkClient) kafka | [2025-06-18 15:21:04,602] DEBUG [Controller id=1] There is no producerId block yet (Zk path version 0), creating the first block (kafka.controller.KafkaController) kafka | [2025-06-18 15:21:04,642] INFO [Controller id=1] Acquired new producerId block ProducerIdsBlock(assignedBrokerId=1, firstProducerId=0, size=1000) by writing to Zk with path version 1 (kafka.controller.KafkaController) kafka | [2025-06-18 15:21:04,683] INFO [Controller id=1] New topics: [Set(policy-pdp-pap, __consumer_offsets)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(policy-pdp-pap,Some(_aOJTUb1QR2K51o55qXHnQ),Map(policy-pdp-pap-0 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=))), TopicIdReplicaAssignment(__consumer_offsets,Some(zkopqUgzQEm01fHoabEl0Q),HashMap(__consumer_offsets-22 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-30 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-25 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-35 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-37 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-38 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-13 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-8 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-21 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-4 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-27 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-7 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-9 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-46 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-41 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-33 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-23 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-49 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-47 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-16 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-28 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-31 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-36 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-42 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-3 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-18 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-15 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-24 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-17 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-48 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-19 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-11 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-2 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-43 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-6 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-14 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-20 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-0 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-44 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-39 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-12 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-45 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-1 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-5 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-26 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-29 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-34 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-10 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-32 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-40 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController) kafka | [2025-06-18 15:21:04,684] INFO [Controller id=1] New partition creation callback for __consumer_offsets-22,__consumer_offsets-30,__consumer_offsets-25,__consumer_offsets-35,__consumer_offsets-38,__consumer_offsets-13,__consumer_offsets-8,__consumer_offsets-21,__consumer_offsets-4,__consumer_offsets-27,__consumer_offsets-7,__consumer_offsets-9,__consumer_offsets-46,__consumer_offsets-41,__consumer_offsets-33,__consumer_offsets-23,__consumer_offsets-49,__consumer_offsets-47,__consumer_offsets-16,__consumer_offsets-28,__consumer_offsets-31,__consumer_offsets-36,__consumer_offsets-42,__consumer_offsets-3,__consumer_offsets-18,__consumer_offsets-37,policy-pdp-pap-0,__consumer_offsets-15,__consumer_offsets-24,__consumer_offsets-17,__consumer_offsets-48,__consumer_offsets-19,__consumer_offsets-11,__consumer_offsets-2,__consumer_offsets-43,__consumer_offsets-6,__consumer_offsets-14,__consumer_offsets-20,__consumer_offsets-0,__consumer_offsets-44,__consumer_offsets-39,__consumer_offsets-12,__consumer_offsets-45,__consumer_offsets-1,__consumer_offsets-5,__consumer_offsets-26,__consumer_offsets-29,__consumer_offsets-34,__consumer_offsets-10,__consumer_offsets-32,__consumer_offsets-40 (kafka.controller.KafkaController) kafka | [2025-06-18 15:21:04,686] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-22 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,686] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-30 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,686] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-25 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,686] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-35 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,686] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-38 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,686] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-13 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,686] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-8 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,686] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-21 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,686] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-4 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-27 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-7 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-9 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-46 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-41 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-33 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-23 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-49 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-47 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-16 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-28 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-31 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-36 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-42 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-3 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-18 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-37 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition policy-pdp-pap-0 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-15 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-24 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-17 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-48 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-19 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-11 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-2 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-43 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-6 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-14 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-20 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-0 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,687] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-44 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-39 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-12 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-45 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-1 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-5 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-26 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-29 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-34 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-10 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-32 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-40 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:04,688] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) kafka | [2025-06-18 15:21:04,699] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-32 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,699] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-5 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-44 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-48 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-46 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-20 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition policy-pdp-pap-0 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-43 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-24 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-6 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-18 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-21 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-1 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-14 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-34 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-16 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,702] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-29 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-11 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-0 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-22 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-47 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-36 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-28 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-42 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-9 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-37 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-13 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-30 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-35 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-39 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-12 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-27 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-45 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-19 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-49 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-40 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-41 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-38 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-8 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-7 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-33 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-25 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-31 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-23 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-10 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-2 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-17 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-4 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,703] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-15 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,704] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-26 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,704] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-3 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:04,704] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-22 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-30 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-25 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-35 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-38 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-13 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-8 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-21 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-4 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-27 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-7 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-9 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-46 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-41 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-33 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-23 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-49 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-47 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-16 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,090] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-28 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,091] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-31 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,091] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-36 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,091] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-42 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,091] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-3 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,091] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-18 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,091] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-37 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition policy-pdp-pap-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-15 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-24 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-17 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-48 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-19 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-11 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-2 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-43 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-6 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-14 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-20 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-44 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-39 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-12 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-45 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-5 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-26 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-29 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-34 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-10 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-32 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,092] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-40 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=13, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-13 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=46, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-46 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=9, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-9 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=42, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-42 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=21, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-21 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=17, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-17 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=30, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-30 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=26, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-26 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=5, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-5 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=38, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-38 (state.change.logger) kafka | [2025-06-18 15:21:05,097] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=1, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-1 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=34, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-34 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=16, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-16 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=45, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-45 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=12, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-12 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=41, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-41 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=24, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-24 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=20, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-20 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=49, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-49 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-0 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=29, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-29 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=25, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-25 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=8, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-8 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=37, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-37 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=4, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-4 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=33, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-33 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=15, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-15 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=48, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-48 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=11, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-11 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=44, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-44 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=23, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-23 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=19, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-19 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=32, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-32 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=28, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-28 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=7, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-7 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=40, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-40 (state.change.logger) kafka | [2025-06-18 15:21:05,098] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=3, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-3 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=36, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-36 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=47, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-47 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=14, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-14 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=43, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-43 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=10, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-10 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=22, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-22 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=18, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-18 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=31, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-31 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=27, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-27 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=39, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-39 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=6, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-6 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=35, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-35 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='policy-pdp-pap', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition policy-pdp-pap-0 (state.change.logger) kafka | [2025-06-18 15:21:05,099] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=2, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-2 (state.change.logger) kafka | [2025-06-18 15:21:05,100] INFO [Controller id=1 epoch=1] Sending LeaderAndIsr request to broker 1 with 51 become-leader and 0 become-follower partitions (state.change.logger) kafka | [2025-06-18 15:21:05,102] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet(1) for 51 partitions (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-32 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-5 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-44 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-48 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-46 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-20 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition policy-pdp-pap-0 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-43 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-24 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-6 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-18 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-21 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-1 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,103] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-14 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-34 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-16 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-29 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-11 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-0 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-22 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-47 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-36 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-28 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-42 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-9 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-37 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-13 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-30 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-35 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-39 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-12 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-27 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-45 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-19 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-49 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-40 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-41 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-38 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-8 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-7 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-33 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-25 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-31 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-23 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-10 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-2 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-17 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,104] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-4 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,105] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-15 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,105] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-26 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,105] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-3 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:05,105] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) kafka | [2025-06-18 15:21:05,109] INFO [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 for 51 partitions (state.change.logger) kafka | [2025-06-18 15:21:05,110] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='policy-pdp-pap', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,110] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=13, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,110] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=46, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,110] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=9, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=42, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=21, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=17, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=30, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=26, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=5, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=38, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=1, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=34, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=16, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=45, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=12, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=41, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=24, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=20, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=49, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=29, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=25, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=8, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=37, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=4, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=33, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=15, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=48, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=11, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=44, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=23, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=19, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=32, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=28, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=7, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=40, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=3, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=36, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,111] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=47, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=14, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=43, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=10, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=22, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=18, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=31, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=27, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=39, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=6, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=35, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,112] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=2, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:05,153] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-3 (state.change.logger) kafka | [2025-06-18 15:21:05,153] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-18 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-41 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-10 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-33 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-48 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-19 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-34 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-4 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-11 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-26 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-49 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-39 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-9 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-24 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-31 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-46 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-1 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-16 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-2 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-25 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-40 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-47 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-17 (state.change.logger) kafka | [2025-06-18 15:21:05,154] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-32 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-37 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-7 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-22 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-29 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-44 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-14 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-23 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-38 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-8 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition policy-pdp-pap-0 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-45 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-15 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-30 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-0 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-35 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-5 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-20 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-27 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-42 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-12 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-21 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-36 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-6 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-43 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-13 (state.change.logger) kafka | [2025-06-18 15:21:05,155] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-28 (state.change.logger) kafka | [2025-06-18 15:21:05,156] INFO [ReplicaFetcherManager on broker 1] Removed fetcher for partitions HashSet(__consumer_offsets-22, __consumer_offsets-30, __consumer_offsets-25, __consumer_offsets-35, __consumer_offsets-38, __consumer_offsets-13, __consumer_offsets-8, __consumer_offsets-21, __consumer_offsets-4, __consumer_offsets-27, __consumer_offsets-7, __consumer_offsets-9, __consumer_offsets-46, __consumer_offsets-41, __consumer_offsets-33, __consumer_offsets-23, __consumer_offsets-49, __consumer_offsets-47, __consumer_offsets-16, __consumer_offsets-28, __consumer_offsets-31, __consumer_offsets-36, __consumer_offsets-42, __consumer_offsets-3, __consumer_offsets-18, __consumer_offsets-37, policy-pdp-pap-0, __consumer_offsets-15, __consumer_offsets-24, __consumer_offsets-17, __consumer_offsets-48, __consumer_offsets-19, __consumer_offsets-11, __consumer_offsets-2, __consumer_offsets-43, __consumer_offsets-6, __consumer_offsets-14, __consumer_offsets-20, __consumer_offsets-0, __consumer_offsets-44, __consumer_offsets-39, __consumer_offsets-12, __consumer_offsets-45, __consumer_offsets-1, __consumer_offsets-5, __consumer_offsets-26, __consumer_offsets-29, __consumer_offsets-34, __consumer_offsets-10, __consumer_offsets-32, __consumer_offsets-40) (kafka.server.ReplicaFetcherManager) kafka | [2025-06-18 15:21:05,156] INFO [Broker id=1] Stopped fetchers as part of LeaderAndIsr request correlationId 1 from controller 1 epoch 1 as part of the become-leader transition for 51 partitions (state.change.logger) kafka | [2025-06-18 15:21:05,200] INFO [LogLoader partition=__consumer_offsets-3, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,209] INFO Created log for partition __consumer_offsets-3 in /var/lib/kafka/data/__consumer_offsets-3 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,212] INFO [Partition __consumer_offsets-3 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-3 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,212] INFO [Partition __consumer_offsets-3 broker=1] Log loaded for partition __consumer_offsets-3 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,213] INFO [Broker id=1] Leader __consumer_offsets-3 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,276] INFO [LogLoader partition=__consumer_offsets-18, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,277] INFO Created log for partition __consumer_offsets-18 in /var/lib/kafka/data/__consumer_offsets-18 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,277] INFO [Partition __consumer_offsets-18 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-18 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,277] INFO [Partition __consumer_offsets-18 broker=1] Log loaded for partition __consumer_offsets-18 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,277] INFO [Broker id=1] Leader __consumer_offsets-18 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,295] INFO [LogLoader partition=__consumer_offsets-41, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,296] INFO Created log for partition __consumer_offsets-41 in /var/lib/kafka/data/__consumer_offsets-41 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,296] INFO [Partition __consumer_offsets-41 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-41 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,296] INFO [Partition __consumer_offsets-41 broker=1] Log loaded for partition __consumer_offsets-41 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,296] INFO [Broker id=1] Leader __consumer_offsets-41 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,309] INFO [LogLoader partition=__consumer_offsets-10, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,310] INFO Created log for partition __consumer_offsets-10 in /var/lib/kafka/data/__consumer_offsets-10 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,310] INFO [Partition __consumer_offsets-10 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-10 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,311] INFO [Partition __consumer_offsets-10 broker=1] Log loaded for partition __consumer_offsets-10 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,311] INFO [Broker id=1] Leader __consumer_offsets-10 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,329] INFO [LogLoader partition=__consumer_offsets-33, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,330] INFO Created log for partition __consumer_offsets-33 in /var/lib/kafka/data/__consumer_offsets-33 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,330] INFO [Partition __consumer_offsets-33 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-33 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,330] INFO [Partition __consumer_offsets-33 broker=1] Log loaded for partition __consumer_offsets-33 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,330] INFO [Broker id=1] Leader __consumer_offsets-33 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,368] INFO [LogLoader partition=__consumer_offsets-48, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,369] INFO Created log for partition __consumer_offsets-48 in /var/lib/kafka/data/__consumer_offsets-48 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,369] INFO [Partition __consumer_offsets-48 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-48 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,369] INFO [Partition __consumer_offsets-48 broker=1] Log loaded for partition __consumer_offsets-48 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,369] INFO [Broker id=1] Leader __consumer_offsets-48 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,407] INFO [LogLoader partition=__consumer_offsets-19, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,408] INFO Created log for partition __consumer_offsets-19 in /var/lib/kafka/data/__consumer_offsets-19 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,408] INFO [Partition __consumer_offsets-19 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-19 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,408] INFO [Partition __consumer_offsets-19 broker=1] Log loaded for partition __consumer_offsets-19 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,408] INFO [Broker id=1] Leader __consumer_offsets-19 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,425] INFO [LogLoader partition=__consumer_offsets-34, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,426] INFO Created log for partition __consumer_offsets-34 in /var/lib/kafka/data/__consumer_offsets-34 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,426] INFO [Partition __consumer_offsets-34 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-34 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,426] INFO [Partition __consumer_offsets-34 broker=1] Log loaded for partition __consumer_offsets-34 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,427] INFO [Broker id=1] Leader __consumer_offsets-34 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,455] INFO [LogLoader partition=__consumer_offsets-4, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,456] INFO Created log for partition __consumer_offsets-4 in /var/lib/kafka/data/__consumer_offsets-4 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,456] INFO [Partition __consumer_offsets-4 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-4 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,456] INFO [Partition __consumer_offsets-4 broker=1] Log loaded for partition __consumer_offsets-4 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,456] INFO [Broker id=1] Leader __consumer_offsets-4 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,547] INFO [LogLoader partition=__consumer_offsets-11, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,548] INFO Created log for partition __consumer_offsets-11 in /var/lib/kafka/data/__consumer_offsets-11 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,548] INFO [Partition __consumer_offsets-11 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-11 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,548] INFO [Partition __consumer_offsets-11 broker=1] Log loaded for partition __consumer_offsets-11 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,548] INFO [Broker id=1] Leader __consumer_offsets-11 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,577] INFO [LogLoader partition=__consumer_offsets-26, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,578] INFO Created log for partition __consumer_offsets-26 in /var/lib/kafka/data/__consumer_offsets-26 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,578] INFO [Partition __consumer_offsets-26 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-26 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,578] INFO [Partition __consumer_offsets-26 broker=1] Log loaded for partition __consumer_offsets-26 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,578] INFO [Broker id=1] Leader __consumer_offsets-26 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,637] INFO [LogLoader partition=__consumer_offsets-49, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,638] INFO Created log for partition __consumer_offsets-49 in /var/lib/kafka/data/__consumer_offsets-49 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,638] INFO [Partition __consumer_offsets-49 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-49 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,638] INFO [Partition __consumer_offsets-49 broker=1] Log loaded for partition __consumer_offsets-49 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,638] INFO [Broker id=1] Leader __consumer_offsets-49 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,658] INFO [LogLoader partition=__consumer_offsets-39, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,659] INFO Created log for partition __consumer_offsets-39 in /var/lib/kafka/data/__consumer_offsets-39 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,660] INFO [Partition __consumer_offsets-39 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-39 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,660] INFO [Partition __consumer_offsets-39 broker=1] Log loaded for partition __consumer_offsets-39 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,660] INFO [Broker id=1] Leader __consumer_offsets-39 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,683] INFO [LogLoader partition=__consumer_offsets-9, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,685] INFO Created log for partition __consumer_offsets-9 in /var/lib/kafka/data/__consumer_offsets-9 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,685] INFO [Partition __consumer_offsets-9 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-9 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,685] INFO [Partition __consumer_offsets-9 broker=1] Log loaded for partition __consumer_offsets-9 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,685] INFO [Broker id=1] Leader __consumer_offsets-9 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,773] INFO [LogLoader partition=__consumer_offsets-24, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,775] INFO Created log for partition __consumer_offsets-24 in /var/lib/kafka/data/__consumer_offsets-24 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,775] INFO [Partition __consumer_offsets-24 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-24 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,775] INFO [Partition __consumer_offsets-24 broker=1] Log loaded for partition __consumer_offsets-24 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,775] INFO [Broker id=1] Leader __consumer_offsets-24 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,809] INFO [LogLoader partition=__consumer_offsets-31, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,811] INFO Created log for partition __consumer_offsets-31 in /var/lib/kafka/data/__consumer_offsets-31 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,811] INFO [Partition __consumer_offsets-31 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-31 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,812] INFO [Partition __consumer_offsets-31 broker=1] Log loaded for partition __consumer_offsets-31 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,812] INFO [Broker id=1] Leader __consumer_offsets-31 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,860] INFO [LogLoader partition=__consumer_offsets-46, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,862] INFO Created log for partition __consumer_offsets-46 in /var/lib/kafka/data/__consumer_offsets-46 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,862] INFO [Partition __consumer_offsets-46 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-46 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,862] INFO [Partition __consumer_offsets-46 broker=1] Log loaded for partition __consumer_offsets-46 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,862] INFO [Broker id=1] Leader __consumer_offsets-46 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,897] INFO [LogLoader partition=__consumer_offsets-1, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,901] INFO Created log for partition __consumer_offsets-1 in /var/lib/kafka/data/__consumer_offsets-1 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,901] INFO [Partition __consumer_offsets-1 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-1 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,902] INFO [Partition __consumer_offsets-1 broker=1] Log loaded for partition __consumer_offsets-1 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,902] INFO [Broker id=1] Leader __consumer_offsets-1 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,923] INFO [LogLoader partition=__consumer_offsets-16, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,924] INFO Created log for partition __consumer_offsets-16 in /var/lib/kafka/data/__consumer_offsets-16 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,924] INFO [Partition __consumer_offsets-16 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-16 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,924] INFO [Partition __consumer_offsets-16 broker=1] Log loaded for partition __consumer_offsets-16 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,925] INFO [Broker id=1] Leader __consumer_offsets-16 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,945] INFO [LogLoader partition=__consumer_offsets-2, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,946] INFO Created log for partition __consumer_offsets-2 in /var/lib/kafka/data/__consumer_offsets-2 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,946] INFO [Partition __consumer_offsets-2 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-2 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,946] INFO [Partition __consumer_offsets-2 broker=1] Log loaded for partition __consumer_offsets-2 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,947] INFO [Broker id=1] Leader __consumer_offsets-2 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:05,960] INFO [LogLoader partition=__consumer_offsets-25, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:05,961] INFO Created log for partition __consumer_offsets-25 in /var/lib/kafka/data/__consumer_offsets-25 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:05,961] INFO [Partition __consumer_offsets-25 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-25 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,961] INFO [Partition __consumer_offsets-25 broker=1] Log loaded for partition __consumer_offsets-25 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:05,961] INFO [Broker id=1] Leader __consumer_offsets-25 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,015] INFO [LogLoader partition=__consumer_offsets-40, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,016] INFO Created log for partition __consumer_offsets-40 in /var/lib/kafka/data/__consumer_offsets-40 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,016] INFO [Partition __consumer_offsets-40 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-40 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,016] INFO [Partition __consumer_offsets-40 broker=1] Log loaded for partition __consumer_offsets-40 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,016] INFO [Broker id=1] Leader __consumer_offsets-40 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,063] INFO [LogLoader partition=__consumer_offsets-47, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,063] INFO Created log for partition __consumer_offsets-47 in /var/lib/kafka/data/__consumer_offsets-47 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,064] INFO [Partition __consumer_offsets-47 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-47 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,064] INFO [Partition __consumer_offsets-47 broker=1] Log loaded for partition __consumer_offsets-47 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,064] INFO [Broker id=1] Leader __consumer_offsets-47 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,085] INFO [LogLoader partition=__consumer_offsets-17, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,086] INFO Created log for partition __consumer_offsets-17 in /var/lib/kafka/data/__consumer_offsets-17 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,086] INFO [Partition __consumer_offsets-17 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-17 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,086] INFO [Partition __consumer_offsets-17 broker=1] Log loaded for partition __consumer_offsets-17 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,087] INFO [Broker id=1] Leader __consumer_offsets-17 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,099] INFO [LogLoader partition=__consumer_offsets-32, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,100] INFO Created log for partition __consumer_offsets-32 in /var/lib/kafka/data/__consumer_offsets-32 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,100] INFO [Partition __consumer_offsets-32 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-32 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,102] INFO [Partition __consumer_offsets-32 broker=1] Log loaded for partition __consumer_offsets-32 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,102] INFO [Broker id=1] Leader __consumer_offsets-32 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,118] INFO [LogLoader partition=__consumer_offsets-37, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,119] INFO Created log for partition __consumer_offsets-37 in /var/lib/kafka/data/__consumer_offsets-37 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,119] INFO [Partition __consumer_offsets-37 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-37 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,119] INFO [Partition __consumer_offsets-37 broker=1] Log loaded for partition __consumer_offsets-37 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,119] INFO [Broker id=1] Leader __consumer_offsets-37 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,131] INFO [LogLoader partition=__consumer_offsets-7, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,131] INFO Created log for partition __consumer_offsets-7 in /var/lib/kafka/data/__consumer_offsets-7 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,131] INFO [Partition __consumer_offsets-7 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-7 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,132] INFO [Partition __consumer_offsets-7 broker=1] Log loaded for partition __consumer_offsets-7 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,132] INFO [Broker id=1] Leader __consumer_offsets-7 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,139] INFO [LogLoader partition=__consumer_offsets-22, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,140] INFO Created log for partition __consumer_offsets-22 in /var/lib/kafka/data/__consumer_offsets-22 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,140] INFO [Partition __consumer_offsets-22 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-22 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,140] INFO [Partition __consumer_offsets-22 broker=1] Log loaded for partition __consumer_offsets-22 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,140] INFO [Broker id=1] Leader __consumer_offsets-22 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,147] INFO [LogLoader partition=__consumer_offsets-29, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,148] INFO Created log for partition __consumer_offsets-29 in /var/lib/kafka/data/__consumer_offsets-29 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,148] INFO [Partition __consumer_offsets-29 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-29 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,148] INFO [Partition __consumer_offsets-29 broker=1] Log loaded for partition __consumer_offsets-29 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,148] INFO [Broker id=1] Leader __consumer_offsets-29 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,158] INFO [LogLoader partition=__consumer_offsets-44, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,158] INFO Created log for partition __consumer_offsets-44 in /var/lib/kafka/data/__consumer_offsets-44 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,158] INFO [Partition __consumer_offsets-44 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-44 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,158] INFO [Partition __consumer_offsets-44 broker=1] Log loaded for partition __consumer_offsets-44 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,158] INFO [Broker id=1] Leader __consumer_offsets-44 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,167] INFO [LogLoader partition=__consumer_offsets-14, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,169] INFO Created log for partition __consumer_offsets-14 in /var/lib/kafka/data/__consumer_offsets-14 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,169] INFO [Partition __consumer_offsets-14 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-14 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,169] INFO [Partition __consumer_offsets-14 broker=1] Log loaded for partition __consumer_offsets-14 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,169] INFO [Broker id=1] Leader __consumer_offsets-14 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,181] INFO [LogLoader partition=__consumer_offsets-23, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,181] INFO Created log for partition __consumer_offsets-23 in /var/lib/kafka/data/__consumer_offsets-23 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,181] INFO [Partition __consumer_offsets-23 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-23 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,181] INFO [Partition __consumer_offsets-23 broker=1] Log loaded for partition __consumer_offsets-23 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,181] INFO [Broker id=1] Leader __consumer_offsets-23 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,192] INFO [LogLoader partition=__consumer_offsets-38, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,193] INFO Created log for partition __consumer_offsets-38 in /var/lib/kafka/data/__consumer_offsets-38 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,193] INFO [Partition __consumer_offsets-38 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-38 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,193] INFO [Partition __consumer_offsets-38 broker=1] Log loaded for partition __consumer_offsets-38 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,193] INFO [Broker id=1] Leader __consumer_offsets-38 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,204] INFO [LogLoader partition=__consumer_offsets-8, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,205] INFO Created log for partition __consumer_offsets-8 in /var/lib/kafka/data/__consumer_offsets-8 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,205] INFO [Partition __consumer_offsets-8 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-8 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,206] INFO [Partition __consumer_offsets-8 broker=1] Log loaded for partition __consumer_offsets-8 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,206] INFO [Broker id=1] Leader __consumer_offsets-8 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,238] INFO [LogLoader partition=policy-pdp-pap-0, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,239] INFO Created log for partition policy-pdp-pap-0 in /var/lib/kafka/data/policy-pdp-pap-0 with properties {} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,239] INFO [Partition policy-pdp-pap-0 broker=1] No checkpointed highwatermark is found for partition policy-pdp-pap-0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,239] INFO [Partition policy-pdp-pap-0 broker=1] Log loaded for partition policy-pdp-pap-0 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,239] INFO [Broker id=1] Leader policy-pdp-pap-0 with topic id Some(_aOJTUb1QR2K51o55qXHnQ) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,247] INFO [LogLoader partition=__consumer_offsets-45, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,247] INFO Created log for partition __consumer_offsets-45 in /var/lib/kafka/data/__consumer_offsets-45 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,247] INFO [Partition __consumer_offsets-45 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-45 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,248] INFO [Partition __consumer_offsets-45 broker=1] Log loaded for partition __consumer_offsets-45 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,248] INFO [Broker id=1] Leader __consumer_offsets-45 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,254] INFO [LogLoader partition=__consumer_offsets-15, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,254] INFO Created log for partition __consumer_offsets-15 in /var/lib/kafka/data/__consumer_offsets-15 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,254] INFO [Partition __consumer_offsets-15 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-15 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,255] INFO [Partition __consumer_offsets-15 broker=1] Log loaded for partition __consumer_offsets-15 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,255] INFO [Broker id=1] Leader __consumer_offsets-15 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,264] INFO [LogLoader partition=__consumer_offsets-30, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,264] INFO Created log for partition __consumer_offsets-30 in /var/lib/kafka/data/__consumer_offsets-30 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,264] INFO [Partition __consumer_offsets-30 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-30 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,265] INFO [Partition __consumer_offsets-30 broker=1] Log loaded for partition __consumer_offsets-30 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,265] INFO [Broker id=1] Leader __consumer_offsets-30 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,278] INFO [LogLoader partition=__consumer_offsets-0, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,279] INFO Created log for partition __consumer_offsets-0 in /var/lib/kafka/data/__consumer_offsets-0 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,279] INFO [Partition __consumer_offsets-0 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,279] INFO [Partition __consumer_offsets-0 broker=1] Log loaded for partition __consumer_offsets-0 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,279] INFO [Broker id=1] Leader __consumer_offsets-0 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,291] INFO [LogLoader partition=__consumer_offsets-35, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,293] INFO Created log for partition __consumer_offsets-35 in /var/lib/kafka/data/__consumer_offsets-35 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,293] INFO [Partition __consumer_offsets-35 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-35 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,293] INFO [Partition __consumer_offsets-35 broker=1] Log loaded for partition __consumer_offsets-35 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,294] INFO [Broker id=1] Leader __consumer_offsets-35 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,304] INFO [LogLoader partition=__consumer_offsets-5, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,305] INFO Created log for partition __consumer_offsets-5 in /var/lib/kafka/data/__consumer_offsets-5 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,305] INFO [Partition __consumer_offsets-5 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-5 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,305] INFO [Partition __consumer_offsets-5 broker=1] Log loaded for partition __consumer_offsets-5 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,305] INFO [Broker id=1] Leader __consumer_offsets-5 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,316] INFO [LogLoader partition=__consumer_offsets-20, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,317] INFO Created log for partition __consumer_offsets-20 in /var/lib/kafka/data/__consumer_offsets-20 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,317] INFO [Partition __consumer_offsets-20 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-20 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,317] INFO [Partition __consumer_offsets-20 broker=1] Log loaded for partition __consumer_offsets-20 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,317] INFO [Broker id=1] Leader __consumer_offsets-20 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,332] INFO [LogLoader partition=__consumer_offsets-27, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,334] INFO Created log for partition __consumer_offsets-27 in /var/lib/kafka/data/__consumer_offsets-27 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,334] INFO [Partition __consumer_offsets-27 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-27 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,334] INFO [Partition __consumer_offsets-27 broker=1] Log loaded for partition __consumer_offsets-27 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,334] INFO [Broker id=1] Leader __consumer_offsets-27 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,343] INFO [LogLoader partition=__consumer_offsets-42, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,346] INFO Created log for partition __consumer_offsets-42 in /var/lib/kafka/data/__consumer_offsets-42 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,346] INFO [Partition __consumer_offsets-42 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-42 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,346] INFO [Partition __consumer_offsets-42 broker=1] Log loaded for partition __consumer_offsets-42 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,348] INFO [Broker id=1] Leader __consumer_offsets-42 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,369] INFO [LogLoader partition=__consumer_offsets-12, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,370] INFO Created log for partition __consumer_offsets-12 in /var/lib/kafka/data/__consumer_offsets-12 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,370] INFO [Partition __consumer_offsets-12 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-12 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,370] INFO [Partition __consumer_offsets-12 broker=1] Log loaded for partition __consumer_offsets-12 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,370] INFO [Broker id=1] Leader __consumer_offsets-12 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,416] INFO [LogLoader partition=__consumer_offsets-21, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,417] INFO Created log for partition __consumer_offsets-21 in /var/lib/kafka/data/__consumer_offsets-21 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,418] INFO [Partition __consumer_offsets-21 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-21 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,418] INFO [Partition __consumer_offsets-21 broker=1] Log loaded for partition __consumer_offsets-21 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,418] INFO [Broker id=1] Leader __consumer_offsets-21 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,431] INFO [LogLoader partition=__consumer_offsets-36, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,432] INFO Created log for partition __consumer_offsets-36 in /var/lib/kafka/data/__consumer_offsets-36 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,432] INFO [Partition __consumer_offsets-36 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-36 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,432] INFO [Partition __consumer_offsets-36 broker=1] Log loaded for partition __consumer_offsets-36 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,432] INFO [Broker id=1] Leader __consumer_offsets-36 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,452] INFO [LogLoader partition=__consumer_offsets-6, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,453] INFO Created log for partition __consumer_offsets-6 in /var/lib/kafka/data/__consumer_offsets-6 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,453] INFO [Partition __consumer_offsets-6 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-6 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,453] INFO [Partition __consumer_offsets-6 broker=1] Log loaded for partition __consumer_offsets-6 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,453] INFO [Broker id=1] Leader __consumer_offsets-6 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,461] INFO [LogLoader partition=__consumer_offsets-43, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,462] INFO Created log for partition __consumer_offsets-43 in /var/lib/kafka/data/__consumer_offsets-43 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,462] INFO [Partition __consumer_offsets-43 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-43 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,462] INFO [Partition __consumer_offsets-43 broker=1] Log loaded for partition __consumer_offsets-43 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,462] INFO [Broker id=1] Leader __consumer_offsets-43 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,478] INFO [LogLoader partition=__consumer_offsets-13, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,478] INFO Created log for partition __consumer_offsets-13 in /var/lib/kafka/data/__consumer_offsets-13 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,478] INFO [Partition __consumer_offsets-13 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-13 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,478] INFO [Partition __consumer_offsets-13 broker=1] Log loaded for partition __consumer_offsets-13 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,479] INFO [Broker id=1] Leader __consumer_offsets-13 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,504] INFO [LogLoader partition=__consumer_offsets-28, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:06,505] INFO Created log for partition __consumer_offsets-28 in /var/lib/kafka/data/__consumer_offsets-28 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2025-06-18 15:21:06,505] INFO [Partition __consumer_offsets-28 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-28 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,505] INFO [Partition __consumer_offsets-28 broker=1] Log loaded for partition __consumer_offsets-28 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:06,505] INFO [Broker id=1] Leader __consumer_offsets-28 with topic id Some(zkopqUgzQEm01fHoabEl0Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-3 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-18 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-41 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-10 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-33 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-48 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-19 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-34 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-4 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-11 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-26 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-49 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-39 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-9 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-24 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-31 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-46 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-1 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-16 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-2 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-25 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-40 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-47 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-17 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-32 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-37 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-7 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-22 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-29 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-44 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-14 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-23 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-38 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-8 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition policy-pdp-pap-0 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-45 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-15 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-30 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-0 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-35 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-5 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-20 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-27 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-42 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-12 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-21 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-36 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-6 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-43 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-13 (state.change.logger) kafka | [2025-06-18 15:21:06,511] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-28 (state.change.logger) kafka | [2025-06-18 15:21:06,515] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 3 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,516] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-3 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 18 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-18 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 41 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-41 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 10 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-10 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 33 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-33 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 48 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-48 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 19 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-19 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 34 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-34 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 4 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-4 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 11 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-11 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 26 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-26 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,518] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 49 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,518] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-49 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 39 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-39 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 9 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-9 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 24 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-24 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 31 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-31 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 46 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-46 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 1 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-1 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 16 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-16 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 2 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-2 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 25 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-25 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 40 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-40 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 47 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-47 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 17 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-17 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 32 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-32 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 37 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-37 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 7 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-7 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 22 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-22 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 29 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-29 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 44 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-44 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 14 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-14 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 23 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-23 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 38 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-38 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 8 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-8 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 45 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,519] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-45 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,519] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 15 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-15 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 30 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-30 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 0 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-0 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 35 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-35 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 5 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-5 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 20 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-20 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 27 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-27 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 42 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-42 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 12 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-12 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 21 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-21 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 36 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-36 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 6 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-6 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 43 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-43 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 13 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-13 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,520] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 28 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,520] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-28 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,522] INFO [Broker id=1] Finished LeaderAndIsr request in 1416ms correlationId 1 from controller 1 for 51 partitions (state.change.logger) kafka | [2025-06-18 15:21:06,524] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-3 in 7 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,525] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-18 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,526] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-41 in 8 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,526] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-10 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,526] TRACE [Controller id=1 epoch=1] Received response LeaderAndIsrResponseData(errorCode=0, partitionErrors=[], topics=[LeaderAndIsrTopicError(topicId=zkopqUgzQEm01fHoabEl0Q, partitionErrors=[LeaderAndIsrPartitionError(topicName='', partitionIndex=13, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=46, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=9, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=42, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=21, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=17, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=30, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=26, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=5, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=38, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=1, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=34, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=16, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=45, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=12, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=41, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=24, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=20, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=49, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=0, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=29, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=25, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=8, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=37, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=4, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=33, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=15, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=48, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=11, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=44, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=23, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=19, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=32, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=28, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=7, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=40, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=3, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=36, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=47, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=14, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=43, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=10, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=22, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=18, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=31, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=27, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=39, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=6, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=35, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=2, errorCode=0)]), LeaderAndIsrTopicError(topicId=_aOJTUb1QR2K51o55qXHnQ, partitionErrors=[LeaderAndIsrPartitionError(topicName='', partitionIndex=0, errorCode=0)])]) for request LEADER_AND_ISR with correlation id 1 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) kafka | [2025-06-18 15:21:06,527] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-33 in 9 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,527] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-48 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,527] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-19 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,527] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-34 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,528] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-4 in 10 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,528] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-11 in 10 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,530] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-26 in 12 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,530] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-49 in 12 milliseconds for epoch 0, of which 12 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,530] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-39 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,530] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-9 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,530] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-24 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,531] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-31 in 12 milliseconds for epoch 0, of which 12 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,531] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-46 in 12 milliseconds for epoch 0, of which 12 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,531] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-1 in 12 milliseconds for epoch 0, of which 12 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,531] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-16 in 12 milliseconds for epoch 0, of which 12 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,532] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-2 in 13 milliseconds for epoch 0, of which 12 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,532] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-25 in 13 milliseconds for epoch 0, of which 13 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,532] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-40 in 13 milliseconds for epoch 0, of which 13 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,532] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-47 in 13 milliseconds for epoch 0, of which 13 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,532] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-17 in 13 milliseconds for epoch 0, of which 13 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,533] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-32 in 14 milliseconds for epoch 0, of which 13 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,533] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-37 in 14 milliseconds for epoch 0, of which 14 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,533] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-7 in 14 milliseconds for epoch 0, of which 14 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,533] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='policy-pdp-pap', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition policy-pdp-pap-0 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,534] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=13, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-13 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,534] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=46, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-46 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,534] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=9, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-9 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,534] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=42, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-42 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,534] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=21, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-21 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,534] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=17, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-17 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,534] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=30, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-30 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,534] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-22 in 15 milliseconds for epoch 0, of which 14 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=26, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-26 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=5, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-5 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=38, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-38 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=1, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-1 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=34, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-34 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=16, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-16 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-29 in 16 milliseconds for epoch 0, of which 16 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=45, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-45 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=12, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-12 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=41, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-41 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=24, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-24 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,535] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=20, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-20 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=49, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-49 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-0 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=29, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-29 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-44 in 17 milliseconds for epoch 0, of which 16 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=25, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-25 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=8, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-8 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=37, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-37 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=4, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-4 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=33, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-33 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=15, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-15 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=48, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-48 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=11, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-11 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=44, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-44 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=23, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-23 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,536] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-14 in 17 milliseconds for epoch 0, of which 17 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=19, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-19 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=32, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-32 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=28, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-28 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=7, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-7 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=40, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-40 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=3, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-3 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=36, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-36 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=47, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-47 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=14, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-14 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=43, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-43 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=10, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-10 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=22, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-22 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,537] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=18, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-18 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,538] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-23 in 18 milliseconds for epoch 0, of which 18 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,538] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=31, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-31 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,538] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=27, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-27 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,538] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=39, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-39 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,538] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=6, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-6 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,538] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=35, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-35 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,538] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=2, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-2 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,538] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-38 in 19 milliseconds for epoch 0, of which 19 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,538] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-8 in 19 milliseconds for epoch 0, of which 19 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,539] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-45 in 19 milliseconds for epoch 0, of which 19 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,539] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-15 in 19 milliseconds for epoch 0, of which 19 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,539] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-30 in 19 milliseconds for epoch 0, of which 19 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,539] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-0 in 19 milliseconds for epoch 0, of which 19 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,539] INFO [Broker id=1] Add 51 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2025-06-18 15:21:06,539] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-35 in 19 milliseconds for epoch 0, of which 19 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,540] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-5 in 20 milliseconds for epoch 0, of which 19 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,540] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-20 in 20 milliseconds for epoch 0, of which 20 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,540] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-27 in 20 milliseconds for epoch 0, of which 20 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,540] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-42 in 20 milliseconds for epoch 0, of which 20 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,541] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-12 in 21 milliseconds for epoch 0, of which 21 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,541] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-21 in 21 milliseconds for epoch 0, of which 21 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,541] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-36 in 21 milliseconds for epoch 0, of which 21 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,541] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-6 in 21 milliseconds for epoch 0, of which 21 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,541] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-43 in 21 milliseconds for epoch 0, of which 21 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,542] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-13 in 22 milliseconds for epoch 0, of which 22 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,542] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-28 in 22 milliseconds for epoch 0, of which 22 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2025-06-18 15:21:06,541] TRACE [Controller id=1 epoch=1] Received response UpdateMetadataResponseData(errorCode=0) for request UPDATE_METADATA with correlation id 2 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) kafka | [2025-06-18 15:21:06,770] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group da487152-880a-42ad-96cd-a24e5d423167 in Empty state. Created a new member id consumer-da487152-880a-42ad-96cd-a24e5d423167-2-ef05e808-ed55-4ebb-8f5e-20cd76b42a50 and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:06,793] INFO [GroupCoordinator 1]: Preparing to rebalance group da487152-880a-42ad-96cd-a24e5d423167 in state PreparingRebalance with old generation 0 (__consumer_offsets-0) (reason: Adding new member consumer-da487152-880a-42ad-96cd-a24e5d423167-2-ef05e808-ed55-4ebb-8f5e-20cd76b42a50 with group instance id None; client reason: need to re-join with the given member-id: consumer-da487152-880a-42ad-96cd-a24e5d423167-2-ef05e808-ed55-4ebb-8f5e-20cd76b42a50) (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:07,402] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group f4d7781e-9aca-4723-836f-cdddc8c4440f in Empty state. Created a new member id consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3-525882a8-64d3-4949-9f06-cc2010cdfa0c and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:07,407] INFO [GroupCoordinator 1]: Preparing to rebalance group f4d7781e-9aca-4723-836f-cdddc8c4440f in state PreparingRebalance with old generation 0 (__consumer_offsets-28) (reason: Adding new member consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3-525882a8-64d3-4949-9f06-cc2010cdfa0c with group instance id None; client reason: need to re-join with the given member-id: consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3-525882a8-64d3-4949-9f06-cc2010cdfa0c) (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:07,487] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group policy-pap in Empty state. Created a new member id consumer-policy-pap-4-578cdc15-d031-42ac-aa0a-6187e4761e82 and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:07,490] INFO [GroupCoordinator 1]: Preparing to rebalance group policy-pap in state PreparingRebalance with old generation 0 (__consumer_offsets-24) (reason: Adding new member consumer-policy-pap-4-578cdc15-d031-42ac-aa0a-6187e4761e82 with group instance id None; client reason: need to re-join with the given member-id: consumer-policy-pap-4-578cdc15-d031-42ac-aa0a-6187e4761e82) (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:09,805] INFO [GroupCoordinator 1]: Stabilized group da487152-880a-42ad-96cd-a24e5d423167 generation 1 (__consumer_offsets-0) with 1 members (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:09,827] INFO [GroupCoordinator 1]: Assignment received from leader consumer-da487152-880a-42ad-96cd-a24e5d423167-2-ef05e808-ed55-4ebb-8f5e-20cd76b42a50 for group da487152-880a-42ad-96cd-a24e5d423167 for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:10,408] INFO [GroupCoordinator 1]: Stabilized group f4d7781e-9aca-4723-836f-cdddc8c4440f generation 1 (__consumer_offsets-28) with 1 members (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:10,428] INFO [GroupCoordinator 1]: Assignment received from leader consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3-525882a8-64d3-4949-9f06-cc2010cdfa0c for group f4d7781e-9aca-4723-836f-cdddc8c4440f for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:10,490] INFO [GroupCoordinator 1]: Stabilized group policy-pap generation 1 (__consumer_offsets-24) with 1 members (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:10,494] INFO [GroupCoordinator 1]: Assignment received from leader consumer-policy-pap-4-578cdc15-d031-42ac-aa0a-6187e4761e82 for group policy-pap for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:21:11,872] INFO Creating topic policy-notification with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(1)) (kafka.zk.AdminZkClient) kafka | [2025-06-18 15:21:11,896] INFO [Controller id=1] New topics: [Set(policy-notification)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(policy-notification,Some(dNr_bGxRQwyO2qM599eCMQ),Map(policy-notification-0 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController) kafka | [2025-06-18 15:21:11,896] INFO [Controller id=1] New partition creation callback for policy-notification-0 (kafka.controller.KafkaController) kafka | [2025-06-18 15:21:11,896] INFO [Controller id=1 epoch=1] Changed partition policy-notification-0 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2025-06-18 15:21:11,896] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) kafka | [2025-06-18 15:21:11,897] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition policy-notification-0 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2025-06-18 15:21:11,897] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) kafka | [2025-06-18 15:21:11,917] INFO [Controller id=1 epoch=1] Changed partition policy-notification-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isr=List(1), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2025-06-18 15:21:11,918] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='policy-notification', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition policy-notification-0 (state.change.logger) kafka | [2025-06-18 15:21:11,918] INFO [Controller id=1 epoch=1] Sending LeaderAndIsr request to broker 1 with 1 become-leader and 0 become-follower partitions (state.change.logger) kafka | [2025-06-18 15:21:11,918] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet(1) for 1 partitions (state.change.logger) kafka | [2025-06-18 15:21:11,918] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition policy-notification-0 from NewReplica to OnlineReplica (state.change.logger) kafka | [2025-06-18 15:21:11,918] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) kafka | [2025-06-18 15:21:11,919] INFO [Broker id=1] Handling LeaderAndIsr request correlationId 3 from controller 1 for 1 partitions (state.change.logger) kafka | [2025-06-18 15:21:11,919] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='policy-notification', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 3 from controller 1 epoch 1 (state.change.logger) kafka | [2025-06-18 15:21:11,920] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 3 from controller 1 epoch 1 starting the become-leader transition for partition policy-notification-0 (state.change.logger) kafka | [2025-06-18 15:21:11,921] INFO [ReplicaFetcherManager on broker 1] Removed fetcher for partitions Set(policy-notification-0) (kafka.server.ReplicaFetcherManager) kafka | [2025-06-18 15:21:11,921] INFO [Broker id=1] Stopped fetchers as part of LeaderAndIsr request correlationId 3 from controller 1 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger) kafka | [2025-06-18 15:21:11,926] INFO [LogLoader partition=policy-notification-0, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2025-06-18 15:21:11,930] INFO Created log for partition policy-notification-0 in /var/lib/kafka/data/policy-notification-0 with properties {} (kafka.log.LogManager) kafka | [2025-06-18 15:21:11,931] INFO [Partition policy-notification-0 broker=1] No checkpointed highwatermark is found for partition policy-notification-0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:11,932] INFO [Partition policy-notification-0 broker=1] Log loaded for partition policy-notification-0 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2025-06-18 15:21:11,932] INFO [Broker id=1] Leader policy-notification-0 with topic id Some(dNr_bGxRQwyO2qM599eCMQ) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas []. Previous leader epoch was -1. (state.change.logger) kafka | [2025-06-18 15:21:11,951] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 3 from controller 1 epoch 1 for the become-leader transition for partition policy-notification-0 (state.change.logger) kafka | [2025-06-18 15:21:11,952] INFO [Broker id=1] Finished LeaderAndIsr request in 33ms correlationId 3 from controller 1 for 1 partitions (state.change.logger) kafka | [2025-06-18 15:21:11,953] TRACE [Controller id=1 epoch=1] Received response LeaderAndIsrResponseData(errorCode=0, partitionErrors=[], topics=[LeaderAndIsrTopicError(topicId=dNr_bGxRQwyO2qM599eCMQ, partitionErrors=[LeaderAndIsrPartitionError(topicName='', partitionIndex=0, errorCode=0)])]) for request LEADER_AND_ISR with correlation id 3 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) kafka | [2025-06-18 15:21:11,954] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='policy-notification', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition policy-notification-0 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 4 (state.change.logger) kafka | [2025-06-18 15:21:11,954] INFO [Broker id=1] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 4 (state.change.logger) kafka | [2025-06-18 15:21:11,955] TRACE [Controller id=1 epoch=1] Received response UpdateMetadataResponseData(errorCode=0) for request UPDATE_METADATA with correlation id 4 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) kafka | [2025-06-18 15:22:42,326] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group testgrp in Empty state. Created a new member id rdkafka-ff20ffc9-3e87-495e-92ae-753f0780c0dc and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:22:42,328] INFO [GroupCoordinator 1]: Preparing to rebalance group testgrp in state PreparingRebalance with old generation 0 (__consumer_offsets-3) (reason: Adding new member rdkafka-ff20ffc9-3e87-495e-92ae-753f0780c0dc with group instance id None; client reason: not provided) (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:22:45,330] INFO [GroupCoordinator 1]: Stabilized group testgrp generation 1 (__consumer_offsets-3) with 1 members (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:22:45,333] INFO [GroupCoordinator 1]: Assignment received from leader rdkafka-ff20ffc9-3e87-495e-92ae-753f0780c0dc for group testgrp for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:22:45,457] INFO [GroupCoordinator 1]: Preparing to rebalance group testgrp in state PreparingRebalance with old generation 1 (__consumer_offsets-3) (reason: Removing member rdkafka-ff20ffc9-3e87-495e-92ae-753f0780c0dc on LeaveGroup; client reason: not provided) (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:22:45,458] INFO [GroupCoordinator 1]: Group testgrp with generation 2 is now empty (__consumer_offsets-3) (kafka.coordinator.group.GroupCoordinator) kafka | [2025-06-18 15:22:45,459] INFO [GroupCoordinator 1]: Member MemberMetadata(memberId=rdkafka-ff20ffc9-3e87-495e-92ae-753f0780c0dc, groupInstanceId=None, clientId=rdkafka, clientHost=/172.17.0.5, sessionTimeoutMs=45000, rebalanceTimeoutMs=300000, supportedProtocols=List(range, roundrobin)) has left group testgrp through explicit `LeaveGroup`; client reason: not provided (kafka.coordinator.group.GroupCoordinator) policy-api | Waiting for policy-db-migrator port 6824... policy-api | policy-db-migrator (172.17.0.5:6824) open policy-api | Policy api config file: /opt/app/policy/api/etc/apiParameters.yaml policy-api | policy-api | . ____ _ __ _ _ policy-api | /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \ policy-api | ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \ policy-api | \\/ ___)| |_)| | | | | || (_| | ) ) ) ) policy-api | ' |____| .__|_| |_|_| |_\__, | / / / / policy-api | =========|_|==============|___/=/_/_/_/ policy-api | policy-api | :: Spring Boot :: (v3.4.6) policy-api | policy-api | [2025-06-18T15:20:43.350+00:00|INFO|Version|background-preinit] HV000001: Hibernate Validator 8.0.2.Final policy-api | [2025-06-18T15:20:43.416+00:00|INFO|PolicyApiApplication|main] Starting PolicyApiApplication using Java 17.0.15 with PID 39 (/app/api.jar started by policy in /opt/app/policy/api/bin) policy-api | [2025-06-18T15:20:43.418+00:00|INFO|PolicyApiApplication|main] The following 1 profile is active: "default" policy-api | [2025-06-18T15:20:44.930+00:00|INFO|RepositoryConfigurationDelegate|main] Bootstrapping Spring Data JPA repositories in DEFAULT mode. policy-api | [2025-06-18T15:20:45.101+00:00|INFO|RepositoryConfigurationDelegate|main] Finished Spring Data repository scanning in 160 ms. Found 6 JPA repository interfaces. policy-api | [2025-06-18T15:20:45.790+00:00|INFO|TomcatWebServer|main] Tomcat initialized with port 6969 (http) policy-api | [2025-06-18T15:20:45.805+00:00|INFO|Http11NioProtocol|main] Initializing ProtocolHandler ["http-nio-6969"] policy-api | [2025-06-18T15:20:45.807+00:00|INFO|StandardService|main] Starting service [Tomcat] policy-api | [2025-06-18T15:20:45.807+00:00|INFO|StandardEngine|main] Starting Servlet engine: [Apache Tomcat/10.1.41] policy-api | [2025-06-18T15:20:45.850+00:00|INFO|[/policy/api/v1]|main] Initializing Spring embedded WebApplicationContext policy-api | [2025-06-18T15:20:45.851+00:00|INFO|ServletWebServerApplicationContext|main] Root WebApplicationContext: initialization completed in 2368 ms policy-api | [2025-06-18T15:20:46.157+00:00|INFO|LogHelper|main] HHH000204: Processing PersistenceUnitInfo [name: default] policy-api | [2025-06-18T15:20:46.252+00:00|INFO|Version|main] HHH000412: Hibernate ORM core version 6.6.16.Final policy-api | [2025-06-18T15:20:46.308+00:00|INFO|RegionFactoryInitiator|main] HHH000026: Second-level cache disabled policy-api | [2025-06-18T15:20:46.701+00:00|INFO|SpringPersistenceUnitInfo|main] No LoadTimeWeaver setup: ignoring JPA class transformer policy-api | [2025-06-18T15:20:46.736+00:00|INFO|HikariDataSource|main] HikariPool-1 - Starting... policy-api | [2025-06-18T15:20:46.969+00:00|INFO|HikariPool|main] HikariPool-1 - Added connection org.postgresql.jdbc.PgConnection@612bb755 policy-api | [2025-06-18T15:20:46.971+00:00|INFO|HikariDataSource|main] HikariPool-1 - Start completed. policy-api | [2025-06-18T15:20:47.063+00:00|INFO|pooling|main] HHH10001005: Database info: policy-api | Database JDBC URL [Connecting through datasource 'HikariDataSource (HikariPool-1)'] policy-api | Database driver: undefined/unknown policy-api | Database version: 16.4 policy-api | Autocommit mode: undefined/unknown policy-api | Isolation level: undefined/unknown policy-api | Minimum pool size: undefined/unknown policy-api | Maximum pool size: undefined/unknown policy-api | [2025-06-18T15:20:49.261+00:00|INFO|JtaPlatformInitiator|main] HHH000489: No JTA platform available (set 'hibernate.transaction.jta.platform' to enable JTA platform integration) policy-api | [2025-06-18T15:20:49.273+00:00|INFO|LocalContainerEntityManagerFactoryBean|main] Initialized JPA EntityManagerFactory for persistence unit 'default' policy-api | [2025-06-18T15:20:49.895+00:00|WARN|ApiDatabaseInitializer|main] Detected multi-versioned type: policytypes/onap.policies.monitoring.tcagen2.v2.yaml policy-api | [2025-06-18T15:20:50.796+00:00|INFO|ApiDatabaseInitializer|main] Multi-versioned Service Template [onap.policies.Monitoring, onap.policies.monitoring.tcagen2] policy-api | [2025-06-18T15:20:51.916+00:00|WARN|JpaBaseConfiguration$JpaWebConfiguration|main] spring.jpa.open-in-view is enabled by default. Therefore, database queries may be performed during view rendering. Explicitly configure spring.jpa.open-in-view to disable this warning policy-api | [2025-06-18T15:20:51.961+00:00|INFO|InitializeUserDetailsBeanManagerConfigurer$InitializeUserDetailsManagerConfigurer|main] Global AuthenticationManager configured with UserDetailsService bean with name inMemoryUserDetailsManager policy-api | [2025-06-18T15:20:52.610+00:00|INFO|EndpointLinksResolver|main] Exposing 3 endpoints beneath base path '' policy-api | [2025-06-18T15:20:52.755+00:00|INFO|Http11NioProtocol|main] Starting ProtocolHandler ["http-nio-6969"] policy-api | [2025-06-18T15:20:52.772+00:00|INFO|TomcatWebServer|main] Tomcat started on port 6969 (http) with context path '/policy/api/v1' policy-api | [2025-06-18T15:20:52.793+00:00|INFO|PolicyApiApplication|main] Started PolicyApiApplication in 10.28 seconds (process running for 10.871) policy-api | [2025-06-18T15:21:39.919+00:00|INFO|[/policy/api/v1]|http-nio-6969-exec-2] Initializing Spring DispatcherServlet 'dispatcherServlet' policy-api | [2025-06-18T15:21:39.919+00:00|INFO|DispatcherServlet|http-nio-6969-exec-2] Initializing Servlet 'dispatcherServlet' policy-api | [2025-06-18T15:21:39.921+00:00|INFO|DispatcherServlet|http-nio-6969-exec-2] Completed initialization in 1 ms policy-api | [2025-06-18T15:22:18.058+00:00|INFO|OrderedServiceImpl|http-nio-6969-exec-3] ***** OrderedServiceImpl implementers: policy-api | [] policy-csit | Invoking the robot tests from: xacml-pdp-test.robot xacml-pdp-slas.robot policy-csit | Run Robot test policy-csit | ROBOT_VARIABLES=-v DATA:/opt/robotworkspace/models/models-examples/src/main/resources/policies policy-csit | -v NODETEMPLATES:/opt/robotworkspace/models/models-examples/src/main/resources/nodetemplates policy-csit | -v POLICY_API_IP:policy-api:6969 policy-csit | -v POLICY_RUNTIME_ACM_IP:policy-clamp-runtime-acm:6969 policy-csit | -v POLICY_PARTICIPANT_SIM_IP:policy-clamp-ac-sim-ppnt:6969 policy-csit | -v POLICY_PAP_IP:policy-pap:6969 policy-csit | -v APEX_IP:policy-apex-pdp:6969 policy-csit | -v APEX_EVENTS_IP:policy-apex-pdp:23324 policy-csit | -v KAFKA_IP:kafka:9092 policy-csit | -v PROMETHEUS_IP:prometheus:9090 policy-csit | -v POLICY_PDPX_IP:policy-xacml-pdp:6969 policy-csit | -v POLICY_OPA_IP:policy-opa-pdp:8282 policy-csit | -v POLICY_DROOLS_IP:policy-drools-pdp:9696 policy-csit | -v DROOLS_IP:policy-drools-apps:6969 policy-csit | -v DROOLS_IP_2:policy-drools-apps:9696 policy-csit | -v TEMP_FOLDER:/tmp/distribution policy-csit | -v DISTRIBUTION_IP:policy-distribution:6969 policy-csit | -v TEST_ENV:docker policy-csit | -v JAEGER_IP:jaeger:16686 policy-csit | Starting Robot test suites ... policy-csit | ============================================================================== policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas policy-csit | ============================================================================== policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas.Xacml-Pdp-Test policy-csit | ============================================================================== policy-csit | Healthcheck :: Verify policy xacml-pdp health check | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | Metrics :: Verify policy-xacml-pdp is exporting prometheus metrics | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | MakeTopics :: Creates the Policy topics | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | ExecuteXacmlPolicy | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas.Xacml-Pdp-Test | PASS | policy-csit | 4 tests, 4 passed, 0 failed policy-csit | ============================================================================== policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas.Xacml-Pdp-Slas policy-csit | ============================================================================== policy-csit | WaitForPrometheusServer :: Sleep time to wait for Prometheus serve... | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | ValidatePolicyDecisionsTotalCounter :: Validate policy decision co... | PASS | policy-csit | ------------------------------------------------------------------------------ policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas.Xacml-Pdp-Slas | PASS | policy-csit | 2 tests, 2 passed, 0 failed policy-csit | ============================================================================== policy-csit | Xacml-Pdp-Test & Xacml-Pdp-Slas | PASS | policy-csit | 6 tests, 6 passed, 0 failed policy-csit | ============================================================================== policy-csit | Output: /tmp/results/output.xml policy-csit | Log: /tmp/results/log.html policy-csit | Report: /tmp/results/report.html policy-csit | RESULT: 0 policy-db-migrator | Waiting for postgres port 5432... policy-db-migrator | nc: connect to postgres (172.17.0.3) port 5432 (tcp) failed: Connection refused policy-db-migrator | nc: connect to postgres (172.17.0.3) port 5432 (tcp) failed: Connection refused policy-db-migrator | nc: connect to postgres (172.17.0.3) port 5432 (tcp) failed: Connection refused policy-db-migrator | nc: connect to postgres (172.17.0.3) port 5432 (tcp) failed: Connection refused policy-db-migrator | Connection to postgres (172.17.0.3) 5432 port [tcp/postgresql] succeeded! policy-db-migrator | Initializing policyadmin... policy-db-migrator | 321 blocks policy-db-migrator | Preparing upgrade release version: 0800 policy-db-migrator | Preparing upgrade release version: 0900 policy-db-migrator | Preparing upgrade release version: 1000 policy-db-migrator | Preparing upgrade release version: 1100 policy-db-migrator | Preparing upgrade release version: 1200 policy-db-migrator | Preparing upgrade release version: 1300 policy-db-migrator | Done policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | CREATE TABLE policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | name | version policy-db-migrator | -------------+--------- policy-db-migrator | policyadmin | 0 policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | id | script | operation | from_version | to_version | tag | success | attime policy-db-migrator | ----+--------+-----------+--------------+------------+-----+---------+-------- policy-db-migrator | (0 rows) policy-db-migrator | policy-db-migrator | policyadmin: upgrade available: 0 -> 1300 policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "policyadmin_schema_changelog" already exists, skipping policy-db-migrator | upgrade: 0 -> 1300 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-jpapdpgroup_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0110-jpapdpstatistics_enginestats.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0120-jpapdpsubgroup_policies.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0130-jpapdpsubgroup_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0140-jpapdpsubgroup_supportedpolicytypes.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0150-jpatoscacapabilityassignment_attributes.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0160-jpatoscacapabilityassignment_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0170-jpatoscacapabilityassignment_occurrences.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0180-jpatoscacapabilityassignment_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0190-jpatoscacapabilitytype_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0200-jpatoscacapabilitytype_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0210-jpatoscadatatype_constraints.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0220-jpatoscadatatype_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0230-jpatoscadatatype_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0240-jpatoscanodetemplate_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0250-jpatoscanodetemplate_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0260-jpatoscanodetype_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0270-jpatoscanodetype_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0280-jpatoscapolicy_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0290-jpatoscapolicy_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0300-jpatoscapolicy_targets.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0310-jpatoscapolicytype_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0320-jpatoscapolicytype_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0330-jpatoscapolicytype_targets.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0340-jpatoscapolicytype_triggers.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0350-jpatoscaproperty_constraints.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0360-jpatoscaproperty_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0370-jpatoscarelationshiptype_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0380-jpatoscarelationshiptype_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0390-jpatoscarequirement_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0400-jpatoscarequirement_occurrences.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0410-jpatoscarequirement_properties.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0420-jpatoscaservicetemplate_metadata.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0430-jpatoscatopologytemplate_inputs.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0440-pdpgroup_pdpsubgroup.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0450-pdpgroup.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0460-pdppolicystatus.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0470-pdp.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0480-pdpstatistics.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0490-pdpsubgroup_pdp.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0500-pdpsubgroup.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0510-toscacapabilityassignment.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0520-toscacapabilityassignments.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0530-toscacapabilityassignments_toscacapabilityassignment.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0540-toscacapabilitytype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0550-toscacapabilitytypes.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0560-toscacapabilitytypes_toscacapabilitytype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0570-toscadatatype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0580-toscadatatypes.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0590-toscadatatypes_toscadatatype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0600-toscanodetemplate.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0610-toscanodetemplates.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0620-toscanodetemplates_toscanodetemplate.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0630-toscanodetype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0640-toscanodetypes.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0650-toscanodetypes_toscanodetype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0660-toscaparameter.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0670-toscapolicies.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0680-toscapolicies_toscapolicy.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0690-toscapolicy.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0700-toscapolicytype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0710-toscapolicytypes.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0720-toscapolicytypes_toscapolicytype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0730-toscaproperty.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0740-toscarelationshiptype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0750-toscarelationshiptypes.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0760-toscarelationshiptypes_toscarelationshiptype.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0770-toscarequirement.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0780-toscarequirements.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0790-toscarequirements_toscarequirement.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0800-toscaservicetemplate.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0810-toscatopologytemplate.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0820-toscatrigger.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0830-FK_ToscaNodeTemplate_capabilitiesName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0840-FK_ToscaNodeTemplate_requirementsName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0850-FK_ToscaNodeType_requirementsName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0860-FK_ToscaServiceTemplate_capabilityTypesName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0870-FK_ToscaServiceTemplate_dataTypesName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0880-FK_ToscaServiceTemplate_nodeTypesName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0890-FK_ToscaServiceTemplate_policyTypesName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0900-FK_ToscaServiceTemplate_relationshipTypesName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0910-FK_ToscaTopologyTemplate_nodeTemplatesName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0920-FK_ToscaTopologyTemplate_policyName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0940-PdpPolicyStatus_PdpGroup.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0950-TscaServiceTemplatetopologyTemplateParentLocalName.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0960-FK_ToscaNodeTemplate_capabilitiesName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0970-FK_ToscaNodeTemplate_requirementsName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0980-FK_ToscaNodeType_requirementsName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0990-FK_ToscaServiceTemplate_capabilityTypesName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1000-FK_ToscaServiceTemplate_dataTypesName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1010-FK_ToscaServiceTemplate_nodeTypesName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1020-FK_ToscaServiceTemplate_policyTypesName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1030-FK_ToscaServiceTemplate_relationshipTypesName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1040-FK_ToscaTopologyTemplate_nodeTemplatesName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1050-FK_ToscaTopologyTemplate_policyName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1060-TscaServiceTemplatetopologyTemplateParentLocalName.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-pdp.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0110-idx_tsidx1.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0120-pk_pdpstatistics.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0130-pdpstatistics.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0140-pk_pdpstatistics.sql policy-db-migrator | UPDATE 0 policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0150-pdpstatistics.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0160-jpapdpstatistics_enginestats.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0170-jpapdpstatistics_enginestats.sql policy-db-migrator | UPDATE 0 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0180-jpapdpstatistics_enginestats.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0190-jpapolicyaudit.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0200-JpaPolicyAuditIndex_timestamp.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0210-sequence.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0220-sequence.sql policy-db-migrator | INSERT 0 1 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-jpatoscapolicy_targets.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0110-jpatoscapolicytype_targets.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0120-toscatrigger.sql policy-db-migrator | DROP TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0130-jpatoscapolicytype_triggers.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0140-toscaparameter.sql policy-db-migrator | DROP TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0150-toscaproperty.sql policy-db-migrator | DROP TABLE policy-db-migrator | DROP TABLE policy-db-migrator | DROP TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0160-jpapolicyaudit_pk.sql policy-db-migrator | ALTER TABLE policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0170-pdpstatistics_pk.sql policy-db-migrator | ALTER TABLE policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0180-jpatoscanodetemplate_metadata.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-upgrade.sql policy-db-migrator | msg policy-db-migrator | --------------------------- policy-db-migrator | upgrade to 1100 completed policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-jpapolicyaudit_renameuser.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0110-idx_tsidx1.sql policy-db-migrator | DROP INDEX policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0120-audit_sequence.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0130-statistics_sequence.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-pdpstatistics.sql policy-db-migrator | DROP TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0110-jpapdpstatistics_enginestats.sql policy-db-migrator | DROP TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0120-statistics_sequence.sql policy-db-migrator | DROP TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | INSERT 0 1 policy-db-migrator | policyadmin: OK: upgrade (1300) policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "policyadmin_schema_changelog" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | name | version policy-db-migrator | -------------+--------- policy-db-migrator | policyadmin | 1300 policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | id | script | operation | from_version | to_version | tag | success | attime policy-db-migrator | -----+---------------------------------------------------------------+-----------+--------------+------------+-------------------+---------+---------------------------- policy-db-migrator | 1 | 0100-jpapdpgroup_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:26.539736 policy-db-migrator | 2 | 0110-jpapdpstatistics_enginestats.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:26.603483 policy-db-migrator | 3 | 0120-jpapdpsubgroup_policies.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:26.679708 policy-db-migrator | 4 | 0130-jpapdpsubgroup_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:26.733382 policy-db-migrator | 5 | 0140-jpapdpsubgroup_supportedpolicytypes.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:26.785127 policy-db-migrator | 6 | 0150-jpatoscacapabilityassignment_attributes.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:26.841079 policy-db-migrator | 7 | 0160-jpatoscacapabilityassignment_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:26.909735 policy-db-migrator | 8 | 0170-jpatoscacapabilityassignment_occurrences.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:26.956106 policy-db-migrator | 9 | 0180-jpatoscacapabilityassignment_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.035937 policy-db-migrator | 10 | 0190-jpatoscacapabilitytype_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.092933 policy-db-migrator | 11 | 0200-jpatoscacapabilitytype_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.142945 policy-db-migrator | 12 | 0210-jpatoscadatatype_constraints.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.21026 policy-db-migrator | 13 | 0220-jpatoscadatatype_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.277666 policy-db-migrator | 14 | 0230-jpatoscadatatype_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.326163 policy-db-migrator | 15 | 0240-jpatoscanodetemplate_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.381798 policy-db-migrator | 16 | 0250-jpatoscanodetemplate_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.43317 policy-db-migrator | 17 | 0260-jpatoscanodetype_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.517389 policy-db-migrator | 18 | 0270-jpatoscanodetype_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.576954 policy-db-migrator | 19 | 0280-jpatoscapolicy_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.663976 policy-db-migrator | 20 | 0290-jpatoscapolicy_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.719891 policy-db-migrator | 21 | 0300-jpatoscapolicy_targets.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.768048 policy-db-migrator | 22 | 0310-jpatoscapolicytype_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.889133 policy-db-migrator | 23 | 0320-jpatoscapolicytype_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.937031 policy-db-migrator | 24 | 0330-jpatoscapolicytype_targets.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:27.99615 policy-db-migrator | 25 | 0340-jpatoscapolicytype_triggers.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.057592 policy-db-migrator | 26 | 0350-jpatoscaproperty_constraints.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.11587 policy-db-migrator | 27 | 0360-jpatoscaproperty_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.177922 policy-db-migrator | 28 | 0370-jpatoscarelationshiptype_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.229305 policy-db-migrator | 29 | 0380-jpatoscarelationshiptype_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.298607 policy-db-migrator | 30 | 0390-jpatoscarequirement_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.362186 policy-db-migrator | 31 | 0400-jpatoscarequirement_occurrences.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.430809 policy-db-migrator | 32 | 0410-jpatoscarequirement_properties.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.488228 policy-db-migrator | 33 | 0420-jpatoscaservicetemplate_metadata.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.571548 policy-db-migrator | 34 | 0430-jpatoscatopologytemplate_inputs.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.63643 policy-db-migrator | 35 | 0440-pdpgroup_pdpsubgroup.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.69705 policy-db-migrator | 36 | 0450-pdpgroup.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.768089 policy-db-migrator | 37 | 0460-pdppolicystatus.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.829528 policy-db-migrator | 38 | 0470-pdp.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:28.933492 policy-db-migrator | 39 | 0480-pdpstatistics.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.000091 policy-db-migrator | 40 | 0490-pdpsubgroup_pdp.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.104997 policy-db-migrator | 41 | 0500-pdpsubgroup.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.167246 policy-db-migrator | 42 | 0510-toscacapabilityassignment.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.266912 policy-db-migrator | 43 | 0520-toscacapabilityassignments.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.322703 policy-db-migrator | 44 | 0530-toscacapabilityassignments_toscacapabilityassignment.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.37947 policy-db-migrator | 45 | 0540-toscacapabilitytype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.471262 policy-db-migrator | 46 | 0550-toscacapabilitytypes.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.530611 policy-db-migrator | 47 | 0560-toscacapabilitytypes_toscacapabilitytype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.591364 policy-db-migrator | 48 | 0570-toscadatatype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.656942 policy-db-migrator | 49 | 0580-toscadatatypes.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.703177 policy-db-migrator | 50 | 0590-toscadatatypes_toscadatatype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.76001 policy-db-migrator | 51 | 0600-toscanodetemplate.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.825666 policy-db-migrator | 52 | 0610-toscanodetemplates.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.891983 policy-db-migrator | 53 | 0620-toscanodetemplates_toscanodetemplate.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:29.961276 policy-db-migrator | 54 | 0630-toscanodetype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.011785 policy-db-migrator | 55 | 0640-toscanodetypes.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.077067 policy-db-migrator | 56 | 0650-toscanodetypes_toscanodetype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.135161 policy-db-migrator | 57 | 0660-toscaparameter.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.20408 policy-db-migrator | 58 | 0670-toscapolicies.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.26097 policy-db-migrator | 59 | 0680-toscapolicies_toscapolicy.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.317183 policy-db-migrator | 60 | 0690-toscapolicy.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.378038 policy-db-migrator | 61 | 0700-toscapolicytype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.450504 policy-db-migrator | 62 | 0710-toscapolicytypes.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.507677 policy-db-migrator | 63 | 0720-toscapolicytypes_toscapolicytype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.569145 policy-db-migrator | 64 | 0730-toscaproperty.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.622863 policy-db-migrator | 65 | 0740-toscarelationshiptype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.689569 policy-db-migrator | 66 | 0750-toscarelationshiptypes.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.738974 policy-db-migrator | 67 | 0760-toscarelationshiptypes_toscarelationshiptype.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.788473 policy-db-migrator | 68 | 0770-toscarequirement.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.84669 policy-db-migrator | 69 | 0780-toscarequirements.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.901677 policy-db-migrator | 70 | 0790-toscarequirements_toscarequirement.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:30.967193 policy-db-migrator | 71 | 0800-toscaservicetemplate.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.019148 policy-db-migrator | 72 | 0810-toscatopologytemplate.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.073326 policy-db-migrator | 73 | 0820-toscatrigger.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.158911 policy-db-migrator | 74 | 0830-FK_ToscaNodeTemplate_capabilitiesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.211195 policy-db-migrator | 75 | 0840-FK_ToscaNodeTemplate_requirementsName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.297456 policy-db-migrator | 76 | 0850-FK_ToscaNodeType_requirementsName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.347944 policy-db-migrator | 77 | 0860-FK_ToscaServiceTemplate_capabilityTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.397501 policy-db-migrator | 78 | 0870-FK_ToscaServiceTemplate_dataTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.465294 policy-db-migrator | 79 | 0880-FK_ToscaServiceTemplate_nodeTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.517528 policy-db-migrator | 80 | 0890-FK_ToscaServiceTemplate_policyTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.595302 policy-db-migrator | 81 | 0900-FK_ToscaServiceTemplate_relationshipTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.682118 policy-db-migrator | 82 | 0910-FK_ToscaTopologyTemplate_nodeTemplatesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.736698 policy-db-migrator | 83 | 0920-FK_ToscaTopologyTemplate_policyName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.801162 policy-db-migrator | 84 | 0940-PdpPolicyStatus_PdpGroup.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.868855 policy-db-migrator | 85 | 0950-TscaServiceTemplatetopologyTemplateParentLocalName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:31.924194 policy-db-migrator | 86 | 0960-FK_ToscaNodeTemplate_capabilitiesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.016402 policy-db-migrator | 87 | 0970-FK_ToscaNodeTemplate_requirementsName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.070111 policy-db-migrator | 88 | 0980-FK_ToscaNodeType_requirementsName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.169479 policy-db-migrator | 89 | 0990-FK_ToscaServiceTemplate_capabilityTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.217371 policy-db-migrator | 90 | 1000-FK_ToscaServiceTemplate_dataTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.26524 policy-db-migrator | 91 | 1010-FK_ToscaServiceTemplate_nodeTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.330883 policy-db-migrator | 92 | 1020-FK_ToscaServiceTemplate_policyTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.389862 policy-db-migrator | 93 | 1030-FK_ToscaServiceTemplate_relationshipTypesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.443044 policy-db-migrator | 94 | 1040-FK_ToscaTopologyTemplate_nodeTemplatesName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.510669 policy-db-migrator | 95 | 1050-FK_ToscaTopologyTemplate_policyName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.564063 policy-db-migrator | 96 | 1060-TscaServiceTemplatetopologyTemplateParentLocalName.sql | upgrade | 0 | 0800 | 1806251520260800u | 1 | 2025-06-18 15:20:32.612377 policy-db-migrator | 97 | 0100-pdp.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:32.664232 policy-db-migrator | 98 | 0110-idx_tsidx1.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:32.766139 policy-db-migrator | 99 | 0120-pk_pdpstatistics.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:32.882114 policy-db-migrator | 100 | 0130-pdpstatistics.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:32.959458 policy-db-migrator | 101 | 0140-pk_pdpstatistics.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:33.091215 policy-db-migrator | 102 | 0150-pdpstatistics.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:33.162353 policy-db-migrator | 103 | 0160-jpapdpstatistics_enginestats.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:33.226303 policy-db-migrator | 104 | 0170-jpapdpstatistics_enginestats.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:33.275931 policy-db-migrator | 105 | 0180-jpapdpstatistics_enginestats.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:33.339504 policy-db-migrator | 106 | 0190-jpapolicyaudit.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:33.489761 policy-db-migrator | 107 | 0200-JpaPolicyAuditIndex_timestamp.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:33.648096 policy-db-migrator | 108 | 0210-sequence.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:33.791814 policy-db-migrator | 109 | 0220-sequence.sql | upgrade | 0800 | 0900 | 1806251520260900u | 1 | 2025-06-18 15:20:33.884803 policy-db-migrator | 110 | 0100-jpatoscapolicy_targets.sql | upgrade | 0900 | 1000 | 1806251520261000u | 1 | 2025-06-18 15:20:33.96382 policy-db-migrator | 111 | 0110-jpatoscapolicytype_targets.sql | upgrade | 0900 | 1000 | 1806251520261000u | 1 | 2025-06-18 15:20:34.041416 policy-db-migrator | 112 | 0120-toscatrigger.sql | upgrade | 0900 | 1000 | 1806251520261000u | 1 | 2025-06-18 15:20:34.136298 policy-db-migrator | 113 | 0130-jpatoscapolicytype_triggers.sql | upgrade | 0900 | 1000 | 1806251520261000u | 1 | 2025-06-18 15:20:34.315202 policy-db-migrator | 114 | 0140-toscaparameter.sql | upgrade | 0900 | 1000 | 1806251520261000u | 1 | 2025-06-18 15:20:34.389575 policy-db-migrator | 115 | 0150-toscaproperty.sql | upgrade | 0900 | 1000 | 1806251520261000u | 1 | 2025-06-18 15:20:34.507954 policy-db-migrator | 116 | 0160-jpapolicyaudit_pk.sql | upgrade | 0900 | 1000 | 1806251520261000u | 1 | 2025-06-18 15:20:34.636236 policy-db-migrator | 117 | 0170-pdpstatistics_pk.sql | upgrade | 0900 | 1000 | 1806251520261000u | 1 | 2025-06-18 15:20:34.832267 policy-db-migrator | 118 | 0180-jpatoscanodetemplate_metadata.sql | upgrade | 0900 | 1000 | 1806251520261000u | 1 | 2025-06-18 15:20:34.899064 policy-db-migrator | 119 | 0100-upgrade.sql | upgrade | 1000 | 1100 | 1806251520261100u | 1 | 2025-06-18 15:20:34.944122 policy-db-migrator | 120 | 0100-jpapolicyaudit_renameuser.sql | upgrade | 1100 | 1200 | 1806251520261200u | 1 | 2025-06-18 15:20:35.098682 policy-db-migrator | 121 | 0110-idx_tsidx1.sql | upgrade | 1100 | 1200 | 1806251520261200u | 1 | 2025-06-18 15:20:35.168312 policy-db-migrator | 122 | 0120-audit_sequence.sql | upgrade | 1100 | 1200 | 1806251520261200u | 1 | 2025-06-18 15:20:35.234033 policy-db-migrator | 123 | 0130-statistics_sequence.sql | upgrade | 1100 | 1200 | 1806251520261200u | 1 | 2025-06-18 15:20:35.350989 policy-db-migrator | 124 | 0100-pdpstatistics.sql | upgrade | 1200 | 1300 | 1806251520261300u | 1 | 2025-06-18 15:20:35.419143 policy-db-migrator | 125 | 0110-jpapdpstatistics_enginestats.sql | upgrade | 1200 | 1300 | 1806251520261300u | 1 | 2025-06-18 15:20:35.531991 policy-db-migrator | 126 | 0120-statistics_sequence.sql | upgrade | 1200 | 1300 | 1806251520261300u | 1 | 2025-06-18 15:20:35.59638 policy-db-migrator | (126 rows) policy-db-migrator | policy-db-migrator | policyadmin: OK @ 1300 policy-db-migrator | Initializing clampacm... policy-db-migrator | 97 blocks policy-db-migrator | Preparing upgrade release version: 1400 policy-db-migrator | Preparing upgrade release version: 1500 policy-db-migrator | Preparing upgrade release version: 1600 policy-db-migrator | Preparing upgrade release version: 1601 policy-db-migrator | Preparing upgrade release version: 1700 policy-db-migrator | Preparing upgrade release version: 1701 policy-db-migrator | Done policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | name | version policy-db-migrator | ----------+--------- policy-db-migrator | clampacm | 0 policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | id | script | operation | from_version | to_version | tag | success | attime policy-db-migrator | ----+--------+-----------+--------------+------------+-----+---------+-------- policy-db-migrator | (0 rows) policy-db-migrator | policy-db-migrator | clampacm: upgrade available: 0 -> 1701 policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | NOTICE: relation "clampacm_schema_changelog" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | upgrade: 0 -> 1701 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-automationcomposition.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0200-automationcompositiondefinition.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0300-automationcompositionelement.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0400-nodetemplatestate.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0500-participant.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0600-participantsupportedelements.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0700-ac_compositionId_index.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0800-ac_element_fk_index.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0900-dt_element_fk_index.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1000-supported_element_fk_index.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1100-automationcompositionelement_fk.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1200-nodetemplate_fk.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 1300-participantsupportedelements_fk.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-automationcomposition.sql policy-db-migrator | ALTER TABLE policy-db-migrator | UPDATE 0 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0200-automationcompositiondefinition.sql policy-db-migrator | ALTER TABLE policy-db-migrator | UPDATE 0 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0300-participantreplica.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0400-participant.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0500-participant_replica_fk_index.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0600-participant_replica_fk.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0700-automationcompositionelement.sql policy-db-migrator | UPDATE 0 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0800-nodetemplatestate.sql policy-db-migrator | UPDATE 0 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-automationcomposition.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0200-automationcompositionelement.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-automationcomposition.sql policy-db-migrator | UPDATE 0 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0200-automationcompositionelement.sql policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-message.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0200-messagejob.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0300-messagejob_identificationId_index.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-automationcompositionrollback.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0200-automationcomposition.sql policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0300-automationcompositionelement.sql policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0400-automationcomposition_fk.sql policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0500-automationcompositiondefinition.sql policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0600-nodetemplatestate.sql policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0700-mb_identificationId_index.sql policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0800-participantreplica.sql policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0900-participantsupportedacelements.sql policy-db-migrator | UPDATE 0 policy-db-migrator | UPDATE 0 policy-db-migrator | ALTER TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | INSERT 0 1 policy-db-migrator | clampacm: OK: upgrade (1701) policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "clampacm_schema_changelog" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | name | version policy-db-migrator | ----------+--------- policy-db-migrator | clampacm | 1701 policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | id | script | operation | from_version | to_version | tag | success | attime policy-db-migrator | ----+--------------------------------------------+-----------+--------------+------------+-------------------+---------+---------------------------- policy-db-migrator | 1 | 0100-automationcomposition.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:36.307614 policy-db-migrator | 2 | 0200-automationcompositiondefinition.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:36.420702 policy-db-migrator | 3 | 0300-automationcompositionelement.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:36.550082 policy-db-migrator | 4 | 0400-nodetemplatestate.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:36.650759 policy-db-migrator | 5 | 0500-participant.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:36.755744 policy-db-migrator | 6 | 0600-participantsupportedelements.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:36.837015 policy-db-migrator | 7 | 0700-ac_compositionId_index.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:36.942169 policy-db-migrator | 8 | 0800-ac_element_fk_index.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:37.080377 policy-db-migrator | 9 | 0900-dt_element_fk_index.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:37.14645 policy-db-migrator | 10 | 1000-supported_element_fk_index.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:37.221195 policy-db-migrator | 11 | 1100-automationcompositionelement_fk.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:37.284441 policy-db-migrator | 12 | 1200-nodetemplate_fk.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:37.339341 policy-db-migrator | 13 | 1300-participantsupportedelements_fk.sql | upgrade | 1300 | 1400 | 1806251520361400u | 1 | 2025-06-18 15:20:37.425615 policy-db-migrator | 14 | 0100-automationcomposition.sql | upgrade | 1400 | 1500 | 1806251520361500u | 1 | 2025-06-18 15:20:37.485121 policy-db-migrator | 15 | 0200-automationcompositiondefinition.sql | upgrade | 1400 | 1500 | 1806251520361500u | 1 | 2025-06-18 15:20:37.582069 policy-db-migrator | 16 | 0300-participantreplica.sql | upgrade | 1400 | 1500 | 1806251520361500u | 1 | 2025-06-18 15:20:37.670691 policy-db-migrator | 17 | 0400-participant.sql | upgrade | 1400 | 1500 | 1806251520361500u | 1 | 2025-06-18 15:20:37.762411 policy-db-migrator | 18 | 0500-participant_replica_fk_index.sql | upgrade | 1400 | 1500 | 1806251520361500u | 1 | 2025-06-18 15:20:37.856887 policy-db-migrator | 19 | 0600-participant_replica_fk.sql | upgrade | 1400 | 1500 | 1806251520361500u | 1 | 2025-06-18 15:20:37.921065 policy-db-migrator | 20 | 0700-automationcompositionelement.sql | upgrade | 1400 | 1500 | 1806251520361500u | 1 | 2025-06-18 15:20:37.975282 policy-db-migrator | 21 | 0800-nodetemplatestate.sql | upgrade | 1400 | 1500 | 1806251520361500u | 1 | 2025-06-18 15:20:38.025989 policy-db-migrator | 22 | 0100-automationcomposition.sql | upgrade | 1500 | 1600 | 1806251520361600u | 1 | 2025-06-18 15:20:38.08811 policy-db-migrator | 23 | 0200-automationcompositionelement.sql | upgrade | 1500 | 1600 | 1806251520361600u | 1 | 2025-06-18 15:20:38.150391 policy-db-migrator | 24 | 0100-automationcomposition.sql | upgrade | 1501 | 1601 | 1806251520361601u | 1 | 2025-06-18 15:20:38.204609 policy-db-migrator | 25 | 0200-automationcompositionelement.sql | upgrade | 1501 | 1601 | 1806251520361601u | 1 | 2025-06-18 15:20:38.283443 policy-db-migrator | 26 | 0100-message.sql | upgrade | 1600 | 1700 | 1806251520361700u | 1 | 2025-06-18 15:20:38.371056 policy-db-migrator | 27 | 0200-messagejob.sql | upgrade | 1600 | 1700 | 1806251520361700u | 1 | 2025-06-18 15:20:38.493044 policy-db-migrator | 28 | 0300-messagejob_identificationId_index.sql | upgrade | 1600 | 1700 | 1806251520361700u | 1 | 2025-06-18 15:20:38.565956 policy-db-migrator | 29 | 0100-automationcompositionrollback.sql | upgrade | 1601 | 1701 | 1806251520361701u | 1 | 2025-06-18 15:20:38.670172 policy-db-migrator | 30 | 0200-automationcomposition.sql | upgrade | 1601 | 1701 | 1806251520361701u | 1 | 2025-06-18 15:20:38.740702 policy-db-migrator | 31 | 0300-automationcompositionelement.sql | upgrade | 1601 | 1701 | 1806251520361701u | 1 | 2025-06-18 15:20:38.840681 policy-db-migrator | 32 | 0400-automationcomposition_fk.sql | upgrade | 1601 | 1701 | 1806251520361701u | 1 | 2025-06-18 15:20:38.899506 policy-db-migrator | 33 | 0500-automationcompositiondefinition.sql | upgrade | 1601 | 1701 | 1806251520361701u | 1 | 2025-06-18 15:20:39.00325 policy-db-migrator | 34 | 0600-nodetemplatestate.sql | upgrade | 1601 | 1701 | 1806251520361701u | 1 | 2025-06-18 15:20:39.060605 policy-db-migrator | 35 | 0700-mb_identificationId_index.sql | upgrade | 1601 | 1701 | 1806251520361701u | 1 | 2025-06-18 15:20:39.162882 policy-db-migrator | 36 | 0800-participantreplica.sql | upgrade | 1601 | 1701 | 1806251520361701u | 1 | 2025-06-18 15:20:39.228022 policy-db-migrator | 37 | 0900-participantsupportedacelements.sql | upgrade | 1601 | 1701 | 1806251520361701u | 1 | 2025-06-18 15:20:39.285305 policy-db-migrator | (37 rows) policy-db-migrator | policy-db-migrator | clampacm: OK @ 1701 policy-db-migrator | Initializing pooling... policy-db-migrator | 4 blocks policy-db-migrator | Preparing upgrade release version: 1600 policy-db-migrator | Done policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | name | version policy-db-migrator | ---------+--------- policy-db-migrator | pooling | 0 policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | id | script | operation | from_version | to_version | tag | success | attime policy-db-migrator | ----+--------+-----------+--------------+------------+-----+---------+-------- policy-db-migrator | (0 rows) policy-db-migrator | policy-db-migrator | pooling: upgrade available: 0 -> 1600 policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "pooling_schema_changelog" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | upgrade: 0 -> 1600 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-distributed.locking.sql policy-db-migrator | CREATE TABLE policy-db-migrator | CREATE INDEX policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | INSERT 0 1 policy-db-migrator | pooling: OK: upgrade (1600) policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "pooling_schema_changelog" already exists, skipping policy-db-migrator | name | version policy-db-migrator | ---------+--------- policy-db-migrator | pooling | 1600 policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | id | script | operation | from_version | to_version | tag | success | attime policy-db-migrator | ----+------------------------------+-----------+--------------+------------+-------------------+---------+---------------------------- policy-db-migrator | 1 | 0100-distributed.locking.sql | upgrade | 1500 | 1600 | 1806251520401600u | 1 | 2025-06-18 15:20:40.181487 policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | pooling: OK @ 1600 policy-db-migrator | Initializing operationshistory... policy-db-migrator | 6 blocks policy-db-migrator | Preparing upgrade release version: 1600 policy-db-migrator | Done policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | name | version policy-db-migrator | -------------------+--------- policy-db-migrator | operationshistory | 0 policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | id | script | operation | from_version | to_version | tag | success | attime policy-db-migrator | ----+--------+-----------+--------------+------------+-----+---------+-------- policy-db-migrator | (0 rows) policy-db-migrator | policy-db-migrator | operationshistory: upgrade available: 0 -> 1600 policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "operationshistory_schema_changelog" already exists, skipping policy-db-migrator | upgrade: 0 -> 1600 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0100-ophistory_id_sequence.sql policy-db-migrator | CREATE TABLE policy-db-migrator | INSERT 0 1 policy-db-migrator | INSERT 0 1 policy-db-migrator | rc=0 policy-db-migrator | policy-db-migrator | > upgrade 0110-operationshistory.sql policy-db-migrator | CREATE TABLE policy-db-migrator | CREATE INDEX policy-db-migrator | CREATE INDEX policy-db-migrator | INSERT 0 1 policy-db-migrator | INSERT 0 1 policy-db-migrator | operationshistory: OK: upgrade (1600) policy-db-migrator | List of databases policy-db-migrator | Name | Owner | Encoding | Locale Provider | Collate | Ctype | ICU Locale | ICU Rules | Access privileges policy-db-migrator | -------------------+-------------+----------+-----------------+------------+------------+------------+-----------+----------------------------- policy-db-migrator | clampacm | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | migration | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | operationshistory | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyadmin | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | policyclamp | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | pooling | policy_user | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =Tc/policy_user + policy-db-migrator | | | | | | | | | policy_user=CTc/policy_user policy-db-migrator | postgres | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | policy-db-migrator | template0 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | template1 | postgres | UTF8 | libc | en_US.utf8 | en_US.utf8 | | | =c/postgres + policy-db-migrator | | | | | | | | | postgres=CTc/postgres policy-db-migrator | (9 rows) policy-db-migrator | policy-db-migrator | NOTICE: relation "schema_versions" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | NOTICE: relation "operationshistory_schema_changelog" already exists, skipping policy-db-migrator | CREATE TABLE policy-db-migrator | name | version policy-db-migrator | -------------------+--------- policy-db-migrator | operationshistory | 1600 policy-db-migrator | (1 row) policy-db-migrator | policy-db-migrator | id | script | operation | from_version | to_version | tag | success | attime policy-db-migrator | ----+--------------------------------+-----------+--------------+------------+-------------------+---------+---------------------------- policy-db-migrator | 1 | 0100-ophistory_id_sequence.sql | upgrade | 1500 | 1600 | 1806251520401600u | 1 | 2025-06-18 15:20:40.960594 policy-db-migrator | 2 | 0110-operationshistory.sql | upgrade | 1500 | 1600 | 1806251520401600u | 1 | 2025-06-18 15:20:41.110606 policy-db-migrator | (2 rows) policy-db-migrator | policy-db-migrator | operationshistory: OK @ 1600 policy-pap | Waiting for api port 6969... policy-pap | api (172.17.0.6:6969) open policy-pap | Waiting for kafka port 9092... policy-pap | kafka (172.17.0.7:9092) open policy-pap | Policy pap config file: /opt/app/policy/pap/etc/papParameters.yaml policy-pap | PDP group configuration file: /opt/app/policy/pap/etc/mounted/groups.json policy-pap | policy-pap | . ____ _ __ _ _ policy-pap | /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \ policy-pap | ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \ policy-pap | \\/ ___)| |_)| | | | | || (_| | ) ) ) ) policy-pap | ' |____| .__|_| |_|_| |_\__, | / / / / policy-pap | =========|_|==============|___/=/_/_/_/ policy-pap | policy-pap | :: Spring Boot :: (v3.4.6) policy-pap | policy-pap | [2025-06-18T15:20:55.129+00:00|INFO|PolicyPapApplication|main] Starting PolicyPapApplication using Java 17.0.15 with PID 59 (/app/pap.jar started by policy in /opt/app/policy/pap/bin) policy-pap | [2025-06-18T15:20:55.130+00:00|INFO|PolicyPapApplication|main] The following 1 profile is active: "default" policy-pap | [2025-06-18T15:20:56.528+00:00|INFO|RepositoryConfigurationDelegate|main] Bootstrapping Spring Data JPA repositories in DEFAULT mode. policy-pap | [2025-06-18T15:20:56.621+00:00|INFO|RepositoryConfigurationDelegate|main] Finished Spring Data repository scanning in 81 ms. Found 7 JPA repository interfaces. policy-pap | [2025-06-18T15:20:57.574+00:00|INFO|TomcatWebServer|main] Tomcat initialized with port 6969 (http) policy-pap | [2025-06-18T15:20:57.590+00:00|INFO|Http11NioProtocol|main] Initializing ProtocolHandler ["http-nio-6969"] policy-pap | [2025-06-18T15:20:57.592+00:00|INFO|StandardService|main] Starting service [Tomcat] policy-pap | [2025-06-18T15:20:57.592+00:00|INFO|StandardEngine|main] Starting Servlet engine: [Apache Tomcat/10.1.41] policy-pap | [2025-06-18T15:20:57.655+00:00|INFO|[/policy/pap/v1]|main] Initializing Spring embedded WebApplicationContext policy-pap | [2025-06-18T15:20:57.656+00:00|INFO|ServletWebServerApplicationContext|main] Root WebApplicationContext: initialization completed in 2471 ms policy-pap | [2025-06-18T15:20:58.091+00:00|INFO|LogHelper|main] HHH000204: Processing PersistenceUnitInfo [name: default] policy-pap | [2025-06-18T15:20:58.177+00:00|INFO|Version|main] HHH000412: Hibernate ORM core version 6.6.16.Final policy-pap | [2025-06-18T15:20:58.226+00:00|INFO|RegionFactoryInitiator|main] HHH000026: Second-level cache disabled policy-pap | [2025-06-18T15:20:58.642+00:00|INFO|SpringPersistenceUnitInfo|main] No LoadTimeWeaver setup: ignoring JPA class transformer policy-pap | [2025-06-18T15:20:58.686+00:00|INFO|HikariDataSource|main] HikariPool-1 - Starting... policy-pap | [2025-06-18T15:20:58.912+00:00|INFO|HikariPool|main] HikariPool-1 - Added connection org.postgresql.jdbc.PgConnection@1d6a22dd policy-pap | [2025-06-18T15:20:58.914+00:00|INFO|HikariDataSource|main] HikariPool-1 - Start completed. policy-pap | [2025-06-18T15:20:59.002+00:00|INFO|pooling|main] HHH10001005: Database info: policy-pap | Database JDBC URL [Connecting through datasource 'HikariDataSource (HikariPool-1)'] policy-pap | Database driver: undefined/unknown policy-pap | Database version: 16.4 policy-pap | Autocommit mode: undefined/unknown policy-pap | Isolation level: undefined/unknown policy-pap | Minimum pool size: undefined/unknown policy-pap | Maximum pool size: undefined/unknown policy-pap | [2025-06-18T15:21:00.976+00:00|INFO|JtaPlatformInitiator|main] HHH000489: No JTA platform available (set 'hibernate.transaction.jta.platform' to enable JTA platform integration) policy-pap | [2025-06-18T15:21:00.980+00:00|INFO|LocalContainerEntityManagerFactoryBean|main] Initialized JPA EntityManagerFactory for persistence unit 'default' policy-pap | [2025-06-18T15:21:02.175+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-pap | allow.auto.create.topics = true policy-pap | auto.commit.interval.ms = 5000 policy-pap | auto.include.jmx.reporter = true policy-pap | auto.offset.reset = latest policy-pap | bootstrap.servers = [kafka:9092] policy-pap | check.crcs = true policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-1 policy-pap | client.rack = policy-pap | connections.max.idle.ms = 540000 policy-pap | default.api.timeout.ms = 60000 policy-pap | enable.auto.commit = true policy-pap | enable.metrics.push = true policy-pap | exclude.internal.topics = true policy-pap | fetch.max.bytes = 52428800 policy-pap | fetch.max.wait.ms = 500 policy-pap | fetch.min.bytes = 1 policy-pap | group.id = f4d7781e-9aca-4723-836f-cdddc8c4440f policy-pap | group.instance.id = null policy-pap | group.protocol = classic policy-pap | group.remote.assignor = null policy-pap | heartbeat.interval.ms = 3000 policy-pap | interceptor.classes = [] policy-pap | internal.leave.group.on.close = true policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false policy-pap | isolation.level = read_uncommitted policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | max.partition.fetch.bytes = 1048576 policy-pap | max.poll.interval.ms = 300000 policy-pap | max.poll.records = 500 policy-pap | metadata.max.age.ms = 300000 policy-pap | metadata.recovery.strategy = none policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-pap | receive.buffer.bytes = 65536 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retry.backoff.max.ms = 1000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.header.urlencode = false policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | session.timeout.ms = 45000 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | policy-pap | [2025-06-18T15:21:02.235+00:00|INFO|KafkaMetricsCollector|main] initializing Kafka metrics collector policy-pap | [2025-06-18T15:21:02.371+00:00|INFO|AppInfoParser|main] Kafka version: 3.9.1 policy-pap | [2025-06-18T15:21:02.371+00:00|INFO|AppInfoParser|main] Kafka commitId: f745dfdcee2b9851 policy-pap | [2025-06-18T15:21:02.371+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1750260062370 policy-pap | [2025-06-18T15:21:02.374+00:00|INFO|ClassicKafkaConsumer|main] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-1, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Subscribed to topic(s): policy-pdp-pap policy-pap | [2025-06-18T15:21:02.374+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-pap | allow.auto.create.topics = true policy-pap | auto.commit.interval.ms = 5000 policy-pap | auto.include.jmx.reporter = true policy-pap | auto.offset.reset = latest policy-pap | bootstrap.servers = [kafka:9092] policy-pap | check.crcs = true policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = consumer-policy-pap-2 policy-pap | client.rack = policy-pap | connections.max.idle.ms = 540000 policy-pap | default.api.timeout.ms = 60000 policy-pap | enable.auto.commit = true policy-pap | enable.metrics.push = true policy-pap | exclude.internal.topics = true policy-pap | fetch.max.bytes = 52428800 policy-pap | fetch.max.wait.ms = 500 policy-pap | fetch.min.bytes = 1 policy-pap | group.id = policy-pap policy-pap | group.instance.id = null policy-pap | group.protocol = classic policy-pap | group.remote.assignor = null policy-pap | heartbeat.interval.ms = 3000 policy-pap | interceptor.classes = [] policy-pap | internal.leave.group.on.close = true policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false policy-pap | isolation.level = read_uncommitted policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | max.partition.fetch.bytes = 1048576 policy-pap | max.poll.interval.ms = 300000 policy-pap | max.poll.records = 500 policy-pap | metadata.max.age.ms = 300000 policy-pap | metadata.recovery.strategy = none policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-pap | receive.buffer.bytes = 65536 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retry.backoff.max.ms = 1000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.header.urlencode = false policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | session.timeout.ms = 45000 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | policy-pap | [2025-06-18T15:21:02.375+00:00|INFO|KafkaMetricsCollector|main] initializing Kafka metrics collector policy-pap | [2025-06-18T15:21:02.382+00:00|INFO|AppInfoParser|main] Kafka version: 3.9.1 policy-pap | [2025-06-18T15:21:02.382+00:00|INFO|AppInfoParser|main] Kafka commitId: f745dfdcee2b9851 policy-pap | [2025-06-18T15:21:02.382+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1750260062382 policy-pap | [2025-06-18T15:21:02.383+00:00|INFO|ClassicKafkaConsumer|main] [Consumer clientId=consumer-policy-pap-2, groupId=policy-pap] Subscribed to topic(s): policy-pdp-pap policy-pap | [2025-06-18T15:21:02.769+00:00|INFO|PapDatabaseInitializer|main] Created initial pdpGroup in DB - PdpGroups(groups=[PdpGroup(name=defaultGroup, description=The default group that registers all supported policy types and pdps., pdpGroupState=ACTIVE, properties=null, pdpSubgroups=[PdpSubGroup(pdpType=xacml, supportedPolicyTypes=[onap.policies.controlloop.guard.common.FrequencyLimiter 1.0.0, onap.policies.controlloop.guard.common.MinMax 1.0.0, onap.policies.controlloop.guard.common.Blacklist 1.0.0, onap.policies.controlloop.guard.common.Filter 1.0.0, onap.policies.controlloop.guard.coordination.FirstBlocksSecond 1.0.0, onap.policies.monitoring.* 1.0.0, onap.policies.optimization.* 1.0.0, onap.policies.optimization.resource.AffinityPolicy 1.0.0, onap.policies.optimization.resource.DistancePolicy 1.0.0, onap.policies.optimization.resource.HpaPolicy 1.0.0, onap.policies.optimization.resource.OptimizationPolicy 1.0.0, onap.policies.optimization.resource.PciPolicy 1.0.0, onap.policies.optimization.service.QueryPolicy 1.0.0, onap.policies.optimization.service.SubscriberPolicy 1.0.0, onap.policies.optimization.resource.Vim_fit 1.0.0, onap.policies.optimization.resource.VnfPolicy 1.0.0, onap.policies.native.Xacml 1.0.0, onap.policies.Naming 1.0.0, onap.policies.match.* 1.0.0], policies=[SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP 1.0.0], currentInstanceCount=0, desiredInstanceCount=1, properties=null, pdpInstances=null)])]) from /opt/app/policy/pap/etc/mounted/groups.json policy-pap | [2025-06-18T15:21:02.886+00:00|WARN|JpaBaseConfiguration$JpaWebConfiguration|main] spring.jpa.open-in-view is enabled by default. Therefore, database queries may be performed during view rendering. Explicitly configure spring.jpa.open-in-view to disable this warning policy-pap | [2025-06-18T15:21:02.980+00:00|INFO|InitializeUserDetailsBeanManagerConfigurer$InitializeUserDetailsManagerConfigurer|main] Global AuthenticationManager configured with UserDetailsService bean with name inMemoryUserDetailsManager policy-pap | [2025-06-18T15:21:03.176+00:00|INFO|OptionalValidatorFactoryBean|main] Failed to set up a Bean Validation provider: jakarta.validation.NoProviderFoundException: Unable to create a Configuration, because no Jakarta Validation provider could be found. Add a provider like Hibernate Validator (RI) to your classpath. policy-pap | [2025-06-18T15:21:03.874+00:00|INFO|EndpointLinksResolver|main] Exposing 3 endpoints beneath base path '' policy-pap | [2025-06-18T15:21:03.977+00:00|INFO|Http11NioProtocol|main] Starting ProtocolHandler ["http-nio-6969"] policy-pap | [2025-06-18T15:21:03.995+00:00|INFO|TomcatWebServer|main] Tomcat started on port 6969 (http) with context path '/policy/pap/v1' policy-pap | [2025-06-18T15:21:04.019+00:00|INFO|ServiceManager|main] Policy PAP starting policy-pap | [2025-06-18T15:21:04.019+00:00|INFO|ServiceManager|main] Policy PAP starting Meter Registry policy-pap | [2025-06-18T15:21:04.020+00:00|INFO|ServiceManager|main] Policy PAP starting PAP parameters policy-pap | [2025-06-18T15:21:04.021+00:00|INFO|ServiceManager|main] Policy PAP starting Pdp Heartbeat Listener policy-pap | [2025-06-18T15:21:04.021+00:00|INFO|ServiceManager|main] Policy PAP starting Response Request ID Dispatcher policy-pap | [2025-06-18T15:21:04.021+00:00|INFO|ServiceManager|main] Policy PAP starting Heartbeat Request ID Dispatcher policy-pap | [2025-06-18T15:21:04.022+00:00|INFO|ServiceManager|main] Policy PAP starting Response Message Dispatcher policy-pap | [2025-06-18T15:21:04.024+00:00|INFO|TopicBase|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=f4d7781e-9aca-4723-836f-cdddc8c4440f, consumerInstance=policy-pap, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: registering org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher@438cb294 policy-pap | [2025-06-18T15:21:04.034+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=f4d7781e-9aca-4723-836f-cdddc8c4440f, consumerInstance=policy-pap, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-pap | [2025-06-18T15:21:04.035+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-pap | allow.auto.create.topics = true policy-pap | auto.commit.interval.ms = 5000 policy-pap | auto.include.jmx.reporter = true policy-pap | auto.offset.reset = latest policy-pap | bootstrap.servers = [kafka:9092] policy-pap | check.crcs = true policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3 policy-pap | client.rack = policy-pap | connections.max.idle.ms = 540000 policy-pap | default.api.timeout.ms = 60000 policy-pap | enable.auto.commit = true policy-pap | enable.metrics.push = true policy-pap | exclude.internal.topics = true policy-pap | fetch.max.bytes = 52428800 policy-pap | fetch.max.wait.ms = 500 policy-pap | fetch.min.bytes = 1 policy-pap | group.id = f4d7781e-9aca-4723-836f-cdddc8c4440f policy-pap | group.instance.id = null policy-pap | group.protocol = classic policy-pap | group.remote.assignor = null policy-pap | heartbeat.interval.ms = 3000 policy-pap | interceptor.classes = [] policy-pap | internal.leave.group.on.close = true policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false policy-pap | isolation.level = read_uncommitted policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | max.partition.fetch.bytes = 1048576 policy-pap | max.poll.interval.ms = 300000 policy-pap | max.poll.records = 500 policy-pap | metadata.max.age.ms = 300000 policy-pap | metadata.recovery.strategy = none policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-pap | receive.buffer.bytes = 65536 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retry.backoff.max.ms = 1000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.header.urlencode = false policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | session.timeout.ms = 45000 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | policy-pap | [2025-06-18T15:21:04.035+00:00|INFO|KafkaMetricsCollector|main] initializing Kafka metrics collector policy-pap | [2025-06-18T15:21:04.041+00:00|INFO|AppInfoParser|main] Kafka version: 3.9.1 policy-pap | [2025-06-18T15:21:04.042+00:00|INFO|AppInfoParser|main] Kafka commitId: f745dfdcee2b9851 policy-pap | [2025-06-18T15:21:04.042+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1750260064041 policy-pap | [2025-06-18T15:21:04.042+00:00|INFO|ClassicKafkaConsumer|main] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Subscribed to topic(s): policy-pdp-pap policy-pap | [2025-06-18T15:21:04.043+00:00|INFO|ServiceManager|main] Policy PAP starting Heartbeat Message Dispatcher policy-pap | [2025-06-18T15:21:04.043+00:00|INFO|TopicBase|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=policy-pap, consumerInstance=02637adf-aa1a-41ac-9629-4b5d6918709a, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-heartbeat, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: registering org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher@7b787996 policy-pap | [2025-06-18T15:21:04.043+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=policy-pap, consumerInstance=02637adf-aa1a-41ac-9629-4b5d6918709a, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-heartbeat, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-pap | [2025-06-18T15:21:04.043+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-pap | allow.auto.create.topics = true policy-pap | auto.commit.interval.ms = 5000 policy-pap | auto.include.jmx.reporter = true policy-pap | auto.offset.reset = latest policy-pap | bootstrap.servers = [kafka:9092] policy-pap | check.crcs = true policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = consumer-policy-pap-4 policy-pap | client.rack = policy-pap | connections.max.idle.ms = 540000 policy-pap | default.api.timeout.ms = 60000 policy-pap | enable.auto.commit = true policy-pap | enable.metrics.push = true policy-pap | exclude.internal.topics = true policy-pap | fetch.max.bytes = 52428800 policy-pap | fetch.max.wait.ms = 500 policy-pap | fetch.min.bytes = 1 policy-pap | group.id = policy-pap policy-pap | group.instance.id = null policy-pap | group.protocol = classic policy-pap | group.remote.assignor = null policy-pap | heartbeat.interval.ms = 3000 policy-pap | interceptor.classes = [] policy-pap | internal.leave.group.on.close = true policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false policy-pap | isolation.level = read_uncommitted policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | max.partition.fetch.bytes = 1048576 policy-pap | max.poll.interval.ms = 300000 policy-pap | max.poll.records = 500 policy-pap | metadata.max.age.ms = 300000 policy-pap | metadata.recovery.strategy = none policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-pap | receive.buffer.bytes = 65536 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retry.backoff.max.ms = 1000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.header.urlencode = false policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | session.timeout.ms = 45000 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | policy-pap | [2025-06-18T15:21:04.043+00:00|INFO|KafkaMetricsCollector|main] initializing Kafka metrics collector policy-pap | [2025-06-18T15:21:04.049+00:00|INFO|AppInfoParser|main] Kafka version: 3.9.1 policy-pap | [2025-06-18T15:21:04.049+00:00|INFO|AppInfoParser|main] Kafka commitId: f745dfdcee2b9851 policy-pap | [2025-06-18T15:21:04.049+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1750260064049 policy-pap | [2025-06-18T15:21:04.050+00:00|INFO|ClassicKafkaConsumer|main] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Subscribed to topic(s): policy-pdp-pap policy-pap | [2025-06-18T15:21:04.050+00:00|INFO|ServiceManager|main] Policy PAP starting topics policy-pap | [2025-06-18T15:21:04.050+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=policy-pap, consumerInstance=02637adf-aa1a-41ac-9629-4b5d6918709a, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-heartbeat,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-heartbeat, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-pap | [2025-06-18T15:21:04.050+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=f4d7781e-9aca-4723-836f-cdddc8c4440f, consumerInstance=policy-pap, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-pap | [2025-06-18T15:21:04.050+00:00|INFO|InlineBusTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=e1493501-3b88-46eb-a20c-6b629157925a, alive=false, publisher=null]]: starting policy-pap | [2025-06-18T15:21:04.063+00:00|INFO|ProducerConfig|main] ProducerConfig values: policy-pap | acks = -1 policy-pap | auto.include.jmx.reporter = true policy-pap | batch.size = 16384 policy-pap | bootstrap.servers = [kafka:9092] policy-pap | buffer.memory = 33554432 policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = producer-1 policy-pap | compression.gzip.level = -1 policy-pap | compression.lz4.level = 9 policy-pap | compression.type = none policy-pap | compression.zstd.level = 3 policy-pap | connections.max.idle.ms = 540000 policy-pap | delivery.timeout.ms = 120000 policy-pap | enable.idempotence = true policy-pap | enable.metrics.push = true policy-pap | interceptor.classes = [] policy-pap | key.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-pap | linger.ms = 0 policy-pap | max.block.ms = 60000 policy-pap | max.in.flight.requests.per.connection = 5 policy-pap | max.request.size = 1048576 policy-pap | metadata.max.age.ms = 300000 policy-pap | metadata.max.idle.ms = 300000 policy-pap | metadata.recovery.strategy = none policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partitioner.adaptive.partitioning.enable = true policy-pap | partitioner.availability.timeout.ms = 0 policy-pap | partitioner.class = null policy-pap | partitioner.ignore.keys = false policy-pap | receive.buffer.bytes = 32768 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retries = 2147483647 policy-pap | retry.backoff.max.ms = 1000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.header.urlencode = false policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | transaction.timeout.ms = 60000 policy-pap | transactional.id = null policy-pap | value.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-pap | policy-pap | [2025-06-18T15:21:04.064+00:00|INFO|KafkaMetricsCollector|main] initializing Kafka metrics collector policy-pap | [2025-06-18T15:21:04.076+00:00|INFO|KafkaProducer|main] [Producer clientId=producer-1] Instantiated an idempotent producer. policy-pap | [2025-06-18T15:21:04.091+00:00|INFO|AppInfoParser|main] Kafka version: 3.9.1 policy-pap | [2025-06-18T15:21:04.092+00:00|INFO|AppInfoParser|main] Kafka commitId: f745dfdcee2b9851 policy-pap | [2025-06-18T15:21:04.092+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1750260064091 policy-pap | [2025-06-18T15:21:04.092+00:00|INFO|InlineKafkaTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=e1493501-3b88-46eb-a20c-6b629157925a, alive=false, publisher=KafkaPublisherWrapper []]]: KAFKA SINK created policy-pap | [2025-06-18T15:21:04.092+00:00|INFO|InlineBusTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=cfafba30-542f-4f04-9615-889ce44ed100, alive=false, publisher=null]]: starting policy-pap | [2025-06-18T15:21:04.093+00:00|INFO|ProducerConfig|main] ProducerConfig values: policy-pap | acks = -1 policy-pap | auto.include.jmx.reporter = true policy-pap | batch.size = 16384 policy-pap | bootstrap.servers = [kafka:9092] policy-pap | buffer.memory = 33554432 policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = producer-2 policy-pap | compression.gzip.level = -1 policy-pap | compression.lz4.level = 9 policy-pap | compression.type = none policy-pap | compression.zstd.level = 3 policy-pap | connections.max.idle.ms = 540000 policy-pap | delivery.timeout.ms = 120000 policy-pap | enable.idempotence = true policy-pap | enable.metrics.push = true policy-pap | interceptor.classes = [] policy-pap | key.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-pap | linger.ms = 0 policy-pap | max.block.ms = 60000 policy-pap | max.in.flight.requests.per.connection = 5 policy-pap | max.request.size = 1048576 policy-pap | metadata.max.age.ms = 300000 policy-pap | metadata.max.idle.ms = 300000 policy-pap | metadata.recovery.strategy = none policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partitioner.adaptive.partitioning.enable = true policy-pap | partitioner.availability.timeout.ms = 0 policy-pap | partitioner.class = null policy-pap | partitioner.ignore.keys = false policy-pap | receive.buffer.bytes = 32768 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retries = 2147483647 policy-pap | retry.backoff.max.ms = 1000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.header.urlencode = false policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | transaction.timeout.ms = 60000 policy-pap | transactional.id = null policy-pap | value.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-pap | policy-pap | [2025-06-18T15:21:04.093+00:00|INFO|KafkaMetricsCollector|main] initializing Kafka metrics collector policy-pap | [2025-06-18T15:21:04.093+00:00|INFO|KafkaProducer|main] [Producer clientId=producer-2] Instantiated an idempotent producer. policy-pap | [2025-06-18T15:21:04.098+00:00|INFO|AppInfoParser|main] Kafka version: 3.9.1 policy-pap | [2025-06-18T15:21:04.098+00:00|INFO|AppInfoParser|main] Kafka commitId: f745dfdcee2b9851 policy-pap | [2025-06-18T15:21:04.099+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1750260064098 policy-pap | [2025-06-18T15:21:04.099+00:00|INFO|InlineKafkaTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=cfafba30-542f-4f04-9615-889ce44ed100, alive=false, publisher=KafkaPublisherWrapper []]]: KAFKA SINK created policy-pap | [2025-06-18T15:21:04.099+00:00|INFO|ServiceManager|main] Policy PAP starting PAP Activator policy-pap | [2025-06-18T15:21:04.099+00:00|INFO|ServiceManager|main] Policy PAP starting PDP publisher policy-pap | [2025-06-18T15:21:04.101+00:00|INFO|ServiceManager|main] Policy PAP starting Policy Notification publisher policy-pap | [2025-06-18T15:21:04.101+00:00|INFO|ServiceManager|main] Policy PAP starting PDP update timers policy-pap | [2025-06-18T15:21:04.103+00:00|INFO|ServiceManager|main] Policy PAP starting PDP state-change timers policy-pap | [2025-06-18T15:21:04.111+00:00|INFO|ServiceManager|main] Policy PAP starting PDP modification lock policy-pap | [2025-06-18T15:21:04.111+00:00|INFO|ServiceManager|main] Policy PAP starting PDP modification requests policy-pap | [2025-06-18T15:21:04.112+00:00|INFO|TimerManager|Thread-10] timer manager state-change started policy-pap | [2025-06-18T15:21:04.112+00:00|INFO|TimerManager|Thread-9] timer manager update started policy-pap | [2025-06-18T15:21:04.113+00:00|INFO|ServiceManager|main] Policy PAP starting PDP expiration timer policy-pap | [2025-06-18T15:21:04.115+00:00|INFO|ServiceManager|main] Policy PAP started policy-pap | [2025-06-18T15:21:04.115+00:00|INFO|PolicyPapApplication|main] Started PolicyPapApplication in 9.76 seconds (process running for 10.315) policy-pap | [2025-06-18T15:21:04.583+00:00|INFO|Metadata|kafka-producer-network-thread | producer-2] [Producer clientId=producer-2] Cluster ID: 7hUyEsUsSwOpxuI0HNr3Gw policy-pap | [2025-06-18T15:21:04.583+00:00|INFO|Metadata|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] Cluster ID: 7hUyEsUsSwOpxuI0HNr3Gw policy-pap | [2025-06-18T15:21:04.584+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] The metadata response from the cluster reported a recoverable issue with correlation id 3 : {policy-pdp-pap=UNKNOWN_TOPIC_OR_PARTITION} policy-pap | [2025-06-18T15:21:04.585+00:00|INFO|Metadata|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Cluster ID: 7hUyEsUsSwOpxuI0HNr3Gw policy-pap | [2025-06-18T15:21:04.649+00:00|INFO|TransactionManager|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] ProducerId set to 0 with epoch 0 policy-pap | [2025-06-18T15:21:04.649+00:00|INFO|TransactionManager|kafka-producer-network-thread | producer-2] [Producer clientId=producer-2] ProducerId set to 1 with epoch 0 policy-pap | [2025-06-18T15:21:04.672+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] The metadata response from the cluster reported a recoverable issue with correlation id 3 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2025-06-18T15:21:04.672+00:00|INFO|Metadata|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Cluster ID: 7hUyEsUsSwOpxuI0HNr3Gw policy-pap | [2025-06-18T15:21:04.796+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] The metadata response from the cluster reported a recoverable issue with correlation id 7 : {policy-pdp-pap=UNKNOWN_TOPIC_OR_PARTITION} policy-pap | [2025-06-18T15:21:04.870+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] The metadata response from the cluster reported a recoverable issue with correlation id 7 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2025-06-18T15:21:05.090+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] The metadata response from the cluster reported a recoverable issue with correlation id 9 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2025-06-18T15:21:05.153+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] The metadata response from the cluster reported a recoverable issue with correlation id 9 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2025-06-18T15:21:05.556+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] The metadata response from the cluster reported a recoverable issue with correlation id 11 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2025-06-18T15:21:05.573+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] The metadata response from the cluster reported a recoverable issue with correlation id 11 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2025-06-18T15:21:06.365+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] The metadata response from the cluster reported a recoverable issue with correlation id 13 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2025-06-18T15:21:06.468+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] The metadata response from the cluster reported a recoverable issue with correlation id 13 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2025-06-18T15:21:07.379+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Discovered group coordinator kafka:9092 (id: 2147483646 rack: null) policy-pap | [2025-06-18T15:21:07.388+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] (Re-)joining group policy-pap | [2025-06-18T15:21:07.404+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Request joining group due to: need to re-join with the given member-id: consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3-525882a8-64d3-4949-9f06-cc2010cdfa0c policy-pap | [2025-06-18T15:21:07.405+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] (Re-)joining group policy-pap | [2025-06-18T15:21:07.481+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Discovered group coordinator kafka:9092 (id: 2147483646 rack: null) policy-pap | [2025-06-18T15:21:07.484+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] (Re-)joining group policy-pap | [2025-06-18T15:21:07.488+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Request joining group due to: need to re-join with the given member-id: consumer-policy-pap-4-578cdc15-d031-42ac-aa0a-6187e4761e82 policy-pap | [2025-06-18T15:21:07.488+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] (Re-)joining group policy-pap | [2025-06-18T15:21:10.411+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Successfully joined group with generation Generation{generationId=1, memberId='consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3-525882a8-64d3-4949-9f06-cc2010cdfa0c', protocol='range'} policy-pap | [2025-06-18T15:21:10.422+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Finished assignment for group at generation 1: {consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3-525882a8-64d3-4949-9f06-cc2010cdfa0c=Assignment(partitions=[policy-pdp-pap-0])} policy-pap | [2025-06-18T15:21:10.432+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Successfully synced group in generation Generation{generationId=1, memberId='consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3-525882a8-64d3-4949-9f06-cc2010cdfa0c', protocol='range'} policy-pap | [2025-06-18T15:21:10.433+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Notifying assignor about the new Assignment(partitions=[policy-pdp-pap-0]) policy-pap | [2025-06-18T15:21:10.436+00:00|INFO|ConsumerRebalanceListenerInvoker|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Adding newly assigned partitions: policy-pdp-pap-0 policy-pap | [2025-06-18T15:21:10.444+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Found no committed offset for partition policy-pdp-pap-0 policy-pap | [2025-06-18T15:21:10.450+00:00|INFO|SubscriptionState|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-f4d7781e-9aca-4723-836f-cdddc8c4440f-3, groupId=f4d7781e-9aca-4723-836f-cdddc8c4440f] Resetting offset for partition policy-pdp-pap-0 to position FetchPosition{offset=1, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:9092 (id: 1 rack: null)], epoch=0}}. policy-pap | [2025-06-18T15:21:10.491+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Successfully joined group with generation Generation{generationId=1, memberId='consumer-policy-pap-4-578cdc15-d031-42ac-aa0a-6187e4761e82', protocol='range'} policy-pap | [2025-06-18T15:21:10.492+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Finished assignment for group at generation 1: {consumer-policy-pap-4-578cdc15-d031-42ac-aa0a-6187e4761e82=Assignment(partitions=[policy-pdp-pap-0])} policy-pap | [2025-06-18T15:21:10.498+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Successfully synced group in generation Generation{generationId=1, memberId='consumer-policy-pap-4-578cdc15-d031-42ac-aa0a-6187e4761e82', protocol='range'} policy-pap | [2025-06-18T15:21:10.498+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Notifying assignor about the new Assignment(partitions=[policy-pdp-pap-0]) policy-pap | [2025-06-18T15:21:10.498+00:00|INFO|ConsumerRebalanceListenerInvoker|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Adding newly assigned partitions: policy-pdp-pap-0 policy-pap | [2025-06-18T15:21:10.500+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Found no committed offset for partition policy-pdp-pap-0 policy-pap | [2025-06-18T15:21:10.503+00:00|INFO|SubscriptionState|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Resetting offset for partition policy-pdp-pap-0 to position FetchPosition{offset=1, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:9092 (id: 1 rack: null)], epoch=0}}. policy-pap | [2025-06-18T15:21:10.923+00:00|INFO|OrderedServiceImpl|KAFKA-source-policy-heartbeat] ***** OrderedServiceImpl implementers: policy-pap | [] policy-pap | [2025-06-18T15:21:10.924+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"messageName":"PDP_TOPIC_CHECK","requestId":"1a80f344-84db-4fe8-bc72-5ce0a89e0807","timestampMs":1750260065978,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c"} policy-pap | [2025-06-18T15:21:10.924+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"messageName":"PDP_TOPIC_CHECK","requestId":"1a80f344-84db-4fe8-bc72-5ce0a89e0807","timestampMs":1750260065978,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c"} policy-pap | [2025-06-18T15:21:10.928+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_TOPIC_CHECK policy-pap | [2025-06-18T15:21:10.929+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_TOPIC_CHECK policy-pap | [2025-06-18T15:21:10.933+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[],"messageName":"PDP_STATUS","requestId":"0f2cc3be-8452-4160-b2fe-aad52c1534ca","timestampMs":1750260070892,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup"} policy-pap | [2025-06-18T15:21:10.937+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[],"messageName":"PDP_STATUS","requestId":"0f2cc3be-8452-4160-b2fe-aad52c1534ca","timestampMs":1750260070892,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup"} policy-pap | [2025-06-18T15:21:10.941+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-pdp-pap] no listeners for autonomous message of type PdpStatus policy-pap | [2025-06-18T15:21:11.646+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting policy-pap | [2025-06-18T15:21:11.646+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting listener policy-pap | [2025-06-18T15:21:11.647+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting timer policy-pap | [2025-06-18T15:21:11.648+00:00|INFO|TimerManager|KAFKA-source-policy-heartbeat] update timer registered Timer [name=955bae4c-4069-40dc-add3-851fff567958, expireMs=1750260101648] policy-pap | [2025-06-18T15:21:11.650+00:00|INFO|TimerManager|Thread-9] update timer waiting 29998ms Timer [name=955bae4c-4069-40dc-add3-851fff567958, expireMs=1750260101648] policy-pap | [2025-06-18T15:21:11.651+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting enqueue policy-pap | [2025-06-18T15:21:11.651+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate started policy-pap | [2025-06-18T15:21:11.657+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[{"type":"onap.policies.Naming","type_version":"1.0.0","properties":{"policy-instance-name":"ONAP_NF_NAMING_TIMESTAMP","naming-models":[{"naming-type":"VNF","naming-recipe":"AIC_CLOUD_REGION|DELIMITER|CONSTANT|DELIMITER|TIMESTAMP","name-operation":"to_lower_case()","naming-properties":[{"property-name":"AIC_CLOUD_REGION"},{"property-name":"CONSTANT","property-value":"onap-nf"},{"property-name":"TIMESTAMP"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VNFC","naming-recipe":"VNF_NAME|DELIMITER|NFC_NAMING_CODE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"ENTIRETY","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}},{"property-name":"NFC_NAMING_CODE"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VF-MODULE","naming-recipe":"VNF_NAME|DELIMITER|VF_MODULE_LABEL|DELIMITER|VF_MODULE_TYPE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-value":"-","property-name":"DELIMITER"},{"property-name":"VF_MODULE_LABEL"},{"property-name":"VF_MODULE_TYPE"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"PRECEEDING","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}}]}]},"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0","metadata":{"policy-id":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","policy-version":"1.0.0"}}],"messageName":"PDP_UPDATE","requestId":"955bae4c-4069-40dc-add3-851fff567958","timestampMs":1750260071595,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:11.708+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[{"type":"onap.policies.Naming","type_version":"1.0.0","properties":{"policy-instance-name":"ONAP_NF_NAMING_TIMESTAMP","naming-models":[{"naming-type":"VNF","naming-recipe":"AIC_CLOUD_REGION|DELIMITER|CONSTANT|DELIMITER|TIMESTAMP","name-operation":"to_lower_case()","naming-properties":[{"property-name":"AIC_CLOUD_REGION"},{"property-name":"CONSTANT","property-value":"onap-nf"},{"property-name":"TIMESTAMP"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VNFC","naming-recipe":"VNF_NAME|DELIMITER|NFC_NAMING_CODE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"ENTIRETY","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}},{"property-name":"NFC_NAMING_CODE"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VF-MODULE","naming-recipe":"VNF_NAME|DELIMITER|VF_MODULE_LABEL|DELIMITER|VF_MODULE_TYPE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-value":"-","property-name":"DELIMITER"},{"property-name":"VF_MODULE_LABEL"},{"property-name":"VF_MODULE_TYPE"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"PRECEEDING","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}}]}]},"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0","metadata":{"policy-id":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","policy-version":"1.0.0"}}],"messageName":"PDP_UPDATE","requestId":"955bae4c-4069-40dc-add3-851fff567958","timestampMs":1750260071595,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:11.709+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_UPDATE policy-pap | [2025-06-18T15:21:11.711+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[{"type":"onap.policies.Naming","type_version":"1.0.0","properties":{"policy-instance-name":"ONAP_NF_NAMING_TIMESTAMP","naming-models":[{"naming-type":"VNF","naming-recipe":"AIC_CLOUD_REGION|DELIMITER|CONSTANT|DELIMITER|TIMESTAMP","name-operation":"to_lower_case()","naming-properties":[{"property-name":"AIC_CLOUD_REGION"},{"property-name":"CONSTANT","property-value":"onap-nf"},{"property-name":"TIMESTAMP"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VNFC","naming-recipe":"VNF_NAME|DELIMITER|NFC_NAMING_CODE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"ENTIRETY","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}},{"property-name":"NFC_NAMING_CODE"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VF-MODULE","naming-recipe":"VNF_NAME|DELIMITER|VF_MODULE_LABEL|DELIMITER|VF_MODULE_TYPE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-value":"-","property-name":"DELIMITER"},{"property-name":"VF_MODULE_LABEL"},{"property-name":"VF_MODULE_TYPE"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"PRECEEDING","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}}]}]},"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0","metadata":{"policy-id":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","policy-version":"1.0.0"}}],"messageName":"PDP_UPDATE","requestId":"955bae4c-4069-40dc-add3-851fff567958","timestampMs":1750260071595,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:11.711+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_UPDATE policy-pap | [2025-06-18T15:21:11.836+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"response":{"responseTo":"955bae4c-4069-40dc-add3-851fff567958","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"ee08dede-b370-432e-ad86-33ea5b385d5f","timestampMs":1750260071789,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:11.836+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping policy-pap | [2025-06-18T15:21:11.837+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping enqueue policy-pap | [2025-06-18T15:21:11.837+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping timer policy-pap | [2025-06-18T15:21:11.837+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer cancelled Timer [name=955bae4c-4069-40dc-add3-851fff567958, expireMs=1750260101648] policy-pap | [2025-06-18T15:21:11.837+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping listener policy-pap | [2025-06-18T15:21:11.837+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopped policy-pap | [2025-06-18T15:21:11.840+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"response":{"responseTo":"955bae4c-4069-40dc-add3-851fff567958","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"ee08dede-b370-432e-ad86-33ea5b385d5f","timestampMs":1750260071789,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:11.842+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id 955bae4c-4069-40dc-add3-851fff567958 policy-pap | [2025-06-18T15:21:11.849+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate successful policy-pap | [2025-06-18T15:21:11.849+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c start publishing next request policy-pap | [2025-06-18T15:21:11.849+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange starting policy-pap | [2025-06-18T15:21:11.849+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange starting listener policy-pap | [2025-06-18T15:21:11.849+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange starting timer policy-pap | [2025-06-18T15:21:11.849+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] state-change timer registered Timer [name=a42fb34f-6c41-47b7-b8c3-4403ad562bf0, expireMs=1750260101849] policy-pap | [2025-06-18T15:21:11.849+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange starting enqueue policy-pap | [2025-06-18T15:21:11.849+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange started policy-pap | [2025-06-18T15:21:11.849+00:00|INFO|TimerManager|Thread-10] state-change timer waiting 30000ms Timer [name=a42fb34f-6c41-47b7-b8c3-4403ad562bf0, expireMs=1750260101849] policy-pap | [2025-06-18T15:21:11.850+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"a42fb34f-6c41-47b7-b8c3-4403ad562bf0","timestampMs":1750260071595,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:11.850+00:00|INFO|network|Thread-8] [OUT|KAFKA|policy-notification] policy-pap | {"deployed-policies":[{"policy-type":"onap.policies.Naming","policy-type-version":"1.0.0","policy-id":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","policy-version":"1.0.0","success-count":1,"failure-count":0,"incomplete-count":0}],"undeployed-policies":[]} policy-pap | [2025-06-18T15:21:11.855+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"messageName":"PDP_STATUS","requestId":"245482cd-f55f-4400-8775-4300ae6de2c5","timestampMs":1750260071798,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:11.889+00:00|WARN|NetworkClient|kafka-producer-network-thread | producer-2] [Producer clientId=producer-2] The metadata response from the cluster reported a recoverable issue with correlation id 7 : {policy-notification=LEADER_NOT_AVAILABLE} policy-pap | [2025-06-18T15:21:12.184+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"messageName":"PDP_STATUS","requestId":"245482cd-f55f-4400-8775-4300ae6de2c5","timestampMs":1750260071798,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.185+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-pdp-pap] no listeners for autonomous message of type PdpStatus policy-pap | [2025-06-18T15:21:12.191+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"a42fb34f-6c41-47b7-b8c3-4403ad562bf0","timestampMs":1750260071595,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.192+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATE_CHANGE policy-pap | [2025-06-18T15:21:12.192+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","response":{"responseTo":"a42fb34f-6c41-47b7-b8c3-4403ad562bf0","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"098c88d2-eaa6-4bb4-8bb5-d4d90287d8d2","timestampMs":1750260071864,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.482+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange stopping policy-pap | [2025-06-18T15:21:12.482+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange stopping enqueue policy-pap | [2025-06-18T15:21:12.482+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange stopping timer policy-pap | [2025-06-18T15:21:12.482+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] state-change timer cancelled Timer [name=a42fb34f-6c41-47b7-b8c3-4403ad562bf0, expireMs=1750260101849] policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange stopping listener policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange stopped policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpStateChange successful policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c start publishing next request policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting listener policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting timer policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer registered Timer [name=56da2523-d0d0-4363-80a8-ed2e3fe91e07, expireMs=1750260102483] policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting enqueue policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate started policy-pap | [2025-06-18T15:21:12.483+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"56da2523-d0d0-4363-80a8-ed2e3fe91e07","timestampMs":1750260072173,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.487+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"a42fb34f-6c41-47b7-b8c3-4403ad562bf0","timestampMs":1750260071595,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.488+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_STATE_CHANGE policy-pap | [2025-06-18T15:21:12.494+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","response":{"responseTo":"a42fb34f-6c41-47b7-b8c3-4403ad562bf0","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"098c88d2-eaa6-4bb4-8bb5-d4d90287d8d2","timestampMs":1750260071864,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.494+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id a42fb34f-6c41-47b7-b8c3-4403ad562bf0 policy-pap | [2025-06-18T15:21:12.497+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"56da2523-d0d0-4363-80a8-ed2e3fe91e07","timestampMs":1750260072173,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.497+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_UPDATE policy-pap | [2025-06-18T15:21:12.498+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"56da2523-d0d0-4363-80a8-ed2e3fe91e07","timestampMs":1750260072173,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.499+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_UPDATE policy-pap | [2025-06-18T15:21:12.507+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"response":{"responseTo":"56da2523-d0d0-4363-80a8-ed2e3fe91e07","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"7a9878aa-3df8-4441-a737-1eabfb3c9f90","timestampMs":1750260072498,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.507+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping policy-pap | [2025-06-18T15:21:12.507+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping enqueue policy-pap | [2025-06-18T15:21:12.507+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping timer policy-pap | [2025-06-18T15:21:12.508+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer cancelled Timer [name=56da2523-d0d0-4363-80a8-ed2e3fe91e07, expireMs=1750260102483] policy-pap | [2025-06-18T15:21:12.508+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping listener policy-pap | [2025-06-18T15:21:12.508+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopped policy-pap | [2025-06-18T15:21:12.510+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"response":{"responseTo":"56da2523-d0d0-4363-80a8-ed2e3fe91e07","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"7a9878aa-3df8-4441-a737-1eabfb3c9f90","timestampMs":1750260072498,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:21:12.511+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id 56da2523-d0d0-4363-80a8-ed2e3fe91e07 policy-pap | [2025-06-18T15:21:12.517+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate successful policy-pap | [2025-06-18T15:21:12.517+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c has no more requests policy-pap | [2025-06-18T15:21:41.610+00:00|INFO|[/policy/pap/v1]|http-nio-6969-exec-2] Initializing Spring DispatcherServlet 'dispatcherServlet' policy-pap | [2025-06-18T15:21:41.610+00:00|INFO|DispatcherServlet|http-nio-6969-exec-2] Initializing Servlet 'dispatcherServlet' policy-pap | [2025-06-18T15:21:41.612+00:00|INFO|DispatcherServlet|http-nio-6969-exec-2] Completed initialization in 2 ms policy-pap | [2025-06-18T15:21:41.649+00:00|INFO|TimerManager|Thread-9] update timer discarded (expired) Timer [name=955bae4c-4069-40dc-add3-851fff567958, expireMs=1750260101648] policy-pap | [2025-06-18T15:21:41.849+00:00|INFO|TimerManager|Thread-10] state-change timer discarded (expired) Timer [name=a42fb34f-6c41-47b7-b8c3-4403ad562bf0, expireMs=1750260101849] policy-pap | [2025-06-18T15:22:21.193+00:00|INFO|SessionData|http-nio-6969-exec-3] cache group defaultGroup policy-pap | [2025-06-18T15:22:21.194+00:00|INFO|PdpGroupDeployProvider|http-nio-6969-exec-3] add policy onap.restart.tca 1.0.0 to subgroup defaultGroup xacml count=2 policy-pap | [2025-06-18T15:22:21.195+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-3] Registering a deploy for policy onap.restart.tca 1.0.0 policy-pap | [2025-06-18T15:22:21.196+00:00|INFO|SessionData|http-nio-6969-exec-3] add update xacml-3e8650f6-1f12-4eed-9943-19c94576807c defaultGroup xacml policies=1 policy-pap | [2025-06-18T15:22:21.196+00:00|INFO|SessionData|http-nio-6969-exec-3] update cached group defaultGroup policy-pap | [2025-06-18T15:22:21.233+00:00|INFO|SessionData|http-nio-6969-exec-3] use cached group defaultGroup policy-pap | [2025-06-18T15:22:21.233+00:00|INFO|PdpGroupDeployProvider|http-nio-6969-exec-3] add policy OSDF_CASABLANCA.Affinity_Default 1.0.0 to subgroup defaultGroup xacml count=3 policy-pap | [2025-06-18T15:22:21.233+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-3] Registering a deploy for policy OSDF_CASABLANCA.Affinity_Default 1.0.0 policy-pap | [2025-06-18T15:22:21.233+00:00|INFO|SessionData|http-nio-6969-exec-3] add update xacml-3e8650f6-1f12-4eed-9943-19c94576807c defaultGroup xacml policies=2 policy-pap | [2025-06-18T15:22:21.233+00:00|INFO|SessionData|http-nio-6969-exec-3] update cached group defaultGroup policy-pap | [2025-06-18T15:22:21.234+00:00|INFO|SessionData|http-nio-6969-exec-3] updating DB group defaultGroup policy-pap | [2025-06-18T15:22:21.251+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-3] sending audit records to database: [PolicyAudit(auditId=null, pdpGroup=defaultGroup, pdpType=xacml, policy=onap.restart.tca 1.0.0, action=DEPLOYMENT, timestamp=2025-06-18T15:22:21Z, user=policyadmin), PolicyAudit(auditId=null, pdpGroup=defaultGroup, pdpType=xacml, policy=OSDF_CASABLANCA.Affinity_Default 1.0.0, action=DEPLOYMENT, timestamp=2025-06-18T15:22:21Z, user=policyadmin)] policy-pap | [2025-06-18T15:22:21.282+00:00|INFO|ServiceManager|http-nio-6969-exec-3] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting policy-pap | [2025-06-18T15:22:21.282+00:00|INFO|ServiceManager|http-nio-6969-exec-3] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting listener policy-pap | [2025-06-18T15:22:21.282+00:00|INFO|ServiceManager|http-nio-6969-exec-3] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting timer policy-pap | [2025-06-18T15:22:21.282+00:00|INFO|TimerManager|http-nio-6969-exec-3] update timer registered Timer [name=fc472ffb-8754-4900-a29c-b7a068d958fd, expireMs=1750260171282] policy-pap | [2025-06-18T15:22:21.282+00:00|INFO|TimerManager|Thread-9] update timer waiting 30000ms Timer [name=fc472ffb-8754-4900-a29c-b7a068d958fd, expireMs=1750260171282] policy-pap | [2025-06-18T15:22:21.283+00:00|INFO|ServiceManager|http-nio-6969-exec-3] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting enqueue policy-pap | [2025-06-18T15:22:21.283+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","description":"The default group that registers all supported policy types and pdps.","policiesToBeDeployed":[{"type":"onap.policies.monitoring.tcagen2","type_version":"1.0.0","properties":{"tca.policy":{"domain":"measurementsForVfScaling","metricsPerEventName":[{"eventName":"Measurement_vGMUX","controlLoopSchemaType":"VNF","policyScope":"DCAE","policyName":"DCAE.Config_tca-hi-lo","policyVersion":"v0.0.1","thresholds":[{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"EQUAL","severity":"MAJOR","closedLoopEventStatus":"ABATED"},{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"GREATER","severity":"CRITICAL","closedLoopEventStatus":"ONSET"}]}]}},"name":"onap.restart.tca","version":"1.0.0","metadata":{"policy-id":"onap.restart.tca","policy-version":"1.0.0"}},{"type":"onap.policies.optimization.resource.AffinityPolicy","type_version":"1.0.0","properties":{"geography":[],"identity":"affinity_vCPE","scope":[],"affinityProperties":{"qualifier":"same","category":"complex"},"resources":[],"services":[],"applicableResources":"any"},"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0","metadata":{"policy-id":"OSDF_CASABLANCA.Affinity_Default","policy-version":"1.0.0"}}],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"fc472ffb-8754-4900-a29c-b7a068d958fd","timestampMs":1750260141233,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:21.283+00:00|INFO|ServiceManager|http-nio-6969-exec-3] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate started policy-pap | [2025-06-18T15:22:21.291+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","description":"The default group that registers all supported policy types and pdps.","policiesToBeDeployed":[{"type":"onap.policies.monitoring.tcagen2","type_version":"1.0.0","properties":{"tca.policy":{"domain":"measurementsForVfScaling","metricsPerEventName":[{"eventName":"Measurement_vGMUX","controlLoopSchemaType":"VNF","policyScope":"DCAE","policyName":"DCAE.Config_tca-hi-lo","policyVersion":"v0.0.1","thresholds":[{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"EQUAL","severity":"MAJOR","closedLoopEventStatus":"ABATED"},{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"GREATER","severity":"CRITICAL","closedLoopEventStatus":"ONSET"}]}]}},"name":"onap.restart.tca","version":"1.0.0","metadata":{"policy-id":"onap.restart.tca","policy-version":"1.0.0"}},{"type":"onap.policies.optimization.resource.AffinityPolicy","type_version":"1.0.0","properties":{"geography":[],"identity":"affinity_vCPE","scope":[],"affinityProperties":{"qualifier":"same","category":"complex"},"resources":[],"services":[],"applicableResources":"any"},"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0","metadata":{"policy-id":"OSDF_CASABLANCA.Affinity_Default","policy-version":"1.0.0"}}],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"fc472ffb-8754-4900-a29c-b7a068d958fd","timestampMs":1750260141233,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:21.291+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_UPDATE policy-pap | [2025-06-18T15:22:21.293+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","description":"The default group that registers all supported policy types and pdps.","policiesToBeDeployed":[{"type":"onap.policies.monitoring.tcagen2","type_version":"1.0.0","properties":{"tca.policy":{"domain":"measurementsForVfScaling","metricsPerEventName":[{"eventName":"Measurement_vGMUX","controlLoopSchemaType":"VNF","policyScope":"DCAE","policyName":"DCAE.Config_tca-hi-lo","policyVersion":"v0.0.1","thresholds":[{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"EQUAL","severity":"MAJOR","closedLoopEventStatus":"ABATED"},{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"GREATER","severity":"CRITICAL","closedLoopEventStatus":"ONSET"}]}]}},"name":"onap.restart.tca","version":"1.0.0","metadata":{"policy-id":"onap.restart.tca","policy-version":"1.0.0"}},{"type":"onap.policies.optimization.resource.AffinityPolicy","type_version":"1.0.0","properties":{"geography":[],"identity":"affinity_vCPE","scope":[],"affinityProperties":{"qualifier":"same","category":"complex"},"resources":[],"services":[],"applicableResources":"any"},"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0","metadata":{"policy-id":"OSDF_CASABLANCA.Affinity_Default","policy-version":"1.0.0"}}],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"fc472ffb-8754-4900-a29c-b7a068d958fd","timestampMs":1750260141233,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:21.293+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_UPDATE policy-pap | [2025-06-18T15:22:21.869+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"onap.restart.tca","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"response":{"responseTo":"fc472ffb-8754-4900-a29c-b7a068d958fd","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"b429c97c-fe37-468c-9a06-025331b71180","timestampMs":1750260141862,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:21.869+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"onap.restart.tca","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"response":{"responseTo":"fc472ffb-8754-4900-a29c-b7a068d958fd","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"b429c97c-fe37-468c-9a06-025331b71180","timestampMs":1750260141862,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:21.869+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping policy-pap | [2025-06-18T15:22:21.869+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping enqueue policy-pap | [2025-06-18T15:22:21.870+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id fc472ffb-8754-4900-a29c-b7a068d958fd policy-pap | [2025-06-18T15:22:21.869+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping timer policy-pap | [2025-06-18T15:22:21.870+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer cancelled Timer [name=fc472ffb-8754-4900-a29c-b7a068d958fd, expireMs=1750260171282] policy-pap | [2025-06-18T15:22:21.870+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping listener policy-pap | [2025-06-18T15:22:21.870+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopped policy-pap | [2025-06-18T15:22:21.880+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate successful policy-pap | [2025-06-18T15:22:21.880+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c has no more requests policy-pap | [2025-06-18T15:22:21.880+00:00|INFO|network|Thread-8] [OUT|KAFKA|policy-notification] policy-pap | {"deployed-policies":[{"policy-type":"onap.policies.monitoring.tcagen2","policy-type-version":"1.0.0","policy-id":"onap.restart.tca","policy-version":"1.0.0","success-count":1,"failure-count":0,"incomplete-count":0},{"policy-type":"onap.policies.optimization.resource.AffinityPolicy","policy-type-version":"1.0.0","policy-id":"OSDF_CASABLANCA.Affinity_Default","policy-version":"1.0.0","success-count":1,"failure-count":0,"incomplete-count":0}],"undeployed-policies":[]} policy-pap | [2025-06-18T15:22:46.017+00:00|INFO|SessionData|http-nio-6969-exec-5] cache group defaultGroup policy-pap | [2025-06-18T15:22:46.019+00:00|INFO|PdpGroupDeleteProvider|http-nio-6969-exec-5] remove policy onap.restart.tca 1.0.0 from subgroup defaultGroup xacml count=2 policy-pap | [2025-06-18T15:22:46.019+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-5] Registering an undeploy for policy onap.restart.tca 1.0.0 policy-pap | [2025-06-18T15:22:46.019+00:00|INFO|SessionData|http-nio-6969-exec-5] add update xacml-3e8650f6-1f12-4eed-9943-19c94576807c defaultGroup xacml policies=0 policy-pap | [2025-06-18T15:22:46.019+00:00|INFO|SessionData|http-nio-6969-exec-5] update cached group defaultGroup policy-pap | [2025-06-18T15:22:46.019+00:00|INFO|SessionData|http-nio-6969-exec-5] updating DB group defaultGroup policy-pap | [2025-06-18T15:22:46.033+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-5] sending audit records to database: [PolicyAudit(auditId=null, pdpGroup=defaultGroup, pdpType=xacml, policy=onap.restart.tca 1.0.0, action=UNDEPLOYMENT, timestamp=2025-06-18T15:22:46Z, user=policyadmin)] policy-pap | [2025-06-18T15:22:46.044+00:00|INFO|ServiceManager|http-nio-6969-exec-5] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting policy-pap | [2025-06-18T15:22:46.044+00:00|INFO|ServiceManager|http-nio-6969-exec-5] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting listener policy-pap | [2025-06-18T15:22:46.044+00:00|INFO|ServiceManager|http-nio-6969-exec-5] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting timer policy-pap | [2025-06-18T15:22:46.044+00:00|INFO|TimerManager|http-nio-6969-exec-5] update timer registered Timer [name=e29f5c97-3aec-4ad6-b58e-ee474839d46d, expireMs=1750260196044] policy-pap | [2025-06-18T15:22:46.044+00:00|INFO|ServiceManager|http-nio-6969-exec-5] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate starting enqueue policy-pap | [2025-06-18T15:22:46.045+00:00|INFO|ServiceManager|http-nio-6969-exec-5] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate started policy-pap | [2025-06-18T15:22:46.045+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","description":"The default group that registers all supported policy types and pdps.","policiesToBeDeployed":[],"policiesToBeUndeployed":[{"name":"onap.restart.tca","version":"1.0.0"}],"messageName":"PDP_UPDATE","requestId":"e29f5c97-3aec-4ad6-b58e-ee474839d46d","timestampMs":1750260166019,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:46.053+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","description":"The default group that registers all supported policy types and pdps.","policiesToBeDeployed":[],"policiesToBeUndeployed":[{"name":"onap.restart.tca","version":"1.0.0"}],"messageName":"PDP_UPDATE","requestId":"e29f5c97-3aec-4ad6-b58e-ee474839d46d","timestampMs":1750260166019,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:46.053+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_UPDATE policy-pap | [2025-06-18T15:22:46.053+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","description":"The default group that registers all supported policy types and pdps.","policiesToBeDeployed":[],"policiesToBeUndeployed":[{"name":"onap.restart.tca","version":"1.0.0"}],"messageName":"PDP_UPDATE","requestId":"e29f5c97-3aec-4ad6-b58e-ee474839d46d","timestampMs":1750260166019,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:46.054+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_UPDATE policy-pap | [2025-06-18T15:22:46.064+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"response":{"responseTo":"e29f5c97-3aec-4ad6-b58e-ee474839d46d","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"e1486330-7898-4398-9ab4-834a88c41277","timestampMs":1750260166056,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:46.064+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"response":{"responseTo":"e29f5c97-3aec-4ad6-b58e-ee474839d46d","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"e1486330-7898-4398-9ab4-834a88c41277","timestampMs":1750260166056,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:22:46.064+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping policy-pap | [2025-06-18T15:22:46.064+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping enqueue policy-pap | [2025-06-18T15:22:46.064+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping timer policy-pap | [2025-06-18T15:22:46.064+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer cancelled Timer [name=e29f5c97-3aec-4ad6-b58e-ee474839d46d, expireMs=1750260196044] policy-pap | [2025-06-18T15:22:46.064+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopping listener policy-pap | [2025-06-18T15:22:46.064+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate stopped policy-pap | [2025-06-18T15:22:46.064+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id e29f5c97-3aec-4ad6-b58e-ee474839d46d policy-pap | [2025-06-18T15:22:46.080+00:00|INFO|network|Thread-8] [OUT|KAFKA|policy-notification] policy-pap | {"deployed-policies":[],"undeployed-policies":[{"policy-type":"onap.policies.monitoring.tcagen2","policy-type-version":"1.0.0","policy-id":"onap.restart.tca","policy-version":"1.0.0","success-count":1,"failure-count":0,"incomplete-count":0}]} policy-pap | [2025-06-18T15:22:46.080+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c PdpUpdate successful policy-pap | [2025-06-18T15:22:46.081+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] xacml-3e8650f6-1f12-4eed-9943-19c94576807c has no more requests policy-pap | [2025-06-18T15:22:51.283+00:00|INFO|TimerManager|Thread-9] update timer discarded (expired) Timer [name=fc472ffb-8754-4900-a29c-b7a068d958fd, expireMs=1750260171282] policy-pap | [2025-06-18T15:23:04.114+00:00|INFO|PdpModifyRequestMap|pool-3-thread-1] check for PDP records older than 360000ms policy-pap | [2025-06-18T15:23:11.823+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"messageName":"PDP_STATUS","requestId":"30f5a157-2cce-4257-a432-84d1175d1b45","timestampMs":1750260191813,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:23:11.823+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"messageName":"PDP_STATUS","requestId":"30f5a157-2cce-4257-a432-84d1175d1b45","timestampMs":1750260191813,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-pap | [2025-06-18T15:23:11.824+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-pdp-pap] no listeners for autonomous message of type PdpStatus policy-xacml-pdp | Waiting for pap port 6969... policy-xacml-pdp | pap (172.17.0.9:6969) open policy-xacml-pdp | Waiting for kafka port 9092... policy-xacml-pdp | kafka (172.17.0.7:9092) open policy-xacml-pdp | + KEYSTORE=/opt/app/policy/pdpx/etc/ssl/policy-keystore policy-xacml-pdp | + TRUSTSTORE=/opt/app/policy/pdpx/etc/ssl/policy-truststore policy-xacml-pdp | + KEYSTORE_PASSWD=Pol1cy_0nap policy-xacml-pdp | + TRUSTSTORE_PASSWD=Pol1cy_0nap policy-xacml-pdp | + '[' 0 -ge 1 ] policy-xacml-pdp | + CONFIG_FILE= policy-xacml-pdp | + '[' -z ] policy-xacml-pdp | + CONFIG_FILE=/opt/app/policy/pdpx/etc/defaultConfig.json policy-xacml-pdp | + '[' -f /opt/app/policy/pdpx/etc/mounted/policy-truststore ] policy-xacml-pdp | + '[' -f /opt/app/policy/pdpx/etc/mounted/policy-keystore ] policy-xacml-pdp | + '[' -f /opt/app/policy/pdpx/etc/mounted/xacml.properties ] policy-xacml-pdp | + '[' -f /opt/app/policy/pdpx/etc/mounted/logback.xml ] policy-xacml-pdp | Policy Xacml PDP config file: /opt/app/policy/pdpx/etc/defaultConfig.json policy-xacml-pdp | + echo 'Policy Xacml PDP config file: /opt/app/policy/pdpx/etc/defaultConfig.json' policy-xacml-pdp | + /usr/lib/jvm/default-jvm/bin/java -cp '/opt/app/policy/pdpx/etc:/opt/app/policy/pdpx/lib/*' '-Dlogback.configurationFile=/opt/app/policy/pdpx/etc/logback.xml' '-Djavax.net.ssl.keyStore=/opt/app/policy/pdpx/etc/ssl/policy-keystore' '-Djavax.net.ssl.keyStorePassword=Pol1cy_0nap' '-Djavax.net.ssl.trustStore=/opt/app/policy/pdpx/etc/ssl/policy-truststore' '-Djavax.net.ssl.trustStorePassword=Pol1cy_0nap' org.onap.policy.pdpx.main.startstop.Main -c /opt/app/policy/pdpx/etc/defaultConfig.json policy-xacml-pdp | [2025-06-18T15:21:05.198+00:00|INFO|Main|main] Starting policy xacml pdp service with arguments - [-c, /opt/app/policy/pdpx/etc/defaultConfig.json] policy-xacml-pdp | [2025-06-18T15:21:05.294+00:00|INFO|XacmlPdpActivator|main] Activator initializing using org.onap.policy.pdpx.main.parameters.XacmlPdpParameterGroup@37858383 policy-xacml-pdp | [2025-06-18T15:21:05.338+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-xacml-pdp | allow.auto.create.topics = true policy-xacml-pdp | auto.commit.interval.ms = 5000 policy-xacml-pdp | auto.include.jmx.reporter = true policy-xacml-pdp | auto.offset.reset = latest policy-xacml-pdp | bootstrap.servers = [kafka:9092] policy-xacml-pdp | check.crcs = true policy-xacml-pdp | client.dns.lookup = use_all_dns_ips policy-xacml-pdp | client.id = consumer-da487152-880a-42ad-96cd-a24e5d423167-1 policy-xacml-pdp | client.rack = policy-xacml-pdp | connections.max.idle.ms = 540000 policy-xacml-pdp | default.api.timeout.ms = 60000 policy-xacml-pdp | enable.auto.commit = true policy-xacml-pdp | enable.metrics.push = true policy-xacml-pdp | exclude.internal.topics = true policy-xacml-pdp | fetch.max.bytes = 52428800 policy-xacml-pdp | fetch.max.wait.ms = 500 policy-xacml-pdp | fetch.min.bytes = 1 policy-xacml-pdp | group.id = da487152-880a-42ad-96cd-a24e5d423167 policy-xacml-pdp | group.instance.id = null policy-xacml-pdp | group.protocol = classic policy-xacml-pdp | group.remote.assignor = null policy-xacml-pdp | heartbeat.interval.ms = 3000 policy-xacml-pdp | interceptor.classes = [] policy-xacml-pdp | internal.leave.group.on.close = true policy-xacml-pdp | internal.throw.on.fetch.stable.offset.unsupported = false policy-xacml-pdp | isolation.level = read_uncommitted policy-xacml-pdp | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-xacml-pdp | max.partition.fetch.bytes = 1048576 policy-xacml-pdp | max.poll.interval.ms = 300000 policy-xacml-pdp | max.poll.records = 500 policy-xacml-pdp | metadata.max.age.ms = 300000 policy-xacml-pdp | metadata.recovery.strategy = none policy-xacml-pdp | metric.reporters = [] policy-xacml-pdp | metrics.num.samples = 2 policy-xacml-pdp | metrics.recording.level = INFO policy-xacml-pdp | metrics.sample.window.ms = 30000 policy-xacml-pdp | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-xacml-pdp | receive.buffer.bytes = 65536 policy-xacml-pdp | reconnect.backoff.max.ms = 1000 policy-xacml-pdp | reconnect.backoff.ms = 50 policy-xacml-pdp | request.timeout.ms = 30000 policy-xacml-pdp | retry.backoff.max.ms = 1000 policy-xacml-pdp | retry.backoff.ms = 100 policy-xacml-pdp | sasl.client.callback.handler.class = null policy-xacml-pdp | sasl.jaas.config = null policy-xacml-pdp | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-xacml-pdp | sasl.kerberos.min.time.before.relogin = 60000 policy-xacml-pdp | sasl.kerberos.service.name = null policy-xacml-pdp | sasl.kerberos.ticket.renew.jitter = 0.05 policy-xacml-pdp | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-xacml-pdp | sasl.login.callback.handler.class = null policy-xacml-pdp | sasl.login.class = null policy-xacml-pdp | sasl.login.connect.timeout.ms = null policy-xacml-pdp | sasl.login.read.timeout.ms = null policy-xacml-pdp | sasl.login.refresh.buffer.seconds = 300 policy-xacml-pdp | sasl.login.refresh.min.period.seconds = 60 policy-xacml-pdp | sasl.login.refresh.window.factor = 0.8 policy-xacml-pdp | sasl.login.refresh.window.jitter = 0.05 policy-xacml-pdp | sasl.login.retry.backoff.max.ms = 10000 policy-xacml-pdp | sasl.login.retry.backoff.ms = 100 policy-xacml-pdp | sasl.mechanism = GSSAPI policy-xacml-pdp | sasl.oauthbearer.clock.skew.seconds = 30 policy-xacml-pdp | sasl.oauthbearer.expected.audience = null policy-xacml-pdp | sasl.oauthbearer.expected.issuer = null policy-xacml-pdp | sasl.oauthbearer.header.urlencode = false policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.url = null policy-xacml-pdp | sasl.oauthbearer.scope.claim.name = scope policy-xacml-pdp | sasl.oauthbearer.sub.claim.name = sub policy-xacml-pdp | sasl.oauthbearer.token.endpoint.url = null policy-xacml-pdp | security.protocol = PLAINTEXT policy-xacml-pdp | security.providers = null policy-xacml-pdp | send.buffer.bytes = 131072 policy-xacml-pdp | session.timeout.ms = 45000 policy-xacml-pdp | socket.connection.setup.timeout.max.ms = 30000 policy-xacml-pdp | socket.connection.setup.timeout.ms = 10000 policy-xacml-pdp | ssl.cipher.suites = null policy-xacml-pdp | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-xacml-pdp | ssl.endpoint.identification.algorithm = https policy-xacml-pdp | ssl.engine.factory.class = null policy-xacml-pdp | ssl.key.password = null policy-xacml-pdp | ssl.keymanager.algorithm = SunX509 policy-xacml-pdp | ssl.keystore.certificate.chain = null policy-xacml-pdp | ssl.keystore.key = null policy-xacml-pdp | ssl.keystore.location = null policy-xacml-pdp | ssl.keystore.password = null policy-xacml-pdp | ssl.keystore.type = JKS policy-xacml-pdp | ssl.protocol = TLSv1.3 policy-xacml-pdp | ssl.provider = null policy-xacml-pdp | ssl.secure.random.implementation = null policy-xacml-pdp | ssl.trustmanager.algorithm = PKIX policy-xacml-pdp | ssl.truststore.certificates = null policy-xacml-pdp | ssl.truststore.location = null policy-xacml-pdp | ssl.truststore.password = null policy-xacml-pdp | ssl.truststore.type = JKS policy-xacml-pdp | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-xacml-pdp | policy-xacml-pdp | [2025-06-18T15:21:05.385+00:00|INFO|KafkaMetricsCollector|main] initializing Kafka metrics collector policy-xacml-pdp | [2025-06-18T15:21:05.542+00:00|INFO|AppInfoParser|main] Kafka version: 3.9.1 policy-xacml-pdp | [2025-06-18T15:21:05.543+00:00|INFO|AppInfoParser|main] Kafka commitId: f745dfdcee2b9851 policy-xacml-pdp | [2025-06-18T15:21:05.543+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1750260065541 policy-xacml-pdp | [2025-06-18T15:21:05.545+00:00|INFO|ClassicKafkaConsumer|main] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-1, groupId=da487152-880a-42ad-96cd-a24e5d423167] Subscribed to topic(s): policy-pdp-pap policy-xacml-pdp | [2025-06-18T15:21:05.614+00:00|INFO|XacmlPdpApplicationManager|main] Initialization applications org.onap.policy.pdpx.main.parameters.XacmlApplicationParameters@7ec3394b JerseyClient(name=policyApiParameters, https=false, selfSignedCerts=false, hostname=policy-api, port=6969, basePath=null, userName=policyadmin, password=zb!XztG34, client=org.glassfish.jersey.client.JerseyClient@698122b2, baseUrl=http://policy-api:6969/, alive=true) policy-xacml-pdp | [2025-06-18T15:21:05.626+00:00|INFO|XacmlPdpApplicationManager|main] Application guard supports [onap.policies.controlloop.guard.common.FrequencyLimiter 1.0.0, onap.policies.controlloop.guard.common.MinMax 1.0.0, onap.policies.controlloop.guard.common.Blacklist 1.0.0, onap.policies.controlloop.guard.common.Filter 1.0.0, onap.policies.controlloop.guard.coordination.FirstBlocksSecond 1.0.0] policy-xacml-pdp | [2025-06-18T15:21:05.627+00:00|INFO|XacmlPdpApplicationManager|main] initializeApplicationPath guard at this path /opt/app/policy/pdpx/apps/guard policy-xacml-pdp | [2025-06-18T15:21:05.627+00:00|INFO|StdXacmlApplicationServiceProvider|main] New Path is /opt/app/policy/pdpx/apps/guard policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] Loading xacml properties /opt/app/policy/pdpx/apps/guard/xacml.properties policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] Loaded xacml properties policy-xacml-pdp | {count-recent-operations.persistenceunit=OperationsHistoryPU, get-operation-outcome.name=GetOperationOutcome, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:oasis:names:tc:xacml:3.0:policy-combining-algorithm:deny-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, count-recent-operations.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.CountRecentOperationsPip, get-operation-outcome.description=Returns operation outcome, count-recent-operations.description=Returns operation counts based on time window, jakarta.persistence.jdbc.password=policy_user, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, get-operation-outcome.issuer=urn:org:onap:xacml:guard:get-operation-outcome, get-operation-outcome.persistenceunit=OperationsHistoryPU, jakarta.persistence.jdbc.driver=org.postgresql.Driver, count-recent-operations.name=CountRecentOperations, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, jakarta.persistence.jdbc.url=jdbc:postgresql://postgres:5432/operationshistory, jakarta.persistence.jdbc.user=policy_user, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, count-recent-operations.issuer=urn:org:onap:xacml:guard:count-recent-operations, xacml.pip.engines=count-recent-operations,get-operation-outcome, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, get-operation-outcome.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.GetOperationOutcomePip} policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] count-recent-operations.persistenceunit -> OperationsHistoryPU policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] get-operation-outcome.name -> GetOperationOutcome policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] xacml.att.evaluationContextFactory -> com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] xacml.pepEngineFactory -> com.att.research.xacml.std.pep.StdEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] xacml.dataTypeFactory -> com.att.research.xacml.std.StdDataTypeFactory policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory.combineRootPolicies -> urn:oasis:names:tc:xacml:3.0:policy-combining-algorithm:deny-overrides policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory -> org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] count-recent-operations.classname -> org.onap.policy.pdp.xacml.application.common.operationshistory.CountRecentOperationsPip policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] get-operation-outcome.description -> Returns operation outcome policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] count-recent-operations.description -> Returns operation counts based on time window policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] jakarta.persistence.jdbc.password -> policy_user policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] xacml.att.functionDefinitionFactory -> com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] get-operation-outcome.issuer -> urn:org:onap:xacml:guard:get-operation-outcome policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] get-operation-outcome.persistenceunit -> OperationsHistoryPU policy-xacml-pdp | [2025-06-18T15:21:05.628+00:00|INFO|XacmlPolicyUtils|main] jakarta.persistence.jdbc.driver -> org.postgresql.Driver policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] count-recent-operations.name -> CountRecentOperations policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] xacml.att.combiningAlgorithmFactory -> com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] xacml.pdpEngineFactory -> com.att.research.xacmlatt.pdp.ATTPDPEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] jakarta.persistence.jdbc.url -> jdbc:postgresql://postgres:5432/operationshistory policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] jakarta.persistence.jdbc.user -> policy_user policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] xacml.traceEngineFactory -> com.att.research.xacml.std.trace.LoggingTraceEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] count-recent-operations.issuer -> urn:org:onap:xacml:guard:count-recent-operations policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] xacml.pip.engines -> count-recent-operations,get-operation-outcome policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] xacml.pipFinderFactory -> com.att.research.xacml.std.pip.StdPIPFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|XacmlPolicyUtils|main] get-operation-outcome.classname -> org.onap.policy.pdp.xacml.application.common.operationshistory.GetOperationOutcomePip policy-xacml-pdp | [2025-06-18T15:21:05.629+00:00|INFO|StdXacmlApplicationServiceProvider|main] {count-recent-operations.persistenceunit=OperationsHistoryPU, get-operation-outcome.name=GetOperationOutcome, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:oasis:names:tc:xacml:3.0:policy-combining-algorithm:deny-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, count-recent-operations.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.CountRecentOperationsPip, get-operation-outcome.description=Returns operation outcome, count-recent-operations.description=Returns operation counts based on time window, jakarta.persistence.jdbc.password=policy_user, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, get-operation-outcome.issuer=urn:org:onap:xacml:guard:get-operation-outcome, get-operation-outcome.persistenceunit=OperationsHistoryPU, jakarta.persistence.jdbc.driver=org.postgresql.Driver, count-recent-operations.name=CountRecentOperations, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, jakarta.persistence.jdbc.url=jdbc:postgresql://postgres:5432/operationshistory, jakarta.persistence.jdbc.user=policy_user, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, count-recent-operations.issuer=urn:org:onap:xacml:guard:count-recent-operations, xacml.pip.engines=count-recent-operations,get-operation-outcome, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, get-operation-outcome.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.GetOperationOutcomePip} policy-xacml-pdp | [2025-06-18T15:21:05.631+00:00|WARN|XACMLProperties|main] Properties file /usr/lib/jvm/java-17-openjdk/lib/xacml.properties cannot be read. policy-xacml-pdp | [2025-06-18T15:21:05.660+00:00|INFO|XacmlPdpApplicationManager|main] Application optimization supports [onap.policies.optimization.resource.AffinityPolicy 1.0.0, onap.policies.optimization.resource.DistancePolicy 1.0.0, onap.policies.optimization.resource.HpaPolicy 1.0.0, onap.policies.optimization.resource.OptimizationPolicy 1.0.0, onap.policies.optimization.resource.PciPolicy 1.0.0, onap.policies.optimization.service.QueryPolicy 1.0.0, onap.policies.optimization.service.SubscriberPolicy 1.0.0, onap.policies.optimization.resource.Vim_fit 1.0.0, onap.policies.optimization.resource.VnfPolicy 1.0.0] policy-xacml-pdp | [2025-06-18T15:21:05.661+00:00|INFO|XacmlPdpApplicationManager|main] initializeApplicationPath optimization at this path /opt/app/policy/pdpx/apps/optimization policy-xacml-pdp | [2025-06-18T15:21:05.661+00:00|INFO|StdXacmlApplicationServiceProvider|main] New Path is /opt/app/policy/pdpx/apps/optimization policy-xacml-pdp | [2025-06-18T15:21:05.661+00:00|INFO|XacmlPolicyUtils|main] Loading xacml properties /opt/app/policy/pdpx/apps/optimization/xacml.properties policy-xacml-pdp | [2025-06-18T15:21:05.661+00:00|INFO|XacmlPolicyUtils|main] Loaded xacml properties policy-xacml-pdp | {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.661+00:00|INFO|XacmlPolicyUtils|main] xacml.rootPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.att.evaluationContextFactory -> com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.att.combiningAlgorithmFactory -> com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.pepEngineFactory -> com.att.research.xacml.std.pep.StdEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.dataTypeFactory -> com.att.research.xacml.std.StdDataTypeFactory policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory.combineRootPolicies -> urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.referencedPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory -> org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.pdpEngineFactory -> com.att.research.xacmlatt.pdp.ATTPDPEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.traceEngineFactory -> com.att.research.xacml.std.trace.LoggingTraceEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.pipFinderFactory -> com.att.research.xacml.std.pip.StdPIPFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|XacmlPolicyUtils|main] xacml.att.functionDefinitionFactory -> com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory policy-xacml-pdp | [2025-06-18T15:21:05.662+00:00|INFO|StdXacmlApplicationServiceProvider|main] {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.664+00:00|INFO|XacmlPdpApplicationManager|main] Application naming supports [onap.policies.Naming 1.0.0] policy-xacml-pdp | [2025-06-18T15:21:05.664+00:00|INFO|XacmlPdpApplicationManager|main] initializeApplicationPath naming at this path /opt/app/policy/pdpx/apps/naming policy-xacml-pdp | [2025-06-18T15:21:05.664+00:00|INFO|StdXacmlApplicationServiceProvider|main] New Path is /opt/app/policy/pdpx/apps/naming policy-xacml-pdp | [2025-06-18T15:21:05.664+00:00|INFO|XacmlPolicyUtils|main] Loading xacml properties /opt/app/policy/pdpx/apps/naming/xacml.properties policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] Loaded xacml properties policy-xacml-pdp | {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.rootPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.att.evaluationContextFactory -> com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.att.combiningAlgorithmFactory -> com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.pepEngineFactory -> com.att.research.xacml.std.pep.StdEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.dataTypeFactory -> com.att.research.xacml.std.StdDataTypeFactory policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory.combineRootPolicies -> urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.referencedPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory -> org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.pdpEngineFactory -> com.att.research.xacmlatt.pdp.ATTPDPEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.traceEngineFactory -> com.att.research.xacml.std.trace.LoggingTraceEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.pipFinderFactory -> com.att.research.xacml.std.pip.StdPIPFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|XacmlPolicyUtils|main] xacml.att.functionDefinitionFactory -> com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory policy-xacml-pdp | [2025-06-18T15:21:05.665+00:00|INFO|StdXacmlApplicationServiceProvider|main] {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPdpApplicationManager|main] Application native supports [onap.policies.native.Xacml 1.0.0, onap.policies.native.ToscaXacml 1.0.0] policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPdpApplicationManager|main] initializeApplicationPath native at this path /opt/app/policy/pdpx/apps/native policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|StdXacmlApplicationServiceProvider|main] New Path is /opt/app/policy/pdpx/apps/native policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] Loading xacml properties /opt/app/policy/pdpx/apps/native/xacml.properties policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] Loaded xacml properties policy-xacml-pdp | {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.rootPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.att.evaluationContextFactory -> com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.att.combiningAlgorithmFactory -> com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.pepEngineFactory -> com.att.research.xacml.std.pep.StdEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.dataTypeFactory -> com.att.research.xacml.std.StdDataTypeFactory policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory.combineRootPolicies -> urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.referencedPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory -> org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.pdpEngineFactory -> com.att.research.xacmlatt.pdp.ATTPDPEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.traceEngineFactory -> com.att.research.xacml.std.trace.LoggingTraceEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.pipFinderFactory -> com.att.research.xacml.std.pip.StdPIPFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|XacmlPolicyUtils|main] xacml.att.functionDefinitionFactory -> com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory policy-xacml-pdp | [2025-06-18T15:21:05.668+00:00|INFO|StdXacmlApplicationServiceProvider|main] {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.669+00:00|INFO|XacmlPdpApplicationManager|main] Application match supports [onap.policies.Match 1.0.0] policy-xacml-pdp | [2025-06-18T15:21:05.669+00:00|INFO|XacmlPdpApplicationManager|main] initializeApplicationPath match at this path /opt/app/policy/pdpx/apps/match policy-xacml-pdp | [2025-06-18T15:21:05.669+00:00|INFO|StdXacmlApplicationServiceProvider|main] New Path is /opt/app/policy/pdpx/apps/match policy-xacml-pdp | [2025-06-18T15:21:05.669+00:00|INFO|XacmlPolicyUtils|main] Loading xacml properties /opt/app/policy/pdpx/apps/match/xacml.properties policy-xacml-pdp | [2025-06-18T15:21:05.669+00:00|INFO|XacmlPolicyUtils|main] Loaded xacml properties policy-xacml-pdp | {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.669+00:00|INFO|XacmlPolicyUtils|main] xacml.rootPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.669+00:00|INFO|XacmlPolicyUtils|main] xacml.att.evaluationContextFactory -> com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory policy-xacml-pdp | [2025-06-18T15:21:05.669+00:00|INFO|XacmlPolicyUtils|main] xacml.att.combiningAlgorithmFactory -> com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|XacmlPolicyUtils|main] xacml.pepEngineFactory -> com.att.research.xacml.std.pep.StdEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|XacmlPolicyUtils|main] xacml.dataTypeFactory -> com.att.research.xacml.std.StdDataTypeFactory policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory.combineRootPolicies -> urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|XacmlPolicyUtils|main] xacml.referencedPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory -> org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|XacmlPolicyUtils|main] xacml.pdpEngineFactory -> com.att.research.xacmlatt.pdp.ATTPDPEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|XacmlPolicyUtils|main] xacml.traceEngineFactory -> com.att.research.xacml.std.trace.LoggingTraceEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|XacmlPolicyUtils|main] xacml.pipFinderFactory -> com.att.research.xacml.std.pip.StdPIPFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|XacmlPolicyUtils|main] xacml.att.functionDefinitionFactory -> com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory policy-xacml-pdp | [2025-06-18T15:21:05.670+00:00|INFO|StdXacmlApplicationServiceProvider|main] {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPdpApplicationManager|main] Application monitoring supports [onap.Monitoring 1.0.0] policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPdpApplicationManager|main] initializeApplicationPath monitoring at this path /opt/app/policy/pdpx/apps/monitoring policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|StdXacmlApplicationServiceProvider|main] New Path is /opt/app/policy/pdpx/apps/monitoring policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] Loading xacml properties /opt/app/policy/pdpx/apps/monitoring/xacml.properties policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] Loaded xacml properties policy-xacml-pdp | {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.rootPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.att.evaluationContextFactory -> com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.att.combiningAlgorithmFactory -> com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.pepEngineFactory -> com.att.research.xacml.std.pep.StdEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.dataTypeFactory -> com.att.research.xacml.std.StdDataTypeFactory policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory.combineRootPolicies -> urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.referencedPolicies -> policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.att.policyFinderFactory -> org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.pdpEngineFactory -> com.att.research.xacmlatt.pdp.ATTPDPEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.traceEngineFactory -> com.att.research.xacml.std.trace.LoggingTraceEngineFactory policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.pipFinderFactory -> com.att.research.xacml.std.pip.StdPIPFinderFactory policy-xacml-pdp | [2025-06-18T15:21:05.671+00:00|INFO|XacmlPolicyUtils|main] xacml.att.functionDefinitionFactory -> com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory policy-xacml-pdp | [2025-06-18T15:21:05.672+00:00|INFO|StdXacmlApplicationServiceProvider|main] {xacml.rootPolicies=, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.referencedPolicies=, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory} policy-xacml-pdp | [2025-06-18T15:21:05.672+00:00|INFO|XacmlPdpApplicationManager|main] Finished applications initialization {optimize=org.onap.policy.xacml.pdp.application.optimization.OptimizationPdpApplication@2b95e48b, native=org.onap.policy.xacml.pdp.application.nativ.NativePdpApplication@4a3329b9, guard=org.onap.policy.xacml.pdp.application.guard.GuardPdpApplication@3dddefd8, naming=org.onap.policy.xacml.pdp.application.naming.NamingPdpApplication@160ac7fb, match=org.onap.policy.xacml.pdp.application.match.MatchPdpApplication@12bfd80d, configure=org.onap.policy.xacml.pdp.application.monitoring.MonitoringPdpApplication@41925502} policy-xacml-pdp | [2025-06-18T15:21:05.694+00:00|INFO|XacmlPdpHearbeatPublisher|main] heartbeat topic probe 4000ms policy-xacml-pdp | [2025-06-18T15:21:05.902+00:00|INFO|ServiceManager|main] service manager starting policy-xacml-pdp | [2025-06-18T15:21:05.902+00:00|INFO|ServiceManager|main] service manager starting XACML PDP parameters policy-xacml-pdp | [2025-06-18T15:21:05.902+00:00|INFO|ServiceManager|main] service manager starting Message Dispatcher policy-xacml-pdp | [2025-06-18T15:21:05.902+00:00|INFO|TopicBase|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=da487152-880a-42ad-96cd-a24e5d423167, consumerInstance=policy-xacml-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: registering org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher@5f574cc2 policy-xacml-pdp | [2025-06-18T15:21:05.916+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=da487152-880a-42ad-96cd-a24e5d423167, consumerInstance=policy-xacml-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-xacml-pdp | [2025-06-18T15:21:05.916+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-xacml-pdp | allow.auto.create.topics = true policy-xacml-pdp | auto.commit.interval.ms = 5000 policy-xacml-pdp | auto.include.jmx.reporter = true policy-xacml-pdp | auto.offset.reset = latest policy-xacml-pdp | bootstrap.servers = [kafka:9092] policy-xacml-pdp | check.crcs = true policy-xacml-pdp | client.dns.lookup = use_all_dns_ips policy-xacml-pdp | client.id = consumer-da487152-880a-42ad-96cd-a24e5d423167-2 policy-xacml-pdp | client.rack = policy-xacml-pdp | connections.max.idle.ms = 540000 policy-xacml-pdp | default.api.timeout.ms = 60000 policy-xacml-pdp | enable.auto.commit = true policy-xacml-pdp | enable.metrics.push = true policy-xacml-pdp | exclude.internal.topics = true policy-xacml-pdp | fetch.max.bytes = 52428800 policy-xacml-pdp | fetch.max.wait.ms = 500 policy-xacml-pdp | fetch.min.bytes = 1 policy-xacml-pdp | group.id = da487152-880a-42ad-96cd-a24e5d423167 policy-xacml-pdp | group.instance.id = null policy-xacml-pdp | group.protocol = classic policy-xacml-pdp | group.remote.assignor = null policy-xacml-pdp | heartbeat.interval.ms = 3000 policy-xacml-pdp | interceptor.classes = [] policy-xacml-pdp | internal.leave.group.on.close = true policy-xacml-pdp | internal.throw.on.fetch.stable.offset.unsupported = false policy-xacml-pdp | isolation.level = read_uncommitted policy-xacml-pdp | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-xacml-pdp | max.partition.fetch.bytes = 1048576 policy-xacml-pdp | max.poll.interval.ms = 300000 policy-xacml-pdp | max.poll.records = 500 policy-xacml-pdp | metadata.max.age.ms = 300000 policy-xacml-pdp | metadata.recovery.strategy = none policy-xacml-pdp | metric.reporters = [] policy-xacml-pdp | metrics.num.samples = 2 policy-xacml-pdp | metrics.recording.level = INFO policy-xacml-pdp | metrics.sample.window.ms = 30000 policy-xacml-pdp | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-xacml-pdp | receive.buffer.bytes = 65536 policy-xacml-pdp | reconnect.backoff.max.ms = 1000 policy-xacml-pdp | reconnect.backoff.ms = 50 policy-xacml-pdp | request.timeout.ms = 30000 policy-xacml-pdp | retry.backoff.max.ms = 1000 policy-xacml-pdp | retry.backoff.ms = 100 policy-xacml-pdp | sasl.client.callback.handler.class = null policy-xacml-pdp | sasl.jaas.config = null policy-xacml-pdp | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-xacml-pdp | sasl.kerberos.min.time.before.relogin = 60000 policy-xacml-pdp | sasl.kerberos.service.name = null policy-xacml-pdp | sasl.kerberos.ticket.renew.jitter = 0.05 policy-xacml-pdp | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-xacml-pdp | sasl.login.callback.handler.class = null policy-xacml-pdp | sasl.login.class = null policy-xacml-pdp | sasl.login.connect.timeout.ms = null policy-xacml-pdp | sasl.login.read.timeout.ms = null policy-xacml-pdp | sasl.login.refresh.buffer.seconds = 300 policy-xacml-pdp | sasl.login.refresh.min.period.seconds = 60 policy-xacml-pdp | sasl.login.refresh.window.factor = 0.8 policy-xacml-pdp | sasl.login.refresh.window.jitter = 0.05 policy-xacml-pdp | sasl.login.retry.backoff.max.ms = 10000 policy-xacml-pdp | sasl.login.retry.backoff.ms = 100 policy-xacml-pdp | sasl.mechanism = GSSAPI policy-xacml-pdp | sasl.oauthbearer.clock.skew.seconds = 30 policy-xacml-pdp | sasl.oauthbearer.expected.audience = null policy-xacml-pdp | sasl.oauthbearer.expected.issuer = null policy-xacml-pdp | sasl.oauthbearer.header.urlencode = false policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.url = null policy-xacml-pdp | sasl.oauthbearer.scope.claim.name = scope policy-xacml-pdp | sasl.oauthbearer.sub.claim.name = sub policy-xacml-pdp | sasl.oauthbearer.token.endpoint.url = null policy-xacml-pdp | security.protocol = PLAINTEXT policy-xacml-pdp | security.providers = null policy-xacml-pdp | send.buffer.bytes = 131072 policy-xacml-pdp | session.timeout.ms = 45000 policy-xacml-pdp | socket.connection.setup.timeout.max.ms = 30000 policy-xacml-pdp | socket.connection.setup.timeout.ms = 10000 policy-xacml-pdp | ssl.cipher.suites = null policy-xacml-pdp | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-xacml-pdp | ssl.endpoint.identification.algorithm = https policy-xacml-pdp | ssl.engine.factory.class = null policy-xacml-pdp | ssl.key.password = null policy-xacml-pdp | ssl.keymanager.algorithm = SunX509 policy-xacml-pdp | ssl.keystore.certificate.chain = null policy-xacml-pdp | ssl.keystore.key = null policy-xacml-pdp | ssl.keystore.location = null policy-xacml-pdp | ssl.keystore.password = null policy-xacml-pdp | ssl.keystore.type = JKS policy-xacml-pdp | ssl.protocol = TLSv1.3 policy-xacml-pdp | ssl.provider = null policy-xacml-pdp | ssl.secure.random.implementation = null policy-xacml-pdp | ssl.trustmanager.algorithm = PKIX policy-xacml-pdp | ssl.truststore.certificates = null policy-xacml-pdp | ssl.truststore.location = null policy-xacml-pdp | ssl.truststore.password = null policy-xacml-pdp | ssl.truststore.type = JKS policy-xacml-pdp | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-xacml-pdp | policy-xacml-pdp | [2025-06-18T15:21:05.917+00:00|INFO|KafkaMetricsCollector|main] initializing Kafka metrics collector policy-xacml-pdp | [2025-06-18T15:21:05.929+00:00|INFO|AppInfoParser|main] Kafka version: 3.9.1 policy-xacml-pdp | [2025-06-18T15:21:05.929+00:00|INFO|AppInfoParser|main] Kafka commitId: f745dfdcee2b9851 policy-xacml-pdp | [2025-06-18T15:21:05.929+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1750260065929 policy-xacml-pdp | [2025-06-18T15:21:05.930+00:00|INFO|ClassicKafkaConsumer|main] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Subscribed to topic(s): policy-pdp-pap policy-xacml-pdp | [2025-06-18T15:21:05.930+00:00|INFO|ServiceManager|main] service manager starting topics policy-xacml-pdp | [2025-06-18T15:21:05.930+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=da487152-880a-42ad-96cd-a24e5d423167, consumerInstance=policy-xacml-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-xacml-pdp | [2025-06-18T15:21:05.930+00:00|INFO|InlineBusTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=8a60a121-5003-4e4e-8cbb-8355ecb743e1, alive=false, publisher=null]]: starting policy-xacml-pdp | [2025-06-18T15:21:05.940+00:00|INFO|ProducerConfig|main] ProducerConfig values: policy-xacml-pdp | acks = -1 policy-xacml-pdp | auto.include.jmx.reporter = true policy-xacml-pdp | batch.size = 16384 policy-xacml-pdp | bootstrap.servers = [kafka:9092] policy-xacml-pdp | buffer.memory = 33554432 policy-xacml-pdp | client.dns.lookup = use_all_dns_ips policy-xacml-pdp | client.id = producer-1 policy-xacml-pdp | compression.gzip.level = -1 policy-xacml-pdp | compression.lz4.level = 9 policy-xacml-pdp | compression.type = none policy-xacml-pdp | compression.zstd.level = 3 policy-xacml-pdp | connections.max.idle.ms = 540000 policy-xacml-pdp | delivery.timeout.ms = 120000 policy-xacml-pdp | enable.idempotence = true policy-xacml-pdp | enable.metrics.push = true policy-xacml-pdp | interceptor.classes = [] policy-xacml-pdp | key.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-xacml-pdp | linger.ms = 0 policy-xacml-pdp | max.block.ms = 60000 policy-xacml-pdp | max.in.flight.requests.per.connection = 5 policy-xacml-pdp | max.request.size = 1048576 policy-xacml-pdp | metadata.max.age.ms = 300000 policy-xacml-pdp | metadata.max.idle.ms = 300000 policy-xacml-pdp | metadata.recovery.strategy = none policy-xacml-pdp | metric.reporters = [] policy-xacml-pdp | metrics.num.samples = 2 policy-xacml-pdp | metrics.recording.level = INFO policy-xacml-pdp | metrics.sample.window.ms = 30000 policy-xacml-pdp | partitioner.adaptive.partitioning.enable = true policy-xacml-pdp | partitioner.availability.timeout.ms = 0 policy-xacml-pdp | partitioner.class = null policy-xacml-pdp | partitioner.ignore.keys = false policy-xacml-pdp | receive.buffer.bytes = 32768 policy-xacml-pdp | reconnect.backoff.max.ms = 1000 policy-xacml-pdp | reconnect.backoff.ms = 50 policy-xacml-pdp | request.timeout.ms = 30000 policy-xacml-pdp | retries = 2147483647 policy-xacml-pdp | retry.backoff.max.ms = 1000 policy-xacml-pdp | retry.backoff.ms = 100 policy-xacml-pdp | sasl.client.callback.handler.class = null policy-xacml-pdp | sasl.jaas.config = null policy-xacml-pdp | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-xacml-pdp | sasl.kerberos.min.time.before.relogin = 60000 policy-xacml-pdp | sasl.kerberos.service.name = null policy-xacml-pdp | sasl.kerberos.ticket.renew.jitter = 0.05 policy-xacml-pdp | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-xacml-pdp | sasl.login.callback.handler.class = null policy-xacml-pdp | sasl.login.class = null policy-xacml-pdp | sasl.login.connect.timeout.ms = null policy-xacml-pdp | sasl.login.read.timeout.ms = null policy-xacml-pdp | sasl.login.refresh.buffer.seconds = 300 policy-xacml-pdp | sasl.login.refresh.min.period.seconds = 60 policy-xacml-pdp | sasl.login.refresh.window.factor = 0.8 policy-xacml-pdp | sasl.login.refresh.window.jitter = 0.05 policy-xacml-pdp | sasl.login.retry.backoff.max.ms = 10000 policy-xacml-pdp | sasl.login.retry.backoff.ms = 100 policy-xacml-pdp | sasl.mechanism = GSSAPI policy-xacml-pdp | sasl.oauthbearer.clock.skew.seconds = 30 policy-xacml-pdp | sasl.oauthbearer.expected.audience = null policy-xacml-pdp | sasl.oauthbearer.expected.issuer = null policy-xacml-pdp | sasl.oauthbearer.header.urlencode = false policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-xacml-pdp | sasl.oauthbearer.jwks.endpoint.url = null policy-xacml-pdp | sasl.oauthbearer.scope.claim.name = scope policy-xacml-pdp | sasl.oauthbearer.sub.claim.name = sub policy-xacml-pdp | sasl.oauthbearer.token.endpoint.url = null policy-xacml-pdp | security.protocol = PLAINTEXT policy-xacml-pdp | security.providers = null policy-xacml-pdp | send.buffer.bytes = 131072 policy-xacml-pdp | socket.connection.setup.timeout.max.ms = 30000 policy-xacml-pdp | socket.connection.setup.timeout.ms = 10000 policy-xacml-pdp | ssl.cipher.suites = null policy-xacml-pdp | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-xacml-pdp | ssl.endpoint.identification.algorithm = https policy-xacml-pdp | ssl.engine.factory.class = null policy-xacml-pdp | ssl.key.password = null policy-xacml-pdp | ssl.keymanager.algorithm = SunX509 policy-xacml-pdp | ssl.keystore.certificate.chain = null policy-xacml-pdp | ssl.keystore.key = null policy-xacml-pdp | ssl.keystore.location = null policy-xacml-pdp | ssl.keystore.password = null policy-xacml-pdp | ssl.keystore.type = JKS policy-xacml-pdp | ssl.protocol = TLSv1.3 policy-xacml-pdp | ssl.provider = null policy-xacml-pdp | ssl.secure.random.implementation = null policy-xacml-pdp | ssl.trustmanager.algorithm = PKIX policy-xacml-pdp | ssl.truststore.certificates = null policy-xacml-pdp | ssl.truststore.location = null policy-xacml-pdp | ssl.truststore.password = null policy-xacml-pdp | ssl.truststore.type = JKS policy-xacml-pdp | transaction.timeout.ms = 60000 policy-xacml-pdp | transactional.id = null policy-xacml-pdp | value.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-xacml-pdp | policy-xacml-pdp | [2025-06-18T15:21:05.942+00:00|INFO|KafkaMetricsCollector|main] initializing Kafka metrics collector policy-xacml-pdp | [2025-06-18T15:21:05.951+00:00|INFO|KafkaProducer|main] [Producer clientId=producer-1] Instantiated an idempotent producer. policy-xacml-pdp | [2025-06-18T15:21:05.970+00:00|INFO|AppInfoParser|main] Kafka version: 3.9.1 policy-xacml-pdp | [2025-06-18T15:21:05.970+00:00|INFO|AppInfoParser|main] Kafka commitId: f745dfdcee2b9851 policy-xacml-pdp | [2025-06-18T15:21:05.970+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1750260065970 policy-xacml-pdp | [2025-06-18T15:21:05.971+00:00|INFO|InlineKafkaTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=8a60a121-5003-4e4e-8cbb-8355ecb743e1, alive=false, publisher=KafkaPublisherWrapper []]]: KAFKA SINK created policy-xacml-pdp | [2025-06-18T15:21:05.971+00:00|INFO|ServiceManager|main] service manager starting Terminate PDP policy-xacml-pdp | [2025-06-18T15:21:05.971+00:00|INFO|ServiceManager|main] service manager starting Heartbeat Publisher policy-xacml-pdp | [2025-06-18T15:21:05.972+00:00|INFO|ServiceManager|main] service manager starting REST Server policy-xacml-pdp | [2025-06-18T15:21:05.972+00:00|INFO|ServiceManager|main] service manager starting policy-xacml-pdp | [2025-06-18T15:21:05.972+00:00|INFO|ServiceManager|main] service manager starting REST RestServerParameters policy-xacml-pdp | [2025-06-18T15:21:05.972+00:00|INFO|JettyServletServer|main] JettyJerseyServer [JerseyServlets={/metrics=io.prometheus.metrics.exporter.servlet.jakarta.PrometheusMetricsServlet-6e9c413e==io.prometheus.metrics.exporter.servlet.jakarta.PrometheusMetricsServlet@b94e35e{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-a23a01d==org.glassfish.jersey.servlet.ServletContainer@d5e4ed96{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:,STOPPED}}, swaggerId=swagger-6969, toString()=JettyServletServer(name=RestServerParameters, host=0.0.0.0, port=6969, sniHostCheck=false, user=policyadmin, password=zb!XztG34, contextPath=/, jettyServer=oejs.Server@38b972d7{STOPPED}[12.0.21,sto=0], context=oeje10s.ServletContextHandler@452c8a40{ROOT,/,b=null,a=STOPPED,h=oeje10s.SessionHandler@534243e4{STOPPED}}, connector=RestServerParameters@29006752{HTTP/1.1, (http/1.1)}{0.0.0.0:6969}, jettyThread=null, servlets={/metrics=io.prometheus.metrics.exporter.servlet.jakarta.PrometheusMetricsServlet-6e9c413e==io.prometheus.metrics.exporter.servlet.jakarta.PrometheusMetricsServlet@b94e35e{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-a23a01d==org.glassfish.jersey.servlet.ServletContainer@d5e4ed96{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:,STOPPED}})]: STARTING policy-xacml-pdp | [2025-06-18T15:21:05.985+00:00|INFO|ServiceManager|main] service manager started policy-xacml-pdp | [2025-06-18T15:21:05.985+00:00|INFO|ServiceManager|main] service manager started policy-xacml-pdp | [2025-06-18T15:21:05.985+00:00|INFO|Main|main] Started policy-xacml-pdp service successfully. policy-xacml-pdp | [2025-06-18T15:21:05.985+00:00|INFO|TopicBase|pool-2-thread-1] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=da487152-880a-42ad-96cd-a24e5d423167, consumerInstance=policy-xacml-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: registering org.onap.policy.common.message.bus.event.client.BidirectionalTopicClient$$Lambda$504/0x00007f44ec2adc68@7999ae72 policy-xacml-pdp | [2025-06-18T15:21:05.986+00:00|INFO|JettyServletServer|RestServerParameters-6969] JettyJerseyServer [JerseyServlets={/metrics=io.prometheus.metrics.exporter.servlet.jakarta.PrometheusMetricsServlet-6e9c413e==io.prometheus.metrics.exporter.servlet.jakarta.PrometheusMetricsServlet@b94e35e{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-a23a01d==org.glassfish.jersey.servlet.ServletContainer@d5e4ed96{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:,STOPPED}}, swaggerId=swagger-6969, toString()=JettyServletServer(name=RestServerParameters, host=0.0.0.0, port=6969, sniHostCheck=false, user=policyadmin, password=zb!XztG34, contextPath=/, jettyServer=oejs.Server@38b972d7{STOPPED}[12.0.21,sto=0], context=oeje10s.ServletContextHandler@452c8a40{ROOT,/,b=null,a=STOPPED,h=oeje10s.SessionHandler@534243e4{STOPPED}}, connector=RestServerParameters@29006752{HTTP/1.1, (http/1.1)}{0.0.0.0:6969}, jettyThread=Thread[RestServerParameters-6969,5,main], servlets={/metrics=io.prometheus.metrics.exporter.servlet.jakarta.PrometheusMetricsServlet-6e9c413e==io.prometheus.metrics.exporter.servlet.jakarta.PrometheusMetricsServlet@b94e35e{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-a23a01d==org.glassfish.jersey.servlet.ServletContainer@d5e4ed96{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:,STOPPED}})]: RUN policy-xacml-pdp | [2025-06-18T15:21:05.986+00:00|INFO|SingleThreadedBusTopicSource|pool-2-thread-1] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=da487152-880a-42ad-96cd-a24e5d423167, consumerInstance=policy-xacml-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=2, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=2]]]]: register: start not attempted policy-xacml-pdp | [2025-06-18T15:21:05.993+00:00|INFO|OrderedServiceImpl|pool-2-thread-1] ***** OrderedServiceImpl implementers: policy-xacml-pdp | [] policy-xacml-pdp | [2025-06-18T15:21:05.997+00:00|INFO|network|pool-2-thread-1] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"messageName":"PDP_TOPIC_CHECK","requestId":"1a80f344-84db-4fe8-bc72-5ce0a89e0807","timestampMs":1750260065978,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c"} policy-xacml-pdp | [2025-06-18T15:21:06.379+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] The metadata response from the cluster reported a recoverable issue with correlation id 3 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-xacml-pdp | [2025-06-18T15:21:06.379+00:00|WARN|NetworkClient|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] The metadata response from the cluster reported a recoverable issue with correlation id 2 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-xacml-pdp | [2025-06-18T15:21:06.380+00:00|INFO|Metadata|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Cluster ID: 7hUyEsUsSwOpxuI0HNr3Gw policy-xacml-pdp | [2025-06-18T15:21:06.380+00:00|INFO|Metadata|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] Cluster ID: 7hUyEsUsSwOpxuI0HNr3Gw policy-xacml-pdp | [2025-06-18T15:21:06.381+00:00|INFO|TransactionManager|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] ProducerId set to 2 with epoch 0 policy-xacml-pdp | [2025-06-18T15:21:06.489+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] The metadata response from the cluster reported a recoverable issue with correlation id 7 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-xacml-pdp | [2025-06-18T15:21:06.518+00:00|WARN|NetworkClient|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] The metadata response from the cluster reported a recoverable issue with correlation id 7 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-xacml-pdp | [2025-06-18T15:21:06.601+00:00|INFO|GsonMessageBodyHandler|RestServerParameters-6969] Using GSON for REST calls policy-xacml-pdp | [2025-06-18T15:21:06.601+00:00|INFO|YamlMessageBodyHandler|RestServerParameters-6969] Accepting YAML for REST calls policy-xacml-pdp | [2025-06-18T15:21:06.739+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Discovered group coordinator kafka:9092 (id: 2147483646 rack: null) policy-xacml-pdp | [2025-06-18T15:21:06.744+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] (Re-)joining group policy-xacml-pdp | [2025-06-18T15:21:06.785+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Request joining group due to: need to re-join with the given member-id: consumer-da487152-880a-42ad-96cd-a24e5d423167-2-ef05e808-ed55-4ebb-8f5e-20cd76b42a50 policy-xacml-pdp | [2025-06-18T15:21:06.786+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] (Re-)joining group policy-xacml-pdp | [2025-06-18T15:21:09.809+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Successfully joined group with generation Generation{generationId=1, memberId='consumer-da487152-880a-42ad-96cd-a24e5d423167-2-ef05e808-ed55-4ebb-8f5e-20cd76b42a50', protocol='range'} policy-xacml-pdp | [2025-06-18T15:21:09.818+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Finished assignment for group at generation 1: {consumer-da487152-880a-42ad-96cd-a24e5d423167-2-ef05e808-ed55-4ebb-8f5e-20cd76b42a50=Assignment(partitions=[policy-pdp-pap-0])} policy-xacml-pdp | [2025-06-18T15:21:09.839+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Successfully synced group in generation Generation{generationId=1, memberId='consumer-da487152-880a-42ad-96cd-a24e5d423167-2-ef05e808-ed55-4ebb-8f5e-20cd76b42a50', protocol='range'} policy-xacml-pdp | [2025-06-18T15:21:09.840+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Notifying assignor about the new Assignment(partitions=[policy-pdp-pap-0]) policy-xacml-pdp | [2025-06-18T15:21:09.843+00:00|INFO|ConsumerRebalanceListenerInvoker|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Adding newly assigned partitions: policy-pdp-pap-0 policy-xacml-pdp | [2025-06-18T15:21:09.860+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Found no committed offset for partition policy-pdp-pap-0 policy-xacml-pdp | [2025-06-18T15:21:09.874+00:00|INFO|SubscriptionState|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-da487152-880a-42ad-96cd-a24e5d423167-2, groupId=da487152-880a-42ad-96cd-a24e5d423167] Resetting offset for partition policy-pdp-pap-0 to position FetchPosition{offset=1, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:9092 (id: 1 rack: null)], epoch=0}}. policy-xacml-pdp | [2025-06-18T15:21:10.852+00:00|INFO|network|pool-2-thread-1] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"messageName":"PDP_TOPIC_CHECK","requestId":"1a80f344-84db-4fe8-bc72-5ce0a89e0807","timestampMs":1750260065978,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c"} policy-xacml-pdp | [2025-06-18T15:21:10.885+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"messageName":"PDP_TOPIC_CHECK","requestId":"1a80f344-84db-4fe8-bc72-5ce0a89e0807","timestampMs":1750260065978,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c"} policy-xacml-pdp | [2025-06-18T15:21:10.887+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_TOPIC_CHECK policy-xacml-pdp | [2025-06-18T15:21:10.888+00:00|INFO|BidirectionalTopicClient|KAFKA-source-policy-pdp-pap] topic policy-pdp-pap is ready; found matching message PdpTopicCheck(super=PdpMessage(messageName=PDP_TOPIC_CHECK, requestId=1a80f344-84db-4fe8-bc72-5ce0a89e0807, timestampMs=1750260065978, name=xacml-3e8650f6-1f12-4eed-9943-19c94576807c, pdpGroup=null, pdpSubgroup=null)) policy-xacml-pdp | [2025-06-18T15:21:10.892+00:00|INFO|TopicBase|pool-2-thread-1] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=da487152-880a-42ad-96cd-a24e5d423167, consumerInstance=policy-xacml-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=2, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=1, locked=false, #topicListeners=2]]]]: unregistering org.onap.policy.common.message.bus.event.client.BidirectionalTopicClient$$Lambda$504/0x00007f44ec2adc68@7999ae72 policy-xacml-pdp | [2025-06-18T15:21:10.893+00:00|INFO|XacmlPdpHearbeatPublisher|pool-2-thread-1] Sending Xacml PDP heartbeat to the PAP - PdpStatus(super=PdpMessage(messageName=PDP_STATUS, requestId=0f2cc3be-8452-4160-b2fe-aad52c1534ca, timestampMs=1750260070892, name=xacml-3e8650f6-1f12-4eed-9943-19c94576807c, pdpGroup=defaultGroup, pdpSubgroup=null), pdpType=xacml, state=PASSIVE, healthy=HEALTHY, description=null, policies=[], deploymentInstanceInfo=null, properties=null, response=null) policy-xacml-pdp | [2025-06-18T15:21:10.899+00:00|INFO|network|pool-2-thread-1] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[],"messageName":"PDP_STATUS","requestId":"0f2cc3be-8452-4160-b2fe-aad52c1534ca","timestampMs":1750260070892,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup"} policy-xacml-pdp | [2025-06-18T15:21:10.911+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[],"messageName":"PDP_STATUS","requestId":"0f2cc3be-8452-4160-b2fe-aad52c1534ca","timestampMs":1750260070892,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup"} policy-xacml-pdp | [2025-06-18T15:21:10.911+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-xacml-pdp | [2025-06-18T15:21:11.705+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[{"type":"onap.policies.Naming","type_version":"1.0.0","properties":{"policy-instance-name":"ONAP_NF_NAMING_TIMESTAMP","naming-models":[{"naming-type":"VNF","naming-recipe":"AIC_CLOUD_REGION|DELIMITER|CONSTANT|DELIMITER|TIMESTAMP","name-operation":"to_lower_case()","naming-properties":[{"property-name":"AIC_CLOUD_REGION"},{"property-name":"CONSTANT","property-value":"onap-nf"},{"property-name":"TIMESTAMP"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VNFC","naming-recipe":"VNF_NAME|DELIMITER|NFC_NAMING_CODE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"ENTIRETY","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}},{"property-name":"NFC_NAMING_CODE"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VF-MODULE","naming-recipe":"VNF_NAME|DELIMITER|VF_MODULE_LABEL|DELIMITER|VF_MODULE_TYPE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-value":"-","property-name":"DELIMITER"},{"property-name":"VF_MODULE_LABEL"},{"property-name":"VF_MODULE_TYPE"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"PRECEEDING","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}}]}]},"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0","metadata":{"policy-id":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","policy-version":"1.0.0"}}],"messageName":"PDP_UPDATE","requestId":"955bae4c-4069-40dc-add3-851fff567958","timestampMs":1750260071595,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:11.713+00:00|INFO|XacmlPdpUpdateListener|KAFKA-source-policy-pdp-pap] PDP update message has been received from the PAP - PdpUpdate(super=PdpMessage(messageName=PDP_UPDATE, requestId=955bae4c-4069-40dc-add3-851fff567958, timestampMs=1750260071595, name=xacml-3e8650f6-1f12-4eed-9943-19c94576807c, pdpGroup=defaultGroup, pdpSubgroup=xacml), source=pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac, description=null, pdpHeartbeatIntervalMs=120000, policiesToBeDeployed=[ToscaPolicy(super=ToscaWithTypeAndObjectProperties(type=onap.policies.Naming, typeVersion=1.0.0, properties={policy-instance-name=ONAP_NF_NAMING_TIMESTAMP, naming-models=[{naming-type=VNF, naming-recipe=AIC_CLOUD_REGION|DELIMITER|CONSTANT|DELIMITER|TIMESTAMP, name-operation=to_lower_case(), naming-properties=[{property-name=AIC_CLOUD_REGION}, {property-name=CONSTANT, property-value=onap-nf}, {property-name=TIMESTAMP}, {property-value=-, property-name=DELIMITER}]}, {naming-type=VNFC, naming-recipe=VNF_NAME|DELIMITER|NFC_NAMING_CODE|DELIMITER|SEQUENCE, name-operation=to_lower_case(), naming-properties=[{property-name=VNF_NAME}, {property-name=SEQUENCE, increment-sequence={max=zzz, scope=ENTIRETY, start-value=1, length=3, increment=1, sequence-type=alpha-numeric}}, {property-name=NFC_NAMING_CODE}, {property-value=-, property-name=DELIMITER}]}, {naming-type=VF-MODULE, naming-recipe=VNF_NAME|DELIMITER|VF_MODULE_LABEL|DELIMITER|VF_MODULE_TYPE|DELIMITER|SEQUENCE, name-operation=to_lower_case(), naming-properties=[{property-name=VNF_NAME}, {property-value=-, property-name=DELIMITER}, {property-name=VF_MODULE_LABEL}, {property-name=VF_MODULE_TYPE}, {property-name=SEQUENCE, increment-sequence={max=zzz, scope=PRECEEDING, start-value=1, length=3, increment=1, sequence-type=alpha-numeric}}]}]}))], policiesToBeUndeployed=[]) policy-xacml-pdp | [2025-06-18T15:21:11.721+00:00|INFO|StdBaseTranslator|KAFKA-source-policy-pdp-pap] Obligation Policy id: SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP type: onap.policies.Naming weight: null policy: policy-xacml-pdp | {"type":"onap.policies.Naming","type_version":"1.0.0","properties":{"policy-instance-name":"ONAP_NF_NAMING_TIMESTAMP","naming-models":[{"naming-type":"VNF","naming-recipe":"AIC_CLOUD_REGION|DELIMITER|CONSTANT|DELIMITER|TIMESTAMP","name-operation":"to_lower_case()","naming-properties":[{"property-name":"AIC_CLOUD_REGION"},{"property-name":"CONSTANT","property-value":"onap-nf"},{"property-name":"TIMESTAMP"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VNFC","naming-recipe":"VNF_NAME|DELIMITER|NFC_NAMING_CODE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"ENTIRETY","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}},{"property-name":"NFC_NAMING_CODE"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VF-MODULE","naming-recipe":"VNF_NAME|DELIMITER|VF_MODULE_LABEL|DELIMITER|VF_MODULE_TYPE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-value":"-","property-name":"DELIMITER"},{"property-name":"VF_MODULE_LABEL"},{"property-name":"VF_MODULE_TYPE"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"PRECEEDING","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}}]}]},"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0","metadata":{"policy-id":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","policy-version":"1.0.0"}} policy-xacml-pdp | [2025-06-18T15:21:11.776+00:00|INFO|StdXacmlApplicationServiceProvider|KAFKA-source-policy-pdp-pap] Xacml Policy is policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.policies.Naming policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.policies.Naming policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | 1.0.0 policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | Default is to PERMIT if the policy matches. policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | {"type":"onap.policies.Naming","type_version":"1.0.0","properties":{"policy-instance-name":"ONAP_NF_NAMING_TIMESTAMP","naming-models":[{"naming-type":"VNF","naming-recipe":"AIC_CLOUD_REGION|DELIMITER|CONSTANT|DELIMITER|TIMESTAMP","name-operation":"to_lower_case()","naming-properties":[{"property-name":"AIC_CLOUD_REGION"},{"property-name":"CONSTANT","property-value":"onap-nf"},{"property-name":"TIMESTAMP"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VNFC","naming-recipe":"VNF_NAME|DELIMITER|NFC_NAMING_CODE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"ENTIRETY","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}},{"property-name":"NFC_NAMING_CODE"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VF-MODULE","naming-recipe":"VNF_NAME|DELIMITER|VF_MODULE_LABEL|DELIMITER|VF_MODULE_TYPE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-value":"-","property-name":"DELIMITER"},{"property-name":"VF_MODULE_LABEL"},{"property-name":"VF_MODULE_TYPE"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"PRECEEDING","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}}]}]},"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0","metadata":{"policy-id":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","policy-version":"1.0.0"}} policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.policies.Naming policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | [2025-06-18T15:21:11.782+00:00|INFO|XacmlPolicyUtils|KAFKA-source-policy-pdp-pap] Storing xacml properties {xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, root1.file=/opt/app/policy/pdpx/apps/naming/SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP_1.0.0.xml, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, xacml.rootPolicies=root1, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.referencedPolicies=, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory} policy-xacml-pdp | /opt/app/policy/pdpx/apps/naming/xacml.properties policy-xacml-pdp | [2025-06-18T15:21:11.789+00:00|INFO|XacmlPdpApplicationManager|KAFKA-source-policy-pdp-pap] Loaded ToscaPolicy {policy-id=SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP, policy-version=1.0.0} into application naming policy-xacml-pdp | [2025-06-18T15:21:11.790+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"response":{"responseTo":"955bae4c-4069-40dc-add3-851fff567958","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"ee08dede-b370-432e-ad86-33ea5b385d5f","timestampMs":1750260071789,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:11.798+00:00|INFO|XacmlPdpHearbeatPublisher|pool-2-thread-1] Sending Xacml PDP heartbeat to the PAP - PdpStatus(super=PdpMessage(messageName=PDP_STATUS, requestId=245482cd-f55f-4400-8775-4300ae6de2c5, timestampMs=1750260071798, name=xacml-3e8650f6-1f12-4eed-9943-19c94576807c, pdpGroup=defaultGroup, pdpSubgroup=xacml), pdpType=xacml, state=PASSIVE, healthy=HEALTHY, description=null, policies=[SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP 1.0.0], deploymentInstanceInfo=null, properties=null, response=null) policy-xacml-pdp | [2025-06-18T15:21:11.799+00:00|INFO|network|pool-2-thread-1] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"messageName":"PDP_STATUS","requestId":"245482cd-f55f-4400-8775-4300ae6de2c5","timestampMs":1750260071798,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:11.804+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"response":{"responseTo":"955bae4c-4069-40dc-add3-851fff567958","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"ee08dede-b370-432e-ad86-33ea5b385d5f","timestampMs":1750260071789,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:11.804+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-xacml-pdp | [2025-06-18T15:21:11.854+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"PASSIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"messageName":"PDP_STATUS","requestId":"245482cd-f55f-4400-8775-4300ae6de2c5","timestampMs":1750260071798,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:11.854+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-xacml-pdp | [2025-06-18T15:21:11.862+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"a42fb34f-6c41-47b7-b8c3-4403ad562bf0","timestampMs":1750260071595,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:11.863+00:00|INFO|XacmlPdpStateChangeListener|KAFKA-source-policy-pdp-pap] PDP State Change message has been received from the PAP - PdpStateChange(super=PdpMessage(messageName=PDP_STATE_CHANGE, requestId=a42fb34f-6c41-47b7-b8c3-4403ad562bf0, timestampMs=1750260071595, name=xacml-3e8650f6-1f12-4eed-9943-19c94576807c, pdpGroup=defaultGroup, pdpSubgroup=xacml), source=pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac, state=ACTIVE) policy-xacml-pdp | [2025-06-18T15:21:11.864+00:00|INFO|XacmlState|KAFKA-source-policy-pdp-pap] set state of org.onap.policy.pdpx.main.XacmlState@629ae085 to ACTIVE policy-xacml-pdp | [2025-06-18T15:21:11.864+00:00|INFO|XacmlState|KAFKA-source-policy-pdp-pap] State change: ACTIVE - Starting rest controller policy-xacml-pdp | [2025-06-18T15:21:11.864+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","response":{"responseTo":"a42fb34f-6c41-47b7-b8c3-4403ad562bf0","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"098c88d2-eaa6-4bb4-8bb5-d4d90287d8d2","timestampMs":1750260071864,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:11.879+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","response":{"responseTo":"a42fb34f-6c41-47b7-b8c3-4403ad562bf0","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"098c88d2-eaa6-4bb4-8bb5-d4d90287d8d2","timestampMs":1750260071864,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:11.880+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-xacml-pdp | [2025-06-18T15:21:12.497+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"56da2523-d0d0-4363-80a8-ed2e3fe91e07","timestampMs":1750260072173,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:12.497+00:00|INFO|XacmlPdpUpdateListener|KAFKA-source-policy-pdp-pap] PDP update message has been received from the PAP - PdpUpdate(super=PdpMessage(messageName=PDP_UPDATE, requestId=56da2523-d0d0-4363-80a8-ed2e3fe91e07, timestampMs=1750260072173, name=xacml-3e8650f6-1f12-4eed-9943-19c94576807c, pdpGroup=defaultGroup, pdpSubgroup=xacml), source=pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac, description=null, pdpHeartbeatIntervalMs=120000, policiesToBeDeployed=[], policiesToBeUndeployed=[]) policy-xacml-pdp | [2025-06-18T15:21:12.498+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"response":{"responseTo":"56da2523-d0d0-4363-80a8-ed2e3fe91e07","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"7a9878aa-3df8-4441-a737-1eabfb3c9f90","timestampMs":1750260072498,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:12.507+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"}],"response":{"responseTo":"56da2523-d0d0-4363-80a8-ed2e3fe91e07","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"7a9878aa-3df8-4441-a737-1eabfb3c9f90","timestampMs":1750260072498,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:21:12.507+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-xacml-pdp | [2025-06-18T15:21:29.762+00:00|INFO|RequestLog|qtp2014233765-31] 172.17.0.1 - - [18/Jun/2025:15:21:29 +0000] "GET / HTTP/1.1" 401 423 "" "curl/7.58.0" policy-xacml-pdp | [2025-06-18T15:21:35.600+00:00|INFO|RequestLog|qtp2014233765-30] 172.17.0.4 - policyadmin [18/Jun/2025:15:21:35 +0000] "GET /metrics HTTP/1.1" 200 2132 "" "Prometheus/3.4.1" policy-xacml-pdp | [2025-06-18T15:22:17.796+00:00|INFO|RequestLog|qtp2014233765-31] 172.17.0.5 - policyadmin [18/Jun/2025:15:22:17 +0000] "GET /policy/pdpx/v1/healthcheck?null HTTP/1.1" 200 110 "" "python-requests/2.32.4" policy-xacml-pdp | [2025-06-18T15:22:17.817+00:00|INFO|RequestLog|qtp2014233765-26] 172.17.0.5 - policyadmin [18/Jun/2025:15:22:17 +0000] "GET /metrics?null HTTP/1.1" 200 2056 "" "python-requests/2.32.4" policy-xacml-pdp | [2025-06-18T15:22:19.243+00:00|INFO|GuardTranslator|qtp2014233765-26] Converting Request DecisionRequest(onapName=Guard, onapComponent=Guard-component, onapInstance=Guard-component-instance, requestId=unique-request-guard-1, context=null, action=guard, currentDateTime=null, currentDate=null, currentTime=null, timeZone=null, resource={guard={actor=APPC, operation=ModifyConfig, target=f17face5-69cb-4c88-9e0b-7426db7edddd, requestId=c7c6a4aa-bb61-4a15-b831-ba1472dd4a65, clname=ControlLoop-vFirewall-d0a1dfc6-94f5-4fd4-a5b5-4630b438850a}}) policy-xacml-pdp | [2025-06-18T15:22:19.262+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:oasis:names:tc:xacml:1.0:environment:current-dateTime policy-xacml-pdp | [2025-06-18T15:22:19.262+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:oasis:names:tc:xacml:1.0:environment:current-date policy-xacml-pdp | [2025-06-18T15:22:19.262+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:oasis:names:tc:xacml:1.0:environment:current-time policy-xacml-pdp | [2025-06-18T15:22:19.262+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:org:onap:guard:timezone policy-xacml-pdp | [2025-06-18T15:22:19.264+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:org:onap:guard:target:vf-count policy-xacml-pdp | [2025-06-18T15:22:19.264+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:org:onap:guard:target:generic-vnf.vnf-name policy-xacml-pdp | [2025-06-18T15:22:19.264+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:org:onap:guard:target:generic-vnf.vnf-id policy-xacml-pdp | [2025-06-18T15:22:19.264+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:org:onap:guard:target:generic-vnf.vnf-type policy-xacml-pdp | [2025-06-18T15:22:19.264+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:org:onap:guard:target:generic-vnf.nf-naming-code policy-xacml-pdp | [2025-06-18T15:22:19.265+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:org:onap:guard:target:vserver.vserver-id policy-xacml-pdp | [2025-06-18T15:22:19.265+00:00|WARN|RequestParser|qtp2014233765-26] Unable to extract attribute value from object: urn:org:onap:guard:target:cloud-region.cloud-region-id policy-xacml-pdp | [2025-06-18T15:22:19.269+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-26] Constructed using properties {count-recent-operations.persistenceunit=OperationsHistoryPU, get-operation-outcome.name=GetOperationOutcome, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:oasis:names:tc:xacml:3.0:policy-combining-algorithm:deny-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, count-recent-operations.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.CountRecentOperationsPip, get-operation-outcome.description=Returns operation outcome, count-recent-operations.description=Returns operation counts based on time window, jakarta.persistence.jdbc.password=policy_user, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, get-operation-outcome.issuer=urn:org:onap:xacml:guard:get-operation-outcome, get-operation-outcome.persistenceunit=OperationsHistoryPU, jakarta.persistence.jdbc.driver=org.postgresql.Driver, count-recent-operations.name=CountRecentOperations, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, jakarta.persistence.jdbc.url=jdbc:postgresql://postgres:5432/operationshistory, jakarta.persistence.jdbc.user=policy_user, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, count-recent-operations.issuer=urn:org:onap:xacml:guard:count-recent-operations, xacml.pip.engines=count-recent-operations,get-operation-outcome, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, get-operation-outcome.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.GetOperationOutcomePip} policy-xacml-pdp | [2025-06-18T15:22:19.269+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-26] Initializing OnapPolicyFinderFactory Properties policy-xacml-pdp | [2025-06-18T15:22:19.269+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-26] Combining root policies with urn:oasis:names:tc:xacml:3.0:policy-combining-algorithm:deny-overrides policy-xacml-pdp | [2025-06-18T15:22:19.276+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-26] Root Policies: 1 policy-xacml-pdp | [2025-06-18T15:22:19.276+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-26] Referenced Policies: 0 policy-xacml-pdp | [2025-06-18T15:22:19.277+00:00|INFO|StdPolicyFinder|qtp2014233765-26] Updating policy map with policy 603c075f-7a9f-4988-8f9f-895523d0a2a7 version 1.0 policy-xacml-pdp | [2025-06-18T15:22:19.281+00:00|INFO|StdOnapPip|qtp2014233765-26] Configuring historyDb PIP {count-recent-operations.persistenceunit=OperationsHistoryPU, get-operation-outcome.name=GetOperationOutcome, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:oasis:names:tc:xacml:3.0:policy-combining-algorithm:deny-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, count-recent-operations.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.CountRecentOperationsPip, get-operation-outcome.description=Returns operation outcome, count-recent-operations.description=Returns operation counts based on time window, jakarta.persistence.jdbc.password=policy_user, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, get-operation-outcome.issuer=urn:org:onap:xacml:guard:get-operation-outcome, get-operation-outcome.persistenceunit=OperationsHistoryPU, jakarta.persistence.jdbc.driver=org.postgresql.Driver, count-recent-operations.name=CountRecentOperations, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, jakarta.persistence.jdbc.url=jdbc:postgresql://postgres:5432/operationshistory, jakarta.persistence.jdbc.user=policy_user, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, count-recent-operations.issuer=urn:org:onap:xacml:guard:count-recent-operations, xacml.pip.engines=count-recent-operations,get-operation-outcome, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, get-operation-outcome.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.GetOperationOutcomePip} policy-xacml-pdp | [2025-06-18T15:22:19.374+00:00|INFO|LogHelper|qtp2014233765-26] HHH000204: Processing PersistenceUnitInfo [name: OperationsHistoryPU] policy-xacml-pdp | [2025-06-18T15:22:19.408+00:00|INFO|Version|qtp2014233765-26] HHH000412: Hibernate ORM core version 6.6.16.Final policy-xacml-pdp | [2025-06-18T15:22:19.433+00:00|INFO|RegionFactoryInitiator|qtp2014233765-26] HHH000026: Second-level cache disabled policy-xacml-pdp | [2025-06-18T15:22:19.569+00:00|WARN|pooling|qtp2014233765-26] HHH10001002: Using built-in connection pool (not intended for production use) policy-xacml-pdp | [2025-06-18T15:22:19.770+00:00|INFO|pooling|qtp2014233765-26] HHH10001005: Database info: policy-xacml-pdp | Database JDBC URL [jdbc:postgresql://postgres:5432/operationshistory] policy-xacml-pdp | Database driver: org.postgresql.Driver policy-xacml-pdp | Database version: 16.4 policy-xacml-pdp | Autocommit mode: false policy-xacml-pdp | Isolation level: undefined/unknown policy-xacml-pdp | Minimum pool size: 1 policy-xacml-pdp | Maximum pool size: 20 policy-xacml-pdp | [2025-06-18T15:22:20.631+00:00|INFO|JtaPlatformInitiator|qtp2014233765-26] HHH000489: No JTA platform available (set 'hibernate.transaction.jta.platform' to enable JTA platform integration) policy-xacml-pdp | [2025-06-18T15:22:20.663+00:00|INFO|StdOnapPip|qtp2014233765-26] Configuring historyDb PIP {count-recent-operations.persistenceunit=OperationsHistoryPU, get-operation-outcome.name=GetOperationOutcome, xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:oasis:names:tc:xacml:3.0:policy-combining-algorithm:deny-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, count-recent-operations.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.CountRecentOperationsPip, get-operation-outcome.description=Returns operation outcome, count-recent-operations.description=Returns operation counts based on time window, jakarta.persistence.jdbc.password=policy_user, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, get-operation-outcome.issuer=urn:org:onap:xacml:guard:get-operation-outcome, get-operation-outcome.persistenceunit=OperationsHistoryPU, jakarta.persistence.jdbc.driver=org.postgresql.Driver, count-recent-operations.name=CountRecentOperations, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, jakarta.persistence.jdbc.url=jdbc:postgresql://postgres:5432/operationshistory, jakarta.persistence.jdbc.user=policy_user, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, count-recent-operations.issuer=urn:org:onap:xacml:guard:count-recent-operations, xacml.pip.engines=count-recent-operations,get-operation-outcome, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory, get-operation-outcome.classname=org.onap.policy.pdp.xacml.application.common.operationshistory.GetOperationOutcomePip} policy-xacml-pdp | [2025-06-18T15:22:20.666+00:00|INFO|LogHelper|qtp2014233765-26] HHH000204: Processing PersistenceUnitInfo [name: OperationsHistoryPU] policy-xacml-pdp | [2025-06-18T15:22:20.668+00:00|INFO|RegionFactoryInitiator|qtp2014233765-26] HHH000026: Second-level cache disabled policy-xacml-pdp | [2025-06-18T15:22:20.685+00:00|WARN|pooling|qtp2014233765-26] HHH10001002: Using built-in connection pool (not intended for production use) policy-xacml-pdp | [2025-06-18T15:22:20.702+00:00|INFO|pooling|qtp2014233765-26] HHH10001005: Database info: policy-xacml-pdp | Database JDBC URL [jdbc:postgresql://postgres:5432/operationshistory] policy-xacml-pdp | Database driver: org.postgresql.Driver policy-xacml-pdp | Database version: 16.4 policy-xacml-pdp | Autocommit mode: false policy-xacml-pdp | Isolation level: undefined/unknown policy-xacml-pdp | Minimum pool size: 1 policy-xacml-pdp | Maximum pool size: 20 policy-xacml-pdp | [2025-06-18T15:22:20.733+00:00|INFO|JtaPlatformInitiator|qtp2014233765-26] HHH000489: No JTA platform available (set 'hibernate.transaction.jta.platform' to enable JTA platform integration) policy-xacml-pdp | [2025-06-18T15:22:20.736+00:00|INFO|StdXacmlApplicationServiceProvider|qtp2014233765-26] Elapsed Time: 1471ms policy-xacml-pdp | [2025-06-18T15:22:20.736+00:00|INFO|GuardTranslator|qtp2014233765-26] Converting Response {results=[{decision=NotApplicable,status={statusCode={statusCodeValue=urn:oasis:names:tc:xacml:1.0:status:ok}},attributeCategories=[{category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,attributes=[{attributeId=urn:oasis:names:tc:xacml:1.0:subject:subject-id,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=Guard}],includeInResults=true}{attributeId=urn:org:onap:onap-component,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=Guard-component}],includeInResults=true}{attributeId=urn:org:onap:onap-instance,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=Guard-component-instance}],includeInResults=true}{attributeId=urn:org:onap:guard:request:request-id,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=unique-request-guard-1}],includeInResults=true}]}{category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributes=[{attributeId=urn:org:onap:guard:clname:clname-id,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=ControlLoop-vFirewall-d0a1dfc6-94f5-4fd4-a5b5-4630b438850a}],includeInResults=true}{attributeId=urn:org:onap:guard:actor:actor-id,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=APPC}],includeInResults=true}{attributeId=urn:org:onap:guard:operation:operation-id,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=ModifyConfig}],includeInResults=true}{attributeId=urn:org:onap:guard:target:target-id,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=f17face5-69cb-4c88-9e0b-7426db7edddd}],includeInResults=true}]}]}]} policy-xacml-pdp | [2025-06-18T15:22:20.740+00:00|INFO|RequestLog|qtp2014233765-26] 172.17.0.5 - policyadmin [18/Jun/2025:15:22:19 +0000] "POST /policy/pdpx/v1/decision?abbrev=true HTTP/1.1" 200 19 "" "python-requests/2.32.4" policy-xacml-pdp | [2025-06-18T15:22:21.293+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","description":"The default group that registers all supported policy types and pdps.","policiesToBeDeployed":[{"type":"onap.policies.monitoring.tcagen2","type_version":"1.0.0","properties":{"tca.policy":{"domain":"measurementsForVfScaling","metricsPerEventName":[{"eventName":"Measurement_vGMUX","controlLoopSchemaType":"VNF","policyScope":"DCAE","policyName":"DCAE.Config_tca-hi-lo","policyVersion":"v0.0.1","thresholds":[{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"EQUAL","severity":"MAJOR","closedLoopEventStatus":"ABATED"},{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"GREATER","severity":"CRITICAL","closedLoopEventStatus":"ONSET"}]}]}},"name":"onap.restart.tca","version":"1.0.0","metadata":{"policy-id":"onap.restart.tca","policy-version":"1.0.0"}},{"type":"onap.policies.optimization.resource.AffinityPolicy","type_version":"1.0.0","properties":{"geography":[],"identity":"affinity_vCPE","scope":[],"affinityProperties":{"qualifier":"same","category":"complex"},"resources":[],"services":[],"applicableResources":"any"},"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0","metadata":{"policy-id":"OSDF_CASABLANCA.Affinity_Default","policy-version":"1.0.0"}}],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"fc472ffb-8754-4900-a29c-b7a068d958fd","timestampMs":1750260141233,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:22:21.295+00:00|INFO|XacmlPdpUpdateListener|KAFKA-source-policy-pdp-pap] PDP update message has been received from the PAP - PdpUpdate(super=PdpMessage(messageName=PDP_UPDATE, requestId=fc472ffb-8754-4900-a29c-b7a068d958fd, timestampMs=1750260141233, name=xacml-3e8650f6-1f12-4eed-9943-19c94576807c, pdpGroup=defaultGroup, pdpSubgroup=xacml), source=pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac, description=The default group that registers all supported policy types and pdps., pdpHeartbeatIntervalMs=null, policiesToBeDeployed=[ToscaPolicy(super=ToscaWithTypeAndObjectProperties(type=onap.policies.monitoring.tcagen2, typeVersion=1.0.0, properties={tca.policy={domain=measurementsForVfScaling, metricsPerEventName=[{eventName=Measurement_vGMUX, controlLoopSchemaType=VNF, policyScope=DCAE, policyName=DCAE.Config_tca-hi-lo, policyVersion=v0.0.1, thresholds=[{closedLoopControlName=ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e, version=1.0.2, fieldPath=$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value, thresholdValue=0, direction=EQUAL, severity=MAJOR, closedLoopEventStatus=ABATED}, {closedLoopControlName=ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e, version=1.0.2, fieldPath=$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value, thresholdValue=0, direction=GREATER, severity=CRITICAL, closedLoopEventStatus=ONSET}]}]}})), ToscaPolicy(super=ToscaWithTypeAndObjectProperties(type=onap.policies.optimization.resource.AffinityPolicy, typeVersion=1.0.0, properties={geography=[], identity=affinity_vCPE, scope=[], affinityProperties={qualifier=same, category=complex}, resources=[], services=[], applicableResources=any}))], policiesToBeUndeployed=[]) policy-xacml-pdp | [2025-06-18T15:22:21.295+00:00|INFO|StdBaseTranslator|KAFKA-source-policy-pdp-pap] Obligation Policy id: onap.restart.tca type: onap.policies.monitoring.tcagen2 weight: null policy: policy-xacml-pdp | {"type":"onap.policies.monitoring.tcagen2","type_version":"1.0.0","properties":{"tca.policy":{"domain":"measurementsForVfScaling","metricsPerEventName":[{"eventName":"Measurement_vGMUX","controlLoopSchemaType":"VNF","policyScope":"DCAE","policyName":"DCAE.Config_tca-hi-lo","policyVersion":"v0.0.1","thresholds":[{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"EQUAL","severity":"MAJOR","closedLoopEventStatus":"ABATED"},{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"GREATER","severity":"CRITICAL","closedLoopEventStatus":"ONSET"}]}]}},"name":"onap.restart.tca","version":"1.0.0","metadata":{"policy-id":"onap.restart.tca","policy-version":"1.0.0"}} policy-xacml-pdp | [2025-06-18T15:22:21.315+00:00|INFO|StdXacmlApplicationServiceProvider|KAFKA-source-policy-pdp-pap] Xacml Policy is policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.restart.tca policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.policies.monitoring.tcagen2 policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.policies.monitoring.tcagen2 policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | 1.0.0 policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | Default is to PERMIT if the policy matches. policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.restart.tca policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | {"type":"onap.policies.monitoring.tcagen2","type_version":"1.0.0","properties":{"tca.policy":{"domain":"measurementsForVfScaling","metricsPerEventName":[{"eventName":"Measurement_vGMUX","controlLoopSchemaType":"VNF","policyScope":"DCAE","policyName":"DCAE.Config_tca-hi-lo","policyVersion":"v0.0.1","thresholds":[{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"EQUAL","severity":"MAJOR","closedLoopEventStatus":"ABATED"},{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"GREATER","severity":"CRITICAL","closedLoopEventStatus":"ONSET"}]}]}},"name":"onap.restart.tca","version":"1.0.0","metadata":{"policy-id":"onap.restart.tca","policy-version":"1.0.0"}} policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.policies.monitoring.tcagen2 policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | [2025-06-18T15:22:21.315+00:00|INFO|XacmlPolicyUtils|KAFKA-source-policy-pdp-pap] Storing xacml properties {xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, root1.file=/opt/app/policy/pdpx/apps/monitoring/onap.restart.tca_1.0.0.xml, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, xacml.rootPolicies=root1, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.referencedPolicies=, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory} policy-xacml-pdp | /opt/app/policy/pdpx/apps/monitoring/xacml.properties policy-xacml-pdp | [2025-06-18T15:22:21.316+00:00|INFO|XacmlPdpApplicationManager|KAFKA-source-policy-pdp-pap] Loaded ToscaPolicy {policy-id=onap.restart.tca, policy-version=1.0.0} into application monitoring policy-xacml-pdp | [2025-06-18T15:22:21.316+00:00|INFO|OptimizationPdpApplication|KAFKA-source-policy-pdp-pap] optimization can support onap.policies.optimization.resource.AffinityPolicy 1.0.0 policy-xacml-pdp | [2025-06-18T15:22:21.317+00:00|ERROR|StdMatchableTranslator|KAFKA-source-policy-pdp-pap] PolicyType not found in data area yet /opt/app/policy/pdpx/apps/optimization/onap.policies.optimization.resource.AffinityPolicy-1.0.0.yaml policy-xacml-pdp | java.nio.file.NoSuchFileException: /opt/app/policy/pdpx/apps/optimization/onap.policies.optimization.resource.AffinityPolicy-1.0.0.yaml policy-xacml-pdp | at java.base/sun.nio.fs.UnixException.translateToIOException(UnixException.java:92) policy-xacml-pdp | at java.base/sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:106) policy-xacml-pdp | at java.base/sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:111) policy-xacml-pdp | at java.base/sun.nio.fs.UnixFileSystemProvider.newByteChannel(UnixFileSystemProvider.java:218) policy-xacml-pdp | at java.base/java.nio.file.Files.newByteChannel(Files.java:380) policy-xacml-pdp | at java.base/java.nio.file.Files.newByteChannel(Files.java:432) policy-xacml-pdp | at java.base/java.nio.file.Files.readAllBytes(Files.java:3288) policy-xacml-pdp | at org.onap.policy.pdp.xacml.application.common.std.StdMatchableTranslator.loadPolicyType(StdMatchableTranslator.java:515) policy-xacml-pdp | at org.onap.policy.pdp.xacml.application.common.std.StdMatchableTranslator.findPolicyType(StdMatchableTranslator.java:480) policy-xacml-pdp | at org.onap.policy.pdp.xacml.application.common.std.StdMatchableTranslator.convertPolicy(StdMatchableTranslator.java:241) policy-xacml-pdp | at org.onap.policy.xacml.pdp.application.optimization.OptimizationPdpApplicationTranslator.convertPolicy(OptimizationPdpApplicationTranslator.java:72) policy-xacml-pdp | at org.onap.policy.pdp.xacml.application.common.std.StdXacmlApplicationServiceProvider.loadPolicy(StdXacmlApplicationServiceProvider.java:127) policy-xacml-pdp | at org.onap.policy.pdpx.main.rest.XacmlPdpApplicationManager.loadDeployedPolicy(XacmlPdpApplicationManager.java:199) policy-xacml-pdp | at org.onap.policy.pdpx.main.comm.XacmlPdpUpdatePublisher.handlePdpUpdate(XacmlPdpUpdatePublisher.java:91) policy-xacml-pdp | at org.onap.policy.pdpx.main.comm.listeners.XacmlPdpUpdateListener.onTopicEvent(XacmlPdpUpdateListener.java:72) policy-xacml-pdp | at org.onap.policy.pdpx.main.comm.listeners.XacmlPdpUpdateListener.onTopicEvent(XacmlPdpUpdateListener.java:36) policy-xacml-pdp | at org.onap.policy.common.endpoints.listeners.ScoListener.onTopicEvent(ScoListener.java:75) policy-xacml-pdp | at org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher.onTopicEvent(MessageTypeDispatcher.java:97) policy-xacml-pdp | at org.onap.policy.common.endpoints.listeners.JsonListener.onTopicEvent(JsonListener.java:61) policy-xacml-pdp | at org.onap.policy.common.message.bus.event.base.TopicBase.broadcast(TopicBase.java:170) policy-xacml-pdp | at org.onap.policy.common.message.bus.event.base.SingleThreadedBusTopicSource.fetchAllMessages(SingleThreadedBusTopicSource.java:252) policy-xacml-pdp | at org.onap.policy.common.message.bus.event.base.SingleThreadedBusTopicSource.run(SingleThreadedBusTopicSource.java:235) policy-xacml-pdp | at java.base/java.lang.Thread.run(Thread.java:840) policy-xacml-pdp | [2025-06-18T15:22:21.347+00:00|INFO|GsonMessageBodyHandler|KAFKA-source-policy-pdp-pap] Using GSON for REST calls policy-xacml-pdp | [2025-06-18T15:22:21.350+00:00|INFO|GsonMessageBodyHandler|KAFKA-source-policy-pdp-pap] Using GSON for REST calls policy-xacml-pdp | [2025-06-18T15:22:21.792+00:00|INFO|StdMatchableTranslator|KAFKA-source-policy-pdp-pap] Successfully pulled onap.policies.optimization.resource.AffinityPolicy 1.0.0 policy-xacml-pdp | [2025-06-18T15:22:21.822+00:00|INFO|MatchablePolicyType|KAFKA-source-policy-pdp-pap] Scanning PolicyType onap.policies.optimization.resource.AffinityPolicy:1.0.0 policy-xacml-pdp | [2025-06-18T15:22:21.823+00:00|INFO|MatchablePolicyType|KAFKA-source-policy-pdp-pap] Retrieving datatype policy.data.affinityProperties_properties policy-xacml-pdp | [2025-06-18T15:22:21.823+00:00|INFO|MatchablePolicyType|KAFKA-source-policy-pdp-pap] Scanning PolicyType onap.policies.optimization.Resource:1.0.0 policy-xacml-pdp | [2025-06-18T15:22:21.823+00:00|INFO|MatchablePolicyType|KAFKA-source-policy-pdp-pap] Scanning PolicyType onap.policies.Optimization:1.0.0 policy-xacml-pdp | [2025-06-18T15:22:21.823+00:00|INFO|MatchablePolicyType|KAFKA-source-policy-pdp-pap] Found root - done scanning policy-xacml-pdp | [2025-06-18T15:22:21.823+00:00|INFO|StdBaseTranslator|KAFKA-source-policy-pdp-pap] Obligation Policy id: OSDF_CASABLANCA.Affinity_Default type: onap.policies.optimization.resource.AffinityPolicy weight: 0 policy: policy-xacml-pdp | {"type":"onap.policies.optimization.resource.AffinityPolicy","type_version":"1.0.0","properties":{"geography":[],"identity":"affinity_vCPE","scope":[],"affinityProperties":{"qualifier":"same","category":"complex"},"resources":[],"services":[],"applicableResources":"any"},"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0","metadata":{"policy-id":"OSDF_CASABLANCA.Affinity_Default","policy-version":"1.0.0"}} policy-xacml-pdp | [2025-06-18T15:22:21.840+00:00|INFO|StdMatchableTranslator|KAFKA-source-policy-pdp-pap] policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | Default is to PERMIT if the policy matches. policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | IF exists and is equal policy-xacml-pdp | policy-xacml-pdp | Does the policy-type attribute exist? policy-xacml-pdp | policy-xacml-pdp | Get the size of policy-type attributes policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | 0 policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | Is this policy-type in the list? policy-xacml-pdp | onap.policies.optimization.resource.AffinityPolicy policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | OSDF_CASABLANCA.Affinity_Default policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | {"type":"onap.policies.optimization.resource.AffinityPolicy","type_version":"1.0.0","properties":{"geography":[],"identity":"affinity_vCPE","scope":[],"affinityProperties":{"qualifier":"same","category":"complex"},"resources":[],"services":[],"applicableResources":"any"},"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0","metadata":{"policy-id":"OSDF_CASABLANCA.Affinity_Default","policy-version":"1.0.0"}} policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | 0 policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.policies.optimization.resource.AffinityPolicy policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | [2025-06-18T15:22:21.861+00:00|INFO|StdXacmlApplicationServiceProvider|KAFKA-source-policy-pdp-pap] Xacml Policy is policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | Default is to PERMIT if the policy matches. policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | IF exists and is equal policy-xacml-pdp | policy-xacml-pdp | Does the policy-type attribute exist? policy-xacml-pdp | policy-xacml-pdp | Get the size of policy-type attributes policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | 0 policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | Is this policy-type in the list? policy-xacml-pdp | onap.policies.optimization.resource.AffinityPolicy policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | OSDF_CASABLANCA.Affinity_Default policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | {"type":"onap.policies.optimization.resource.AffinityPolicy","type_version":"1.0.0","properties":{"geography":[],"identity":"affinity_vCPE","scope":[],"affinityProperties":{"qualifier":"same","category":"complex"},"resources":[],"services":[],"applicableResources":"any"},"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0","metadata":{"policy-id":"OSDF_CASABLANCA.Affinity_Default","policy-version":"1.0.0"}} policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | 0 policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | onap.policies.optimization.resource.AffinityPolicy policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | policy-xacml-pdp | [2025-06-18T15:22:21.861+00:00|INFO|XacmlPolicyUtils|KAFKA-source-policy-pdp-pap] Storing xacml properties {xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, root1.file=/opt/app/policy/pdpx/apps/optimization/OSDF_CASABLANCA.Affinity_Default_1.0.0.xml, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, xacml.rootPolicies=root1, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.referencedPolicies=, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory} policy-xacml-pdp | /opt/app/policy/pdpx/apps/optimization/xacml.properties policy-xacml-pdp | [2025-06-18T15:22:21.861+00:00|INFO|XacmlPdpApplicationManager|KAFKA-source-policy-pdp-pap] Loaded ToscaPolicy {policy-id=OSDF_CASABLANCA.Affinity_Default, policy-version=1.0.0} into application optimization policy-xacml-pdp | [2025-06-18T15:22:21.862+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"onap.restart.tca","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"response":{"responseTo":"fc472ffb-8754-4900-a29c-b7a068d958fd","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"b429c97c-fe37-468c-9a06-025331b71180","timestampMs":1750260141862,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:22:21.876+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"onap.restart.tca","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"response":{"responseTo":"fc472ffb-8754-4900-a29c-b7a068d958fd","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"b429c97c-fe37-468c-9a06-025331b71180","timestampMs":1750260141862,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:22:21.877+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-xacml-pdp | [2025-06-18T15:22:35.575+00:00|INFO|RequestLog|qtp2014233765-27] 172.17.0.4 - policyadmin [18/Jun/2025:15:22:35 +0000] "GET /metrics HTTP/1.1" 200 2178 "" "Prometheus/3.4.1" policy-xacml-pdp | [2025-06-18T15:22:45.492+00:00|INFO|StdCombinedPolicyResultsTranslator|qtp2014233765-29] Converting Request DecisionRequest(onapName=DCAE, onapComponent=PolicyHandler, onapInstance=622431a4-9dea-4eae-b443-3b2164639c64, requestId=null, context=null, action=configure, currentDateTime=null, currentDate=null, currentTime=null, timeZone=null, resource={policy-id=onap.restart.tca}) policy-xacml-pdp | [2025-06-18T15:22:45.493+00:00|WARN|RequestParser|qtp2014233765-29] Unable to extract attribute value from object: urn:org:onap:policy-type policy-xacml-pdp | [2025-06-18T15:22:45.494+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-29] Constructed using properties {xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, root1.file=/opt/app/policy/pdpx/apps/monitoring/onap.restart.tca_1.0.0.xml, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, xacml.rootPolicies=root1, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.referencedPolicies=, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory} policy-xacml-pdp | [2025-06-18T15:22:45.494+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-29] Initializing OnapPolicyFinderFactory Properties policy-xacml-pdp | [2025-06-18T15:22:45.494+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-29] Combining root policies with urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides policy-xacml-pdp | [2025-06-18T15:22:45.495+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-29] Loading policy file /opt/app/policy/pdpx/apps/monitoring/onap.restart.tca_1.0.0.xml policy-xacml-pdp | [2025-06-18T15:22:45.521+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-29] Root Policies: 1 policy-xacml-pdp | [2025-06-18T15:22:45.521+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-29] Referenced Policies: 0 policy-xacml-pdp | [2025-06-18T15:22:45.521+00:00|INFO|StdPolicyFinder|qtp2014233765-29] Updating policy map with policy 118afde7-156f-4000-947a-41c82f3d022c version 1.0 policy-xacml-pdp | [2025-06-18T15:22:45.521+00:00|INFO|StdPolicyFinder|qtp2014233765-29] Updating policy map with policy onap.restart.tca version 1.0.0 policy-xacml-pdp | [2025-06-18T15:22:45.542+00:00|INFO|StdXacmlApplicationServiceProvider|qtp2014233765-29] Elapsed Time: 48ms policy-xacml-pdp | [2025-06-18T15:22:45.542+00:00|INFO|StdBaseTranslator|qtp2014233765-29] Converting Response {results=[{decision=Permit,status={statusCode={statusCodeValue=urn:oasis:names:tc:xacml:1.0:status:ok}},obligations=[{id=urn:org:onap:rest:body,attributeAssignments=[{attributeId=urn:org:onap::obligation:policyid,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=onap.restart.tca}}{attributeId=urn:org:onap::obligation:policycontent,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value={"type":"onap.policies.monitoring.tcagen2","type_version":"1.0.0","properties":{"tca.policy":{"domain":"measurementsForVfScaling","metricsPerEventName":[{"eventName":"Measurement_vGMUX","controlLoopSchemaType":"VNF","policyScope":"DCAE","policyName":"DCAE.Config_tca-hi-lo","policyVersion":"v0.0.1","thresholds":[{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"EQUAL","severity":"MAJOR","closedLoopEventStatus":"ABATED"},{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"GREATER","severity":"CRITICAL","closedLoopEventStatus":"ONSET"}]}]}},"name":"onap.restart.tca","version":"1.0.0","metadata":{"policy-id":"onap.restart.tca","policy-version":"1.0.0"}}}}{attributeId=urn:org:onap::obligation:policytype,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=onap.policies.monitoring.tcagen2}}]}],attributeCategories=[{category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,attributes=[{attributeId=urn:oasis:names:tc:xacml:1.0:subject:subject-id,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=DCAE}],includeInResults=true}{attributeId=urn:org:onap:onap-component,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=PolicyHandler}],includeInResults=true}{attributeId=urn:org:onap:onap-instance,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=622431a4-9dea-4eae-b443-3b2164639c64}],includeInResults=true}]}{category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributes=[{attributeId=urn:oasis:names:tc:xacml:1.0:resource:resource-id,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=onap.restart.tca}],includeInResults=true}]}],policyIdentifiers=[{id=onap.restart.tca,version=1.0.0}],policySetIdentifiers=[{id=118afde7-156f-4000-947a-41c82f3d022c,version=1.0}]}]} policy-xacml-pdp | [2025-06-18T15:22:45.543+00:00|INFO|StdCombinedPolicyResultsTranslator|qtp2014233765-29] Obligation: urn:org:onap:rest:body policy-xacml-pdp | [2025-06-18T15:22:45.543+00:00|WARN|StdCombinedPolicyResultsTranslator|qtp2014233765-29] Advice found - not supported in this class class org.onap.policy.pdp.xacml.application.common.std.StdCombinedPolicyResultsTranslator policy-xacml-pdp | [2025-06-18T15:22:45.543+00:00|INFO|MonitoringPdpApplication|qtp2014233765-29] Abbreviating decision results DecisionResponse(status=null, message=null, advice=null, obligations=null, policies={onap.restart.tca={type=onap.policies.monitoring.tcagen2, type_version=1.0.0, properties={tca.policy={domain=measurementsForVfScaling, metricsPerEventName=[{eventName=Measurement_vGMUX, controlLoopSchemaType=VNF, policyScope=DCAE, policyName=DCAE.Config_tca-hi-lo, policyVersion=v0.0.1, thresholds=[{closedLoopControlName=ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e, version=1.0.2, fieldPath=$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value, thresholdValue=0, direction=EQUAL, severity=MAJOR, closedLoopEventStatus=ABATED}, {closedLoopControlName=ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e, version=1.0.2, fieldPath=$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value, thresholdValue=0, direction=GREATER, severity=CRITICAL, closedLoopEventStatus=ONSET}]}]}}, name=onap.restart.tca, version=1.0.0, metadata={policy-id=onap.restart.tca, policy-version=1.0.0}}}, attributes=null) policy-xacml-pdp | [2025-06-18T15:22:45.547+00:00|INFO|RequestLog|qtp2014233765-29] 172.17.0.5 - policyadmin [18/Jun/2025:15:22:45 +0000] "POST /policy/pdpx/v1/decision?abbrev=true HTTP/1.1" 200 146 "" "python-requests/2.32.4" policy-xacml-pdp | [2025-06-18T15:22:45.559+00:00|INFO|StdCombinedPolicyResultsTranslator|qtp2014233765-29] Converting Request DecisionRequest(onapName=DCAE, onapComponent=PolicyHandler, onapInstance=622431a4-9dea-4eae-b443-3b2164639c64, requestId=null, context=null, action=configure, currentDateTime=null, currentDate=null, currentTime=null, timeZone=null, resource={policy-id=onap.restart.tca}) policy-xacml-pdp | [2025-06-18T15:22:45.560+00:00|WARN|RequestParser|qtp2014233765-29] Unable to extract attribute value from object: urn:org:onap:policy-type policy-xacml-pdp | [2025-06-18T15:22:45.561+00:00|INFO|StdXacmlApplicationServiceProvider|qtp2014233765-29] Elapsed Time: 1ms policy-xacml-pdp | [2025-06-18T15:22:45.561+00:00|INFO|StdBaseTranslator|qtp2014233765-29] Converting Response {results=[{decision=Permit,status={statusCode={statusCodeValue=urn:oasis:names:tc:xacml:1.0:status:ok}},obligations=[{id=urn:org:onap:rest:body,attributeAssignments=[{attributeId=urn:org:onap::obligation:policyid,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=onap.restart.tca}}{attributeId=urn:org:onap::obligation:policycontent,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value={"type":"onap.policies.monitoring.tcagen2","type_version":"1.0.0","properties":{"tca.policy":{"domain":"measurementsForVfScaling","metricsPerEventName":[{"eventName":"Measurement_vGMUX","controlLoopSchemaType":"VNF","policyScope":"DCAE","policyName":"DCAE.Config_tca-hi-lo","policyVersion":"v0.0.1","thresholds":[{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"EQUAL","severity":"MAJOR","closedLoopEventStatus":"ABATED"},{"closedLoopControlName":"ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e","version":"1.0.2","fieldPath":"$.event.measurementsForVfScalingFields.additionalMeasurements[*].arrayOfFields[0].value","thresholdValue":0,"direction":"GREATER","severity":"CRITICAL","closedLoopEventStatus":"ONSET"}]}]}},"name":"onap.restart.tca","version":"1.0.0","metadata":{"policy-id":"onap.restart.tca","policy-version":"1.0.0"}}}}{attributeId=urn:org:onap::obligation:policytype,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=onap.policies.monitoring.tcagen2}}]}],attributeCategories=[{category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,attributes=[{attributeId=urn:oasis:names:tc:xacml:1.0:subject:subject-id,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=DCAE}],includeInResults=true}{attributeId=urn:org:onap:onap-component,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=PolicyHandler}],includeInResults=true}{attributeId=urn:org:onap:onap-instance,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=622431a4-9dea-4eae-b443-3b2164639c64}],includeInResults=true}]}{category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributes=[{attributeId=urn:oasis:names:tc:xacml:1.0:resource:resource-id,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=onap.restart.tca}],includeInResults=true}]}],policyIdentifiers=[{id=onap.restart.tca,version=1.0.0}],policySetIdentifiers=[{id=118afde7-156f-4000-947a-41c82f3d022c,version=1.0}]}]} policy-xacml-pdp | [2025-06-18T15:22:45.561+00:00|INFO|StdCombinedPolicyResultsTranslator|qtp2014233765-29] Obligation: urn:org:onap:rest:body policy-xacml-pdp | [2025-06-18T15:22:45.562+00:00|WARN|StdCombinedPolicyResultsTranslator|qtp2014233765-29] Advice found - not supported in this class class org.onap.policy.pdp.xacml.application.common.std.StdCombinedPolicyResultsTranslator policy-xacml-pdp | [2025-06-18T15:22:45.562+00:00|INFO|MonitoringPdpApplication|qtp2014233765-29] Unsupported query param for Monitoring application: {null=[]} policy-xacml-pdp | [2025-06-18T15:22:45.564+00:00|INFO|RequestLog|qtp2014233765-29] 172.17.0.5 - policyadmin [18/Jun/2025:15:22:45 +0000] "POST /policy/pdpx/v1/decision?null HTTP/1.1" 200 1055 "" "python-requests/2.32.4" policy-xacml-pdp | [2025-06-18T15:22:45.580+00:00|INFO|StdCombinedPolicyResultsTranslator|qtp2014233765-32] Converting Request DecisionRequest(onapName=SDNC, onapComponent=SDNC-component, onapInstance=SDNC-component-instance, requestId=unique-request-sdnc-1, context=null, action=naming, currentDateTime=null, currentDate=null, currentTime=null, timeZone=null, resource={nfRole=[], naming-type=[], property-name=[], policy-type=[onap.policies.Naming]}) policy-xacml-pdp | [2025-06-18T15:22:45.581+00:00|WARN|RequestParser|qtp2014233765-32] Unable to extract attribute value from object: urn:oasis:names:tc:xacml:1.0:resource:resource-id policy-xacml-pdp | [2025-06-18T15:22:45.581+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-32] Constructed using properties {xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, root1.file=/opt/app/policy/pdpx/apps/naming/SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP_1.0.0.xml, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, xacml.rootPolicies=root1, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.referencedPolicies=, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory} policy-xacml-pdp | [2025-06-18T15:22:45.581+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-32] Initializing OnapPolicyFinderFactory Properties policy-xacml-pdp | [2025-06-18T15:22:45.581+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-32] Combining root policies with urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides policy-xacml-pdp | [2025-06-18T15:22:45.582+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-32] Loading policy file /opt/app/policy/pdpx/apps/naming/SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP_1.0.0.xml policy-xacml-pdp | [2025-06-18T15:22:45.590+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-32] Root Policies: 1 policy-xacml-pdp | [2025-06-18T15:22:45.590+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-32] Referenced Policies: 0 policy-xacml-pdp | [2025-06-18T15:22:45.590+00:00|INFO|StdPolicyFinder|qtp2014233765-32] Updating policy map with policy 2f92c80c-1eba-409a-8c02-f87f83ae6527 version 1.0 policy-xacml-pdp | [2025-06-18T15:22:45.590+00:00|INFO|StdPolicyFinder|qtp2014233765-32] Updating policy map with policy SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP version 1.0.0 policy-xacml-pdp | [2025-06-18T15:22:45.592+00:00|INFO|StdXacmlApplicationServiceProvider|qtp2014233765-32] Elapsed Time: 11ms policy-xacml-pdp | [2025-06-18T15:22:45.592+00:00|INFO|StdBaseTranslator|qtp2014233765-32] Converting Response {results=[{decision=Permit,status={statusCode={statusCodeValue=urn:oasis:names:tc:xacml:1.0:status:ok}},obligations=[{id=urn:org:onap:rest:body,attributeAssignments=[{attributeId=urn:org:onap::obligation:policyid,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP}}{attributeId=urn:org:onap::obligation:policycontent,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value={"type":"onap.policies.Naming","type_version":"1.0.0","properties":{"policy-instance-name":"ONAP_NF_NAMING_TIMESTAMP","naming-models":[{"naming-type":"VNF","naming-recipe":"AIC_CLOUD_REGION|DELIMITER|CONSTANT|DELIMITER|TIMESTAMP","name-operation":"to_lower_case()","naming-properties":[{"property-name":"AIC_CLOUD_REGION"},{"property-name":"CONSTANT","property-value":"onap-nf"},{"property-name":"TIMESTAMP"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VNFC","naming-recipe":"VNF_NAME|DELIMITER|NFC_NAMING_CODE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"ENTIRETY","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}},{"property-name":"NFC_NAMING_CODE"},{"property-value":"-","property-name":"DELIMITER"}]},{"naming-type":"VF-MODULE","naming-recipe":"VNF_NAME|DELIMITER|VF_MODULE_LABEL|DELIMITER|VF_MODULE_TYPE|DELIMITER|SEQUENCE","name-operation":"to_lower_case()","naming-properties":[{"property-name":"VNF_NAME"},{"property-value":"-","property-name":"DELIMITER"},{"property-name":"VF_MODULE_LABEL"},{"property-name":"VF_MODULE_TYPE"},{"property-name":"SEQUENCE","increment-sequence":{"max":"zzz","scope":"PRECEEDING","start-value":"1","length":"3","increment":"1","sequence-type":"alpha-numeric"}}]}]},"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0","metadata":{"policy-id":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","policy-version":"1.0.0"}}}}{attributeId=urn:org:onap::obligation:policytype,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=onap.policies.Naming}}]}],attributeCategories=[{category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,attributes=[{attributeId=urn:oasis:names:tc:xacml:1.0:subject:subject-id,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=SDNC}],includeInResults=true}{attributeId=urn:org:onap:onap-component,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=SDNC-component}],includeInResults=true}{attributeId=urn:org:onap:onap-instance,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=SDNC-component-instance}],includeInResults=true}]}{category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributes=[{attributeId=urn:org:onap:policy-type,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=onap.policies.Naming}],includeInResults=true}]}],policyIdentifiers=[{id=SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP,version=1.0.0}],policySetIdentifiers=[{id=2f92c80c-1eba-409a-8c02-f87f83ae6527,version=1.0}]}]} policy-xacml-pdp | [2025-06-18T15:22:45.592+00:00|INFO|StdCombinedPolicyResultsTranslator|qtp2014233765-32] Obligation: urn:org:onap:rest:body policy-xacml-pdp | [2025-06-18T15:22:45.592+00:00|WARN|StdCombinedPolicyResultsTranslator|qtp2014233765-32] Advice found - not supported in this class class org.onap.policy.pdp.xacml.application.common.std.StdCombinedPolicyResultsTranslator policy-xacml-pdp | [2025-06-18T15:22:45.595+00:00|INFO|RequestLog|qtp2014233765-32] 172.17.0.5 - policyadmin [18/Jun/2025:15:22:45 +0000] "POST /policy/pdpx/v1/decision?null HTTP/1.1" 200 1598 "" "python-requests/2.32.4" policy-xacml-pdp | [2025-06-18T15:22:45.613+00:00|INFO|StdMatchableTranslator|qtp2014233765-30] Converting Request DecisionRequest(onapName=OOF, onapComponent=OOF-component, onapInstance=OOF-component-instance, requestId=null, context={subscriberName=[]}, action=optimize, currentDateTime=null, currentDate=null, currentTime=null, timeZone=null, resource={scope=[], services=[], resources=[], geography=[]}) policy-xacml-pdp | [2025-06-18T15:22:45.615+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-30] Constructed using properties {xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, root1.file=/opt/app/policy/pdpx/apps/optimization/OSDF_CASABLANCA.Affinity_Default_1.0.0.xml, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, xacml.rootPolicies=root1, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.referencedPolicies=, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory} policy-xacml-pdp | [2025-06-18T15:22:45.615+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-30] Initializing OnapPolicyFinderFactory Properties policy-xacml-pdp | [2025-06-18T15:22:45.615+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-30] Combining root policies with urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides policy-xacml-pdp | [2025-06-18T15:22:45.615+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-30] Loading policy file /opt/app/policy/pdpx/apps/optimization/OSDF_CASABLANCA.Affinity_Default_1.0.0.xml policy-xacml-pdp | [2025-06-18T15:22:45.624+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-30] Root Policies: 1 policy-xacml-pdp | [2025-06-18T15:22:45.624+00:00|INFO|OnapPolicyFinderFactory|qtp2014233765-30] Referenced Policies: 0 policy-xacml-pdp | [2025-06-18T15:22:45.624+00:00|INFO|StdPolicyFinder|qtp2014233765-30] Updating policy map with policy 728d1b7b-c951-4681-8b4d-fc8e4c944cea version 1.0 policy-xacml-pdp | [2025-06-18T15:22:45.624+00:00|INFO|StdPolicyFinder|qtp2014233765-30] Updating policy map with policy OSDF_CASABLANCA.Affinity_Default version 1.0.0 policy-xacml-pdp | [2025-06-18T15:22:45.626+00:00|INFO|StdXacmlApplicationServiceProvider|qtp2014233765-30] Elapsed Time: 12ms policy-xacml-pdp | [2025-06-18T15:22:45.626+00:00|INFO|StdBaseTranslator|qtp2014233765-30] Converting Response {results=[{decision=Permit,status={statusCode={statusCodeValue=urn:oasis:names:tc:xacml:1.0:status:ok}},obligations=[{id=urn:org:onap:rest:body,attributeAssignments=[{attributeId=urn:org:onap::obligation:policyid,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=OSDF_CASABLANCA.Affinity_Default}}{attributeId=urn:org:onap::obligation:policycontent,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value={"type":"onap.policies.optimization.resource.AffinityPolicy","type_version":"1.0.0","properties":{"geography":[],"identity":"affinity_vCPE","scope":[],"affinityProperties":{"qualifier":"same","category":"complex"},"resources":[],"services":[],"applicableResources":"any"},"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0","metadata":{"policy-id":"OSDF_CASABLANCA.Affinity_Default","policy-version":"1.0.0"}}}}{attributeId=urn:org:onap::obligation:weight,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#integer,value=0}}{attributeId=urn:org:onap::obligation:policytype,category=urn:oasis:names:tc:xacml:3.0:attribute-category:resource,attributeValue={dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=onap.policies.optimization.resource.AffinityPolicy}}]}],attributeCategories=[{category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,attributes=[{attributeId=urn:oasis:names:tc:xacml:1.0:subject:subject-id,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=OOF}],includeInResults=true}{attributeId=urn:org:onap:onap-component,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=OOF-component}],includeInResults=true}{attributeId=urn:org:onap:onap-instance,category=urn:oasis:names:tc:xacml:1.0:subject-category:access-subject,values=[{dataTypeId=http://www.w3.org/2001/XMLSchema#string,value=OOF-component-instance}],includeInResults=true}]}],policyIdentifiers=[{id=OSDF_CASABLANCA.Affinity_Default,version=1.0.0}],policySetIdentifiers=[{id=728d1b7b-c951-4681-8b4d-fc8e4c944cea,version=1.0}]}]} policy-xacml-pdp | [2025-06-18T15:22:45.626+00:00|INFO|StdMatchableTranslator|qtp2014233765-30] Obligation: urn:org:onap:rest:body policy-xacml-pdp | [2025-06-18T15:22:45.627+00:00|INFO|StdMatchableTranslator|qtp2014233765-30] New entry onap.policies.optimization.resource.AffinityPolicy weight 0 policy-xacml-pdp | [2025-06-18T15:22:45.627+00:00|INFO|StdMatchableTranslator|qtp2014233765-30] Policy (OSDF_CASABLANCA.Affinity_Default,{type=onap.policies.optimization.resource.AffinityPolicy, type_version=1.0.0, properties={geography=[], identity=affinity_vCPE, scope=[], affinityProperties={qualifier=same, category=complex}, resources=[], services=[], applicableResources=any}, name=OSDF_CASABLANCA.Affinity_Default, version=1.0.0, metadata={policy-id=OSDF_CASABLANCA.Affinity_Default, policy-version=1.0.0}}) policy-xacml-pdp | [2025-06-18T15:22:45.629+00:00|INFO|RequestLog|qtp2014233765-30] 172.17.0.5 - policyadmin [18/Jun/2025:15:22:45 +0000] "POST /policy/pdpx/v1/decision?null HTTP/1.1" 200 467 "" "python-requests/2.32.4" policy-xacml-pdp | [2025-06-18T15:22:46.054+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"source":"pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac","description":"The default group that registers all supported policy types and pdps.","policiesToBeDeployed":[],"policiesToBeUndeployed":[{"name":"onap.restart.tca","version":"1.0.0"}],"messageName":"PDP_UPDATE","requestId":"e29f5c97-3aec-4ad6-b58e-ee474839d46d","timestampMs":1750260166019,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:22:46.054+00:00|INFO|XacmlPdpUpdateListener|KAFKA-source-policy-pdp-pap] PDP update message has been received from the PAP - PdpUpdate(super=PdpMessage(messageName=PDP_UPDATE, requestId=e29f5c97-3aec-4ad6-b58e-ee474839d46d, timestampMs=1750260166019, name=xacml-3e8650f6-1f12-4eed-9943-19c94576807c, pdpGroup=defaultGroup, pdpSubgroup=xacml), source=pap-ccba7b36-c5a4-4176-a47a-89bf8a8c84ac, description=The default group that registers all supported policy types and pdps., pdpHeartbeatIntervalMs=null, policiesToBeDeployed=[], policiesToBeUndeployed=[onap.restart.tca 1.0.0]) policy-xacml-pdp | [2025-06-18T15:22:46.055+00:00|ERROR|StdXacmlApplicationServiceProvider|KAFKA-source-policy-pdp-pap] Failed to find ToscaPolicy {policy-id=onap.restart.tca, policy-version=1.0.0} in our map size 0 policy-xacml-pdp | [2025-06-18T15:22:46.055+00:00|ERROR|StdXacmlApplicationServiceProvider|KAFKA-source-policy-pdp-pap] Failed to find ToscaPolicy {policy-id=onap.restart.tca, policy-version=1.0.0} in our map size 1 policy-xacml-pdp | [2025-06-18T15:22:46.055+00:00|ERROR|StdXacmlApplicationServiceProvider|KAFKA-source-policy-pdp-pap] Failed to find ToscaPolicy {policy-id=onap.restart.tca, policy-version=1.0.0} in our map size 1 policy-xacml-pdp | [2025-06-18T15:22:46.055+00:00|ERROR|StdXacmlApplicationServiceProvider|KAFKA-source-policy-pdp-pap] Failed to find ToscaPolicy {policy-id=onap.restart.tca, policy-version=1.0.0} in our map size 0 policy-xacml-pdp | [2025-06-18T15:22:46.055+00:00|ERROR|StdXacmlApplicationServiceProvider|KAFKA-source-policy-pdp-pap] Failed to find ToscaPolicy {policy-id=onap.restart.tca, policy-version=1.0.0} in our map size 0 policy-xacml-pdp | [2025-06-18T15:22:46.055+00:00|INFO|XacmlPolicyUtils|KAFKA-source-policy-pdp-pap] Storing xacml properties {xacml.att.evaluationContextFactory=com.att.research.xacmlatt.pdp.std.StdEvaluationContextFactory, xacml.pepEngineFactory=com.att.research.xacml.std.pep.StdEngineFactory, xacml.dataTypeFactory=com.att.research.xacml.std.StdDataTypeFactory, xacml.att.policyFinderFactory.combineRootPolicies=urn:com:att:xacml:3.0:policy-combining-algorithm:combined-permit-overrides, xacml.att.policyFinderFactory=org.onap.policy.pdp.xacml.application.common.OnapPolicyFinderFactory, xacml.att.functionDefinitionFactory=com.att.research.xacmlatt.pdp.std.StdFunctionDefinitionFactory, xacml.rootPolicies=, xacml.att.combiningAlgorithmFactory=com.att.research.xacmlatt.pdp.std.StdCombiningAlgorithmFactory, xacml.referencedPolicies=, xacml.pdpEngineFactory=com.att.research.xacmlatt.pdp.ATTPDPEngineFactory, xacml.traceEngineFactory=com.att.research.xacml.std.trace.LoggingTraceEngineFactory, xacml.pipFinderFactory=com.att.research.xacml.std.pip.StdPIPFinderFactory} policy-xacml-pdp | /opt/app/policy/pdpx/apps/monitoring/xacml.properties policy-xacml-pdp | [2025-06-18T15:22:46.056+00:00|INFO|XacmlPdpApplicationManager|KAFKA-source-policy-pdp-pap] Unloaded ToscaPolicy {policy-id=onap.restart.tca, policy-version=1.0.0} from application monitoring policy-xacml-pdp | [2025-06-18T15:22:46.056+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"response":{"responseTo":"e29f5c97-3aec-4ad6-b58e-ee474839d46d","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"e1486330-7898-4398-9ab4-834a88c41277","timestampMs":1750260166056,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:22:46.063+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"response":{"responseTo":"e29f5c97-3aec-4ad6-b58e-ee474839d46d","responseStatus":"SUCCESS"},"messageName":"PDP_STATUS","requestId":"e1486330-7898-4398-9ab4-834a88c41277","timestampMs":1750260166056,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:22:46.063+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-xacml-pdp | [2025-06-18T15:23:11.813+00:00|INFO|XacmlPdpHearbeatPublisher|pool-2-thread-1] Sending Xacml PDP heartbeat to the PAP - PdpStatus(super=PdpMessage(messageName=PDP_STATUS, requestId=30f5a157-2cce-4257-a432-84d1175d1b45, timestampMs=1750260191813, name=xacml-3e8650f6-1f12-4eed-9943-19c94576807c, pdpGroup=defaultGroup, pdpSubgroup=xacml), pdpType=xacml, state=ACTIVE, healthy=HEALTHY, description=null, policies=[SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP 1.0.0, OSDF_CASABLANCA.Affinity_Default 1.0.0], deploymentInstanceInfo=null, properties=null, response=null) policy-xacml-pdp | [2025-06-18T15:23:11.814+00:00|INFO|network|pool-2-thread-1] [OUT|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"messageName":"PDP_STATUS","requestId":"30f5a157-2cce-4257-a432-84d1175d1b45","timestampMs":1750260191813,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:23:11.824+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-xacml-pdp | {"pdpType":"xacml","state":"ACTIVE","healthy":"HEALTHY","policies":[{"name":"SDNC_Policy.ONAP_NF_NAMING_TIMESTAMP","version":"1.0.0"},{"name":"OSDF_CASABLANCA.Affinity_Default","version":"1.0.0"}],"messageName":"PDP_STATUS","requestId":"30f5a157-2cce-4257-a432-84d1175d1b45","timestampMs":1750260191813,"name":"xacml-3e8650f6-1f12-4eed-9943-19c94576807c","pdpGroup":"defaultGroup","pdpSubgroup":"xacml"} policy-xacml-pdp | [2025-06-18T15:23:11.824+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-xacml-pdp | [2025-06-18T15:23:35.575+00:00|INFO|RequestLog|qtp2014233765-27] 172.17.0.4 - policyadmin [18/Jun/2025:15:23:35 +0000] "GET /metrics HTTP/1.1" 200 2220 "" "Prometheus/3.4.1" postgres | The files belonging to this database system will be owned by user "postgres". postgres | This user must also own the server process. postgres | postgres | The database cluster will be initialized with locale "en_US.utf8". postgres | The default database encoding has accordingly been set to "UTF8". postgres | The default text search configuration will be set to "english". postgres | postgres | Data page checksums are disabled. postgres | postgres | fixing permissions on existing directory /var/lib/postgresql/data ... ok postgres | creating subdirectories ... ok postgres | selecting dynamic shared memory implementation ... posix postgres | selecting default max_connections ... 100 postgres | selecting default shared_buffers ... 128MB postgres | selecting default time zone ... Etc/UTC postgres | creating configuration files ... ok postgres | running bootstrap script ... ok postgres | performing post-bootstrap initialization ... ok postgres | syncing data to disk ... ok postgres | postgres | postgres | Success. You can now start the database server using: postgres | postgres | pg_ctl -D /var/lib/postgresql/data -l logfile start postgres | postgres | initdb: warning: enabling "trust" authentication for local connections postgres | initdb: hint: You can change this by editing pg_hba.conf or using the option -A, or --auth-local and --auth-host, the next time you run initdb. postgres | waiting for server to start....2025-06-18 15:20:22.395 UTC [47] LOG: starting PostgreSQL 16.4 (Debian 16.4-1.pgdg120+2) on x86_64-pc-linux-gnu, compiled by gcc (Debian 12.2.0-14) 12.2.0, 64-bit postgres | 2025-06-18 15:20:22.409 UTC [47] LOG: listening on Unix socket "/var/run/postgresql/.s.PGSQL.5432" postgres | 2025-06-18 15:20:22.425 UTC [50] LOG: database system was shut down at 2025-06-18 15:20:21 UTC postgres | 2025-06-18 15:20:22.432 UTC [47] LOG: database system is ready to accept connections postgres | done postgres | server started postgres | postgres | /usr/local/bin/docker-entrypoint.sh: ignoring /docker-entrypoint-initdb.d/db-pg.conf postgres | postgres | /usr/local/bin/docker-entrypoint.sh: running /docker-entrypoint-initdb.d/db-pg.sh postgres | #!/bin/bash -xv postgres | # Copyright (C) 2022, 2024 Nordix Foundation. All rights reserved postgres | # postgres | # Licensed under the Apache License, Version 2.0 (the "License"); postgres | # you may not use this file except in compliance with the License. postgres | # You may obtain a copy of the License at postgres | # postgres | # http://www.apache.org/licenses/LICENSE-2.0 postgres | # postgres | # Unless required by applicable law or agreed to in writing, software postgres | # distributed under the License is distributed on an "AS IS" BASIS, postgres | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. postgres | # See the License for the specific language governing permissions and postgres | # limitations under the License. postgres | postgres | psql -U postgres -d postgres --command "CREATE USER ${PGSQL_USER} WITH PASSWORD '${PGSQL_PASSWORD}';" postgres | + psql -U postgres -d postgres --command 'CREATE USER policy_user WITH PASSWORD '\''policy_user'\'';' postgres | CREATE ROLE postgres | postgres | for db in migration pooling policyadmin policyclamp operationshistory clampacm postgres | do postgres | psql -U postgres -d postgres --command "CREATE DATABASE ${db};" postgres | psql -U postgres -d postgres --command "ALTER DATABASE ${db} OWNER TO ${PGSQL_USER} ;" postgres | psql -U postgres -d postgres --command "GRANT ALL PRIVILEGES ON DATABASE ${db} TO ${PGSQL_USER} ;" postgres | done postgres | + for db in migration pooling policyadmin policyclamp operationshistory clampacm postgres | + psql -U postgres -d postgres --command 'CREATE DATABASE migration;' postgres | CREATE DATABASE postgres | + psql -U postgres -d postgres --command 'ALTER DATABASE migration OWNER TO policy_user ;' postgres | ALTER DATABASE postgres | + psql -U postgres -d postgres --command 'GRANT ALL PRIVILEGES ON DATABASE migration TO policy_user ;' postgres | GRANT postgres | + for db in migration pooling policyadmin policyclamp operationshistory clampacm postgres | + psql -U postgres -d postgres --command 'CREATE DATABASE pooling;' postgres | CREATE DATABASE postgres | + psql -U postgres -d postgres --command 'ALTER DATABASE pooling OWNER TO policy_user ;' postgres | ALTER DATABASE postgres | + psql -U postgres -d postgres --command 'GRANT ALL PRIVILEGES ON DATABASE pooling TO policy_user ;' postgres | GRANT postgres | + for db in migration pooling policyadmin policyclamp operationshistory clampacm postgres | + psql -U postgres -d postgres --command 'CREATE DATABASE policyadmin;' postgres | CREATE DATABASE postgres | + psql -U postgres -d postgres --command 'ALTER DATABASE policyadmin OWNER TO policy_user ;' postgres | ALTER DATABASE postgres | + psql -U postgres -d postgres --command 'GRANT ALL PRIVILEGES ON DATABASE policyadmin TO policy_user ;' postgres | GRANT postgres | + for db in migration pooling policyadmin policyclamp operationshistory clampacm postgres | + psql -U postgres -d postgres --command 'CREATE DATABASE policyclamp;' postgres | CREATE DATABASE postgres | + psql -U postgres -d postgres --command 'ALTER DATABASE policyclamp OWNER TO policy_user ;' postgres | ALTER DATABASE postgres | + psql -U postgres -d postgres --command 'GRANT ALL PRIVILEGES ON DATABASE policyclamp TO policy_user ;' postgres | GRANT postgres | + for db in migration pooling policyadmin policyclamp operationshistory clampacm postgres | + psql -U postgres -d postgres --command 'CREATE DATABASE operationshistory;' postgres | CREATE DATABASE postgres | + psql -U postgres -d postgres --command 'ALTER DATABASE operationshistory OWNER TO policy_user ;' postgres | ALTER DATABASE postgres | + psql -U postgres -d postgres --command 'GRANT ALL PRIVILEGES ON DATABASE operationshistory TO policy_user ;' postgres | GRANT postgres | + for db in migration pooling policyadmin policyclamp operationshistory clampacm postgres | + psql -U postgres -d postgres --command 'CREATE DATABASE clampacm;' postgres | CREATE DATABASE postgres | + psql -U postgres -d postgres --command 'ALTER DATABASE clampacm OWNER TO policy_user ;' postgres | ALTER DATABASE postgres | + psql -U postgres -d postgres --command 'GRANT ALL PRIVILEGES ON DATABASE clampacm TO policy_user ;' postgres | GRANT postgres | postgres | waiting for server to shut down....2025-06-18 15:20:24.064 UTC [47] LOG: received fast shutdown request postgres | 2025-06-18 15:20:24.071 UTC [47] LOG: aborting any active transactions postgres | 2025-06-18 15:20:24.075 UTC [47] LOG: background worker "logical replication launcher" (PID 53) exited with exit code 1 postgres | 2025-06-18 15:20:24.076 UTC [48] LOG: shutting down postgres | 2025-06-18 15:20:24.079 UTC [48] LOG: checkpoint starting: shutdown immediate postgres | .2025-06-18 15:20:25.187 UTC [48] LOG: checkpoint complete: wrote 5511 buffers (33.6%); 0 WAL file(s) added, 0 removed, 1 recycled; write=0.940 s, sync=0.143 s, total=1.111 s; sync files=1788, longest=0.022 s, average=0.001 s; distance=25535 kB, estimate=25535 kB; lsn=0/2DDA218, redo lsn=0/2DDA218 postgres | 2025-06-18 15:20:25.200 UTC [47] LOG: database system is shut down postgres | done postgres | server stopped postgres | postgres | PostgreSQL init process complete; ready for start up. postgres | postgres | 2025-06-18 15:20:25.317 UTC [1] LOG: starting PostgreSQL 16.4 (Debian 16.4-1.pgdg120+2) on x86_64-pc-linux-gnu, compiled by gcc (Debian 12.2.0-14) 12.2.0, 64-bit postgres | 2025-06-18 15:20:25.317 UTC [1] LOG: listening on IPv4 address "0.0.0.0", port 5432 postgres | 2025-06-18 15:20:25.317 UTC [1] LOG: listening on IPv6 address "::", port 5432 postgres | 2025-06-18 15:20:25.328 UTC [1] LOG: listening on Unix socket "/var/run/postgresql/.s.PGSQL.5432" postgres | 2025-06-18 15:20:25.338 UTC [100] LOG: database system was shut down at 2025-06-18 15:20:25 UTC postgres | 2025-06-18 15:20:25.351 UTC [1] LOG: database system is ready to accept connections prometheus | time=2025-06-18T15:20:23.679Z level=INFO source=main.go:674 msg="No time or size retention was set so using the default time retention" duration=15d prometheus | time=2025-06-18T15:20:23.679Z level=INFO source=main.go:725 msg="Starting Prometheus Server" mode=server version="(version=3.4.1, branch=HEAD, revision=aea6503d9bbaad6c5faff3ecf6f1025213356c92)" prometheus | time=2025-06-18T15:20:23.679Z level=INFO source=main.go:730 msg="operational information" build_context="(go=go1.24.3, platform=linux/amd64, user=root@16f976c24db1, date=20250531-10:44:38, tags=netgo,builtinassets,stringlabels)" host_details="(Linux 4.15.0-192-generic #203-Ubuntu SMP Wed Aug 10 17:40:03 UTC 2022 x86_64 prometheus (none))" fd_limits="(soft=1048576, hard=1048576)" vm_limits="(soft=unlimited, hard=unlimited)" prometheus | time=2025-06-18T15:20:23.680Z level=INFO source=main.go:806 msg="Leaving GOMAXPROCS=8: CPU quota undefined" component=automaxprocs prometheus | time=2025-06-18T15:20:23.682Z level=INFO source=web.go:656 msg="Start listening for connections" component=web address=0.0.0.0:9090 prometheus | time=2025-06-18T15:20:23.682Z level=INFO source=main.go:1266 msg="Starting TSDB ..." prometheus | time=2025-06-18T15:20:23.691Z level=INFO source=tls_config.go:347 msg="Listening on" component=web address=[::]:9090 prometheus | time=2025-06-18T15:20:23.691Z level=INFO source=tls_config.go:350 msg="TLS is disabled." component=web http2=false address=[::]:9090 prometheus | time=2025-06-18T15:20:23.692Z level=INFO source=head.go:657 msg="Replaying on-disk memory mappable chunks if any" component=tsdb prometheus | time=2025-06-18T15:20:23.692Z level=INFO source=head.go:744 msg="On-disk memory mappable chunks replay completed" component=tsdb duration=2.47µs prometheus | time=2025-06-18T15:20:23.694Z level=INFO source=head.go:752 msg="Replaying WAL, this may take a while" component=tsdb prometheus | time=2025-06-18T15:20:23.694Z level=INFO source=head.go:825 msg="WAL segment loaded" component=tsdb segment=0 maxSegment=0 duration=307.243µs prometheus | time=2025-06-18T15:20:23.694Z level=INFO source=head.go:862 msg="WAL replay completed" component=tsdb checkpoint_replay_duration=58.83µs wal_replay_duration=338.163µs wbl_replay_duration=200ns chunk_snapshot_load_duration=0s mmap_chunk_replay_duration=2.47µs total_replay_duration=2.576324ms prometheus | time=2025-06-18T15:20:23.699Z level=INFO source=main.go:1287 msg="filesystem information" fs_type=EXT4_SUPER_MAGIC prometheus | time=2025-06-18T15:20:23.699Z level=INFO source=main.go:1290 msg="TSDB started" prometheus | time=2025-06-18T15:20:23.699Z level=INFO source=main.go:1475 msg="Loading configuration file" filename=/etc/prometheus/prometheus.yml prometheus | time=2025-06-18T15:20:23.700Z level=INFO source=main.go:1514 msg="updated GOGC" old=100 new=75 prometheus | time=2025-06-18T15:20:23.700Z level=INFO source=main.go:1524 msg="Completed loading of configuration file" db_storage=1.21µs remote_storage=2.05µs web_handler=590ns query_engine=1.27µs scrape=259.952µs scrape_sd=144.522µs notify=120.511µs notify_sd=14.31µs rules=2.68µs tracing=4.49µs filename=/etc/prometheus/prometheus.yml totalDuration=1.211721ms prometheus | time=2025-06-18T15:20:23.700Z level=INFO source=main.go:1251 msg="Server is ready to receive web requests." prometheus | time=2025-06-18T15:20:23.700Z level=INFO source=manager.go:175 msg="Starting rule manager..." component="rule manager" zookeeper | ===> User zookeeper | uid=1000(appuser) gid=1000(appuser) groups=1000(appuser) zookeeper | ===> Configuring ... zookeeper | ===> Running preflight checks ... zookeeper | ===> Check if /var/lib/zookeeper/data is writable ... zookeeper | ===> Check if /var/lib/zookeeper/log is writable ... zookeeper | ===> Launching ... zookeeper | ===> Launching zookeeper ... zookeeper | [2025-06-18 15:20:25,703] INFO Reading configuration from: /etc/kafka/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,705] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,705] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,705] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,706] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,707] INFO autopurge.snapRetainCount set to 3 (org.apache.zookeeper.server.DatadirCleanupManager) zookeeper | [2025-06-18 15:20:25,707] INFO autopurge.purgeInterval set to 0 (org.apache.zookeeper.server.DatadirCleanupManager) zookeeper | [2025-06-18 15:20:25,707] INFO Purge task is not scheduled. (org.apache.zookeeper.server.DatadirCleanupManager) zookeeper | [2025-06-18 15:20:25,707] WARN Either no config or no quorum defined in config, running in standalone mode (org.apache.zookeeper.server.quorum.QuorumPeerMain) zookeeper | [2025-06-18 15:20:25,708] INFO Log4j 1.2 jmx support not found; jmx disabled. (org.apache.zookeeper.jmx.ManagedUtil) zookeeper | [2025-06-18 15:20:25,708] INFO Reading configuration from: /etc/kafka/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,709] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,709] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,709] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,709] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2025-06-18 15:20:25,709] INFO Starting server (org.apache.zookeeper.server.ZooKeeperServerMain) zookeeper | [2025-06-18 15:20:25,720] INFO ServerMetrics initialized with provider org.apache.zookeeper.metrics.impl.DefaultMetricsProvider@3bbc39f8 (org.apache.zookeeper.server.ServerMetrics) zookeeper | [2025-06-18 15:20:25,723] INFO ACL digest algorithm is: SHA1 (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) zookeeper | [2025-06-18 15:20:25,723] INFO zookeeper.DigestAuthenticationProvider.enabled = true (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) zookeeper | [2025-06-18 15:20:25,725] INFO zookeeper.snapshot.trust.empty : false (org.apache.zookeeper.server.persistence.FileTxnSnapLog) zookeeper | [2025-06-18 15:20:25,733] INFO (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,733] INFO ______ _ (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,733] INFO |___ / | | (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,733] INFO / / ___ ___ | | __ ___ ___ _ __ ___ _ __ (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,733] INFO / / / _ \ / _ \ | |/ / / _ \ / _ \ | '_ \ / _ \ | '__| (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,733] INFO / /__ | (_) | | (_) | | < | __/ | __/ | |_) | | __/ | | (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,733] INFO /_____| \___/ \___/ |_|\_\ \___| \___| | .__/ \___| |_| (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,733] INFO | | (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,733] INFO |_| (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,733] INFO (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,735] INFO Server environment:zookeeper.version=3.8.4-9316c2a7a97e1666d8f4593f34dd6fc36ecc436c, built on 2024-02-12 22:16 UTC (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,735] INFO Server environment:host.name=zookeeper (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,735] INFO Server environment:java.version=17.0.14 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,735] INFO Server environment:java.vendor=Eclipse Adoptium (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,735] INFO Server environment:java.home=/usr/lib/jvm/temurin-17-jre (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,735] INFO Server environment:java.class.path=/usr/bin/../share/java/kafka/kafka-streams-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/jersey-common-2.39.1.jar:/usr/bin/../share/java/kafka/swagger-annotations-2.2.8.jar:/usr/bin/../share/java/kafka/kafka_2.13-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/commons-validator-1.7.jar:/usr/bin/../share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/bin/../share/java/kafka/aopalliance-repackaged-2.6.1.jar:/usr/bin/../share/java/kafka/kafka-transaction-coordinator-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/connect-transforms-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/netty-transport-4.1.118.Final.jar:/usr/bin/../share/java/kafka/kafka-clients-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/rocksdbjni-7.9.2.jar:/usr/bin/../share/java/kafka/javax.activation-api-1.2.0.jar:/usr/bin/../share/java/kafka/jetty-util-ajax-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/kafka-metadata-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/commons-cli-1.4.jar:/usr/bin/../share/java/kafka/connect-mirror-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/slf4j-reload4j-1.7.36.jar:/usr/bin/../share/java/kafka/kafka-streams-scala_2.13-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-datatype-jdk8-2.16.2.jar:/usr/bin/../share/java/kafka/scala-library-2.13.15.jar:/usr/bin/../share/java/kafka/jakarta.ws.rs-api-2.1.6.jar:/usr/bin/../share/java/kafka/jetty-servlet-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/jakarta.annotation-api-1.3.5.jar:/usr/bin/../share/java/kafka/netty-resolver-4.1.118.Final.jar:/usr/bin/../share/java/kafka/scala-java8-compat_2.13-1.0.2.jar:/usr/bin/../share/java/kafka/kafka-tools-api-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/javax.ws.rs-api-2.1.1.jar:/usr/bin/../share/java/kafka/zookeeper-jute-3.8.4.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-base-2.16.2.jar:/usr/bin/../share/java/kafka/connect-runtime-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/hk2-api-2.6.1.jar:/usr/bin/../share/java/kafka/jackson-module-afterburner-2.16.2.jar:/usr/bin/../share/java/kafka/kafka-streams-test-utils-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/kafka.jar:/usr/bin/../share/java/kafka/protobuf-java-3.25.5.jar:/usr/bin/../share/java/kafka/jackson-dataformat-csv-2.16.2.jar:/usr/bin/../share/java/kafka/kafka-server-common-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.inject-2.6.1.jar:/usr/bin/../share/java/kafka/maven-artifact-3.9.6.jar:/usr/bin/../share/java/kafka/jakarta.xml.bind-api-2.3.3.jar:/usr/bin/../share/java/kafka/jose4j-0.9.4.jar:/usr/bin/../share/java/kafka/hk2-locator-2.6.1.jar:/usr/bin/../share/java/kafka/reflections-0.10.2.jar:/usr/bin/../share/java/kafka/slf4j-api-1.7.36.jar:/usr/bin/../share/java/kafka/jetty-continuation-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/paranamer-2.8.jar:/usr/bin/../share/java/kafka/commons-beanutils-1.9.4.jar:/usr/bin/../share/java/kafka/jaxb-api-2.3.1.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-2.39.1.jar:/usr/bin/../share/java/kafka/hk2-utils-2.6.1.jar:/usr/bin/../share/java/kafka/trogdor-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/scala-logging_2.13-3.9.5.jar:/usr/bin/../share/java/kafka/reload4j-1.2.25.jar:/usr/bin/../share/java/kafka/jersey-hk2-2.39.1.jar:/usr/bin/../share/java/kafka/kafka-server-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-shell-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/jersey-client-2.39.1.jar:/usr/bin/../share/java/kafka/osgi-resource-locator-1.0.3.jar:/usr/bin/../share/java/kafka/scala-reflect-2.13.15.jar:/usr/bin/../share/java/kafka/jetty-util-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/netty-common-4.1.118.Final.jar:/usr/bin/../share/java/kafka/kafka-group-coordinator-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/commons-digester-2.1.jar:/usr/bin/../share/java/kafka/argparse4j-0.7.0.jar:/usr/bin/../share/java/kafka/commons-lang3-3.12.0.jar:/usr/bin/../share/java/kafka/kafka-storage-api-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/audience-annotations-0.12.0.jar:/usr/bin/../share/java/kafka/connect-mirror-client-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/javax.annotation-api-1.3.2.jar:/usr/bin/../share/java/kafka/netty-buffer-4.1.118.Final.jar:/usr/bin/../share/java/kafka/zstd-jni-1.5.6-4.jar:/usr/bin/../share/java/kafka/jakarta.validation-api-2.0.2.jar:/usr/bin/../share/java/kafka/jetty-client-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/zookeeper-3.8.4.jar:/usr/bin/../share/java/kafka/jersey-server-2.39.1.jar:/usr/bin/../share/java/kafka/connect-basic-auth-extension-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-servlets-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/netty-transport-native-unix-common-4.1.118.Final.jar:/usr/bin/../share/java/kafka/jopt-simple-5.0.4.jar:/usr/bin/../share/java/kafka/netty-transport-native-epoll-4.1.118.Final.jar:/usr/bin/../share/java/kafka/error_prone_annotations-2.10.0.jar:/usr/bin/../share/java/kafka/lz4-java-1.8.0.jar:/usr/bin/../share/java/kafka/kafka-group-coordinator-api-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.activation-api-1.2.2.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-core-2.39.1.jar:/usr/bin/../share/java/kafka/kafka-tools-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-module-jaxb-annotations-2.16.2.jar:/usr/bin/../share/java/kafka/jetty-server-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/netty-transport-classes-epoll-4.1.118.Final.jar:/usr/bin/../share/java/kafka/jackson-annotations-2.16.2.jar:/usr/bin/../share/java/kafka/jackson-databind-2.16.2.jar:/usr/bin/../share/java/kafka/pcollections-4.0.1.jar:/usr/bin/../share/java/kafka/opentelemetry-proto-1.0.0-alpha.jar:/usr/bin/../share/java/kafka/jetty-security-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/connect-json-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-json-provider-2.16.2.jar:/usr/bin/../share/java/kafka/commons-logging-1.2.jar:/usr/bin/../share/java/kafka/jsr305-3.0.2.jar:/usr/bin/../share/java/kafka/kafka-raft-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/plexus-utils-3.5.1.jar:/usr/bin/../share/java/kafka/jetty-http-9.4.57.v20241219.jar:/usr/bin/../share/java/kafka/connect-api-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/scala-collection-compat_2.13-2.10.0.jar:/usr/bin/../share/java/kafka/metrics-core-2.2.0.jar:/usr/bin/../share/java/kafka/kafka-streams-examples-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/commons-collections-3.2.2.jar:/usr/bin/../share/java/kafka/javassist-3.29.2-GA.jar:/usr/bin/../share/java/kafka/caffeine-2.9.3.jar:/usr/bin/../share/java/kafka/commons-io-2.14.0.jar:/usr/bin/../share/java/kafka/jackson-module-scala_2.13-2.16.2.jar:/usr/bin/../share/java/kafka/activation-1.1.1.jar:/usr/bin/../share/java/kafka/jackson-core-2.16.2.jar:/usr/bin/../share/java/kafka/metrics-core-4.1.12.1.jar:/usr/bin/../share/java/kafka/jline-3.25.1.jar:/usr/bin/../share/java/kafka/netty-codec-4.1.118.Final.jar:/usr/bin/../share/java/kafka/snappy-java-1.1.10.5.jar:/usr/bin/../share/java/kafka/netty-handler-4.1.118.Final.jar:/usr/bin/../share/java/kafka/kafka-storage-7.9.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-io-9.4.57.v20241219.jar:/usr/bin/../share/java/confluent-telemetry/* (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:java.library.path=/usr/local/lib64:/usr/local/lib::/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:java.io.tmpdir=/tmp (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:java.compiler= (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:os.name=Linux (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:os.arch=amd64 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:os.version=4.15.0-192-generic (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:user.name=appuser (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:user.home=/home/appuser (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:user.dir=/home/appuser (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:os.memory.free=495MB (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:os.memory.max=512MB (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO Server environment:os.memory.total=512MB (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO zookeeper.enableEagerACLCheck = false (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO zookeeper.digest.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO zookeeper.closeSessionTxn.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO zookeeper.flushDelay = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO zookeeper.maxWriteQueuePollTime = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO zookeeper.maxBatchSize=1000 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,736] INFO zookeeper.intBufferStartingSizeBytes = 1024 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,737] INFO Weighed connection throttling is disabled (org.apache.zookeeper.server.BlueThrottle) zookeeper | [2025-06-18 15:20:25,738] INFO minSessionTimeout set to 6000 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,738] INFO maxSessionTimeout set to 60000 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,739] INFO getData response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) zookeeper | [2025-06-18 15:20:25,739] INFO getChildren response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) zookeeper | [2025-06-18 15:20:25,740] INFO zookeeper.pathStats.slotCapacity = 60 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2025-06-18 15:20:25,740] INFO zookeeper.pathStats.slotDuration = 15 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2025-06-18 15:20:25,740] INFO zookeeper.pathStats.maxDepth = 6 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2025-06-18 15:20:25,740] INFO zookeeper.pathStats.initialDelay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2025-06-18 15:20:25,740] INFO zookeeper.pathStats.delay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2025-06-18 15:20:25,740] INFO zookeeper.pathStats.enabled = false (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2025-06-18 15:20:25,742] INFO The max bytes for all large requests are set to 104857600 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,742] INFO The large request threshold is set to -1 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,743] INFO zookeeper.enforce.auth.enabled = false (org.apache.zookeeper.server.AuthenticationHelper) zookeeper | [2025-06-18 15:20:25,743] INFO zookeeper.enforce.auth.schemes = [] (org.apache.zookeeper.server.AuthenticationHelper) zookeeper | [2025-06-18 15:20:25,743] INFO Created server with tickTime 3000 ms minSessionTimeout 6000 ms maxSessionTimeout 60000 ms clientPortListenBacklog -1 datadir /var/lib/zookeeper/log/version-2 snapdir /var/lib/zookeeper/data/version-2 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,764] INFO Logging initialized @381ms to org.eclipse.jetty.util.log.Slf4jLog (org.eclipse.jetty.util.log) zookeeper | [2025-06-18 15:20:25,819] WARN o.e.j.s.ServletContextHandler@6150c3ec{/,null,STOPPED} contextPath ends with /* (org.eclipse.jetty.server.handler.ContextHandler) zookeeper | [2025-06-18 15:20:25,819] WARN Empty contextPath (org.eclipse.jetty.server.handler.ContextHandler) zookeeper | [2025-06-18 15:20:25,835] INFO jetty-9.4.57.v20241219; built: 2025-01-08T21:24:30.412Z; git: df524e6b29271c2e09ba9aea83c18dc9db464a31; jvm 17.0.14+7 (org.eclipse.jetty.server.Server) zookeeper | [2025-06-18 15:20:25,869] INFO DefaultSessionIdManager workerName=node0 (org.eclipse.jetty.server.session) zookeeper | [2025-06-18 15:20:25,869] INFO No SessionScavenger set, using defaults (org.eclipse.jetty.server.session) zookeeper | [2025-06-18 15:20:25,871] INFO node0 Scavenging every 660000ms (org.eclipse.jetty.server.session) zookeeper | [2025-06-18 15:20:25,875] WARN ServletContext@o.e.j.s.ServletContextHandler@6150c3ec{/,null,STARTING} has uncovered http methods for path: /* (org.eclipse.jetty.security.SecurityHandler) zookeeper | [2025-06-18 15:20:25,884] INFO Started o.e.j.s.ServletContextHandler@6150c3ec{/,null,AVAILABLE} (org.eclipse.jetty.server.handler.ContextHandler) zookeeper | [2025-06-18 15:20:25,893] INFO Started ServerConnector@222545dc{HTTP/1.1, (http/1.1)}{0.0.0.0:8080} (org.eclipse.jetty.server.AbstractConnector) zookeeper | [2025-06-18 15:20:25,893] INFO Started @515ms (org.eclipse.jetty.server.Server) zookeeper | [2025-06-18 15:20:25,893] INFO Started AdminServer on address 0.0.0.0, port 8080 and command URL /commands (org.apache.zookeeper.server.admin.JettyAdminServer) zookeeper | [2025-06-18 15:20:25,897] INFO Using org.apache.zookeeper.server.NIOServerCnxnFactory as server connection factory (org.apache.zookeeper.server.ServerCnxnFactory) zookeeper | [2025-06-18 15:20:25,898] WARN maxCnxns is not configured, using default value 0. (org.apache.zookeeper.server.ServerCnxnFactory) zookeeper | [2025-06-18 15:20:25,899] INFO Configuring NIO connection handler with 10s sessionless connection timeout, 2 selector thread(s), 16 worker threads, and 64 kB direct buffers. (org.apache.zookeeper.server.NIOServerCnxnFactory) zookeeper | [2025-06-18 15:20:25,900] INFO binding to port 0.0.0.0/0.0.0.0:2181 (org.apache.zookeeper.server.NIOServerCnxnFactory) zookeeper | [2025-06-18 15:20:25,910] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) zookeeper | [2025-06-18 15:20:25,910] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) zookeeper | [2025-06-18 15:20:25,910] INFO zookeeper.snapshotSizeFactor = 0.33 (org.apache.zookeeper.server.ZKDatabase) zookeeper | [2025-06-18 15:20:25,910] INFO zookeeper.commitLogCount=500 (org.apache.zookeeper.server.ZKDatabase) zookeeper | [2025-06-18 15:20:25,914] INFO zookeeper.snapshot.compression.method = CHECKED (org.apache.zookeeper.server.persistence.SnapStream) zookeeper | [2025-06-18 15:20:25,914] INFO Snapshotting: 0x0 to /var/lib/zookeeper/data/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) zookeeper | [2025-06-18 15:20:25,916] INFO Snapshot loaded in 6 ms, highest zxid is 0x0, digest is 1371985504 (org.apache.zookeeper.server.ZKDatabase) zookeeper | [2025-06-18 15:20:25,917] INFO Snapshotting: 0x0 to /var/lib/zookeeper/data/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) zookeeper | [2025-06-18 15:20:25,917] INFO Snapshot taken in 1 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2025-06-18 15:20:25,925] INFO PrepRequestProcessor (sid:0) started, reconfigEnabled=false (org.apache.zookeeper.server.PrepRequestProcessor) zookeeper | [2025-06-18 15:20:25,925] INFO zookeeper.request_throttler.shutdownTimeout = 10000 ms (org.apache.zookeeper.server.RequestThrottler) zookeeper | [2025-06-18 15:20:25,939] INFO Using checkIntervalMs=60000 maxPerMinute=10000 maxNeverUsedIntervalMs=0 (org.apache.zookeeper.server.ContainerManager) zookeeper | [2025-06-18 15:20:25,939] INFO ZooKeeper audit is disabled. (org.apache.zookeeper.audit.ZKAuditProvider) zookeeper | [2025-06-18 15:20:31,529] INFO Creating new log file: log.1 (org.apache.zookeeper.server.persistence.FileTxnLog) Tearing down containers... Container policy-csit Stopping Container grafana Stopping Container policy-csit Stopped Container policy-csit Removing Container policy-xacml-pdp Stopping Container policy-csit Removed Container grafana Stopped Container grafana Removing Container grafana Removed Container prometheus Stopping Container prometheus Stopped Container prometheus Removing Container prometheus Removed Container policy-xacml-pdp Stopped Container policy-xacml-pdp Removing Container policy-xacml-pdp Removed Container policy-pap Stopping Container policy-pap Stopped Container policy-pap Removing Container policy-pap Removed Container kafka Stopping Container policy-api Stopping Container kafka Stopped Container kafka Removing Container kafka Removed Container zookeeper Stopping Container zookeeper Stopped Container zookeeper Removing Container zookeeper Removed Container policy-api Stopped Container policy-api Removing Container policy-api Removed Container policy-db-migrator Stopping Container policy-db-migrator Stopped Container policy-db-migrator Removing Container policy-db-migrator Removed Container postgres Stopping Container postgres Stopped Container postgres Removing Container postgres Removed Network compose_default Removing Network compose_default Removed $ ssh-agent -k unset SSH_AUTH_SOCK; unset SSH_AGENT_PID; echo Agent pid 2127 killed; [ssh-agent] Stopped. Robot results publisher started... INFO: Checking test criticality is deprecated and will be dropped in a future release! -Parsing output xml: Done! -Copying log files to build dir: Done! -Assigning results to build: Done! -Checking thresholds: Done! Done publishing Robot results. [PostBuildScript] - [INFO] Executing post build scripts. [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash /tmp/jenkins8427822414138944462.sh ---> sysstat.sh [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash /tmp/jenkins16023692876844225382.sh ---> package-listing.sh ++ facter osfamily ++ tr '[:upper:]' '[:lower:]' + OS_FAMILY=debian + workspace=/w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp + START_PACKAGES=/tmp/packages_start.txt + END_PACKAGES=/tmp/packages_end.txt + DIFF_PACKAGES=/tmp/packages_diff.txt + PACKAGES=/tmp/packages_start.txt + '[' /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp ']' + PACKAGES=/tmp/packages_end.txt + case "${OS_FAMILY}" in + dpkg -l + grep '^ii' + '[' -f /tmp/packages_start.txt ']' + '[' -f /tmp/packages_end.txt ']' + diff /tmp/packages_start.txt /tmp/packages_end.txt + '[' /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp ']' + mkdir -p /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp/archives/ + cp -f /tmp/packages_diff.txt /tmp/packages_end.txt /tmp/packages_start.txt /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp/archives/ [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash /tmp/jenkins15431081007006614813.sh ---> capture-instance-metadata.sh Setup pyenv: system 3.8.13 3.9.13 * 3.10.6 (set by /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-WIvM from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-WIvM/bin to PATH INFO: Running in OpenStack, capturing instance metadata [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash /tmp/jenkins218350807282303228.sh provisioning config files... copy managed file [jenkins-log-archives-settings] to file:/w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp@tmp/config5402922209698961015tmp Regular expression run condition: Expression=[^.*logs-s3.*], Label=[] Run condition [Regular expression match] preventing perform for step [Provide Configuration files] [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content SERVER_ID=logs [EnvInject] - Variables injected successfully. [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash /tmp/jenkins11963911717720109977.sh ---> create-netrc.sh [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash /tmp/jenkins4420295657183146473.sh ---> python-tools-install.sh Setup pyenv: system 3.8.13 3.9.13 * 3.10.6 (set by /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-WIvM from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-WIvM/bin to PATH [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash /tmp/jenkins11131875840439049213.sh ---> sudo-logs.sh Archiving 'sudo' log.. [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash /tmp/jenkins5924612715698817179.sh ---> job-cost.sh Setup pyenv: system 3.8.13 3.9.13 * 3.10.6 (set by /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-WIvM from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: zipp==1.1.0 python-openstackclient urllib3~=1.26.15 lf-activate-venv(): INFO: Adding /tmp/venv-WIvM/bin to PATH INFO: No Stack... INFO: Retrieving Pricing Info for: v3-standard-8 INFO: Archiving Costs [policy-xacml-pdp-master-project-csit-verify-xacml-pdp] $ /bin/bash -l /tmp/jenkins2093511681561905391.sh ---> logs-deploy.sh Setup pyenv: system 3.8.13 3.9.13 * 3.10.6 (set by /w/workspace/policy-xacml-pdp-master-project-csit-verify-xacml-pdp/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-WIvM from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-WIvM/bin to PATH INFO: Nexus URL https://nexus.onap.org path production/vex-yul-ecomp-jenkins-1/policy-xacml-pdp-master-project-csit-verify-xacml-pdp/819 INFO: archiving workspace using pattern(s): -p **/target/surefire-reports/*-output.txt Archives upload complete. INFO: archiving logs to Nexus ---> uname -a: Linux prd-ubuntu1804-docker-8c-8g-22130 4.15.0-192-generic #203-Ubuntu SMP Wed Aug 10 17:40:03 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux ---> lscpu: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 8 On-line CPU(s) list: 0-7 Thread(s) per core: 1 Core(s) per socket: 1 Socket(s): 8 NUMA node(s): 1 Vendor ID: AuthenticAMD CPU family: 23 Model: 49 Model name: AMD EPYC-Rome Processor Stepping: 0 CPU MHz: 2799.998 BogoMIPS: 5599.99 Virtualization: AMD-V Hypervisor vendor: KVM Virtualization type: full L1d cache: 32K L1i cache: 32K L2 cache: 512K L3 cache: 16384K NUMA node0 CPU(s): 0-7 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm rep_good nopl xtopology cpuid extd_apicid tsc_known_freq pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext perfctr_core ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves clzero xsaveerptr arat npt nrip_save umip rdpid arch_capabilities ---> nproc: 8 ---> df -h: Filesystem Size Used Avail Use% Mounted on udev 16G 0 16G 0% /dev tmpfs 3.2G 708K 3.2G 1% /run /dev/vda1 155G 15G 141G 10% / tmpfs 16G 0 16G 0% /dev/shm tmpfs 5.0M 0 5.0M 0% /run/lock tmpfs 16G 0 16G 0% /sys/fs/cgroup /dev/vda15 105M 4.4M 100M 5% /boot/efi tmpfs 3.2G 0 3.2G 0% /run/user/1001 ---> free -m: total used free shared buff/cache available Mem: 32167 866 24299 0 7001 30845 Swap: 1023 0 1023 ---> ip addr: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: ens3: mtu 1458 qdisc mq state UP group default qlen 1000 link/ether fa:16:3e:03:83:28 brd ff:ff:ff:ff:ff:ff inet 10.30.107.237/23 brd 10.30.107.255 scope global dynamic ens3 valid_lft 85940sec preferred_lft 85940sec inet6 fe80::f816:3eff:fe03:8328/64 scope link valid_lft forever preferred_lft forever 3: docker0: mtu 1500 qdisc noqueue state DOWN group default link/ether 02:42:4c:bb:97:62 brd ff:ff:ff:ff:ff:ff inet 10.250.0.254/24 brd 10.250.0.255 scope global docker0 valid_lft forever preferred_lft forever inet6 fe80::42:4cff:febb:9762/64 scope link valid_lft forever preferred_lft forever ---> sar -b -r -n DEV: Linux 4.15.0-192-generic (prd-ubuntu1804-docker-8c-8g-22130) 06/18/25 _x86_64_ (8 CPU) 15:17:37 LINUX RESTART (8 CPU) 15:18:01 tps rtps wtps bread/s bwrtn/s 15:19:01 294.07 57.73 236.34 3648.78 62526.89 15:20:01 302.33 20.10 282.24 2297.48 134405.60 15:21:02 473.00 4.73 468.27 427.53 122246.56 15:22:01 162.16 0.20 161.96 26.44 27788.92 15:23:01 78.52 0.23 78.29 18.80 16727.21 15:24:01 19.18 0.12 19.06 12.66 376.60 15:25:01 69.04 1.82 67.22 94.12 2613.16 Average: 199.85 12.16 187.69 934.49 52442.49 15:18:01 kbmemfree kbavail kbmemused %memused kbbuffers kbcached kbcommit %commit kbactive kbinact kbdirty 15:19:01 30195136 31724396 2744084 8.33 67384 1774496 1362764 4.01 828200 1632132 141508 15:20:01 25540684 31657364 7398536 22.46 144172 6076556 1672836 4.92 997352 5851248 2085488 15:21:02 23789664 30204424 9149556 27.78 158612 6361896 7377428 21.71 2649472 5871664 1468 15:22:01 22695336 29664232 10243884 31.10 187472 6827484 8215368 24.17 3302580 6235148 117668 15:23:01 22644724 29601512 10294496 31.25 200548 6808604 8300836 24.42 3377880 6217888 784 15:24:01 23126760 30041048 9812460 29.79 200784 6772284 6563040 19.31 2962052 6172108 56 15:25:01 24832516 31533188 8106704 24.61 202888 6546220 1629768 4.80 1515120 5971248 11064 Average: 24689260 30632309 8249960 25.05 165980 5881077 5017434 14.76 2233237 5421634 336862 15:18:01 IFACE rxpck/s txpck/s rxkB/s txkB/s rxcmp/s txcmp/s rxmcst/s %ifutil 15:19:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 15:19:01 lo 1.67 1.67 0.19 0.19 0.00 0.00 0.00 0.00 15:19:01 ens3 604.28 396.67 1643.52 86.50 0.00 0.00 0.00 0.00 15:20:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 15:20:01 lo 13.40 13.40 1.22 1.22 0.00 0.00 0.00 0.00 15:20:01 br-af455255afc3 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 15:20:01 ens3 1440.41 854.51 32261.75 71.13 0.00 0.00 0.00 0.00 15:21:02 veth19c0382 0.15 0.43 0.01 0.02 0.00 0.00 0.00 0.00 15:21:02 veth5d0389c 91.78 91.53 15.60 18.61 0.00 0.00 0.00 0.00 15:21:02 vethc124893 3.08 2.88 0.34 0.33 0.00 0.00 0.00 0.00 15:21:02 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 15:22:01 veth19c0382 0.42 0.42 0.05 1.01 0.00 0.00 0.00 0.00 15:22:01 veth5d0389c 0.14 0.20 0.45 0.02 0.00 0.00 0.00 0.00 15:22:01 vethc124893 143.57 144.74 16.97 34.82 0.00 0.00 0.00 0.00 15:22:01 docker0 138.33 191.20 8.88 1371.76 0.00 0.00 0.00 0.00 15:23:01 veth19c0382 0.50 0.62 0.05 1.26 0.00 0.00 0.00 0.00 15:23:01 veth5d0389c 130.11 129.80 14.96 27.83 0.00 0.00 0.00 0.00 15:23:01 vethc124893 381.49 383.37 41.54 72.16 0.00 0.00 0.00 0.01 15:23:01 veth9a77e97 1.83 1.67 0.67 0.49 0.00 0.00 0.00 0.00 15:24:01 veth5d0389c 0.33 0.45 0.58 0.03 0.00 0.00 0.00 0.00 15:24:01 vethc124893 113.23 115.93 12.74 21.03 0.00 0.00 0.00 0.00 15:24:01 docker0 136.16 188.17 8.73 1348.91 0.00 0.00 0.00 0.00 15:24:01 veth5913464 892.05 916.75 179.97 107.09 0.00 0.00 0.00 0.01 15:25:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 15:25:01 lo 26.83 26.83 2.42 2.42 0.00 0.00 0.00 0.00 15:25:01 ens3 2437.14 1539.04 36567.04 213.43 0.00 0.00 0.00 0.00 Average: docker0 19.50 26.94 1.25 193.16 0.00 0.00 0.00 0.00 Average: lo 3.35 3.35 0.31 0.31 0.00 0.00 0.00 0.00 Average: ens3 346.26 218.33 5225.85 30.32 0.00 0.00 0.00 0.00 ---> sar -P ALL: Linux 4.15.0-192-generic (prd-ubuntu1804-docker-8c-8g-22130) 06/18/25 _x86_64_ (8 CPU) 15:17:37 LINUX RESTART (8 CPU) 15:18:01 CPU %user %nice %system %iowait %steal %idle 15:19:01 all 10.28 0.00 1.08 4.56 0.04 84.03 15:19:01 0 14.04 0.00 1.47 1.69 0.03 82.77 15:19:01 1 10.13 0.00 1.25 0.25 0.05 88.31 15:19:01 2 5.71 0.00 1.10 3.67 0.02 89.50 15:19:01 3 1.99 0.00 0.62 12.39 0.03 84.97 15:19:01 4 11.24 0.00 0.77 0.87 0.03 87.09 15:19:01 5 25.72 0.00 1.61 13.98 0.08 58.60 15:19:01 6 8.12 0.00 0.92 0.36 0.03 90.56 15:19:01 7 5.37 0.00 0.85 3.35 0.03 90.39 15:20:01 all 17.07 0.00 6.08 11.23 0.06 65.56 15:20:01 0 21.48 0.00 5.70 3.48 0.07 69.28 15:20:01 1 20.86 0.00 6.89 4.96 0.08 67.20 15:20:01 2 19.48 0.00 5.53 2.47 0.05 72.47 15:20:01 3 13.23 0.00 6.73 2.53 0.05 77.46 15:20:01 4 13.66 0.00 6.13 2.93 0.05 77.22 15:20:01 5 13.52 0.00 5.90 25.69 0.07 54.83 15:20:01 6 20.77 0.00 5.26 13.91 0.08 59.98 15:20:01 7 13.56 0.00 6.48 33.95 0.07 45.95 15:21:02 all 21.89 0.00 3.34 9.14 0.07 65.55 15:21:02 0 17.84 0.00 3.28 9.11 0.07 69.70 15:21:02 1 27.71 0.00 4.53 28.02 0.08 39.66 15:21:02 2 24.56 0.00 3.16 2.81 0.08 69.38 15:21:02 3 22.47 0.00 3.20 1.51 0.07 72.76 15:21:02 4 24.07 0.00 3.46 14.00 0.07 58.40 15:21:02 5 19.15 0.00 3.02 5.62 0.07 72.14 15:21:02 6 18.33 0.00 2.88 2.56 0.07 76.17 15:21:02 7 20.96 0.00 3.23 9.63 0.07 66.11 15:22:01 all 12.28 0.00 1.85 1.77 0.06 84.05 15:22:01 0 8.44 0.00 1.82 0.15 0.05 89.55 15:22:01 1 14.17 0.00 1.91 2.65 0.07 81.20 15:22:01 2 19.02 0.00 2.26 0.87 0.05 77.81 15:22:01 3 18.18 0.00 2.07 0.54 0.07 79.14 15:22:01 4 10.48 0.00 1.82 3.84 0.07 83.79 15:22:01 5 10.55 0.00 1.85 2.11 0.05 85.44 15:22:01 6 8.06 0.00 1.33 3.55 0.05 87.01 15:22:01 7 9.27 0.00 1.72 0.48 0.05 88.48 15:23:01 all 7.17 0.00 1.64 1.07 0.05 90.08 15:23:01 0 6.33 0.00 1.82 0.52 0.03 91.30 15:23:01 1 6.67 0.00 1.39 0.13 0.05 91.75 15:23:01 2 8.97 0.00 1.49 0.03 0.05 89.45 15:23:01 3 6.24 0.00 1.71 2.21 0.07 89.77 15:23:01 4 5.80 0.00 1.46 4.79 0.05 87.90 15:23:01 5 8.66 0.00 1.63 0.60 0.07 89.05 15:23:01 6 8.84 0.00 1.72 0.00 0.07 89.37 15:23:01 7 5.82 0.00 1.91 0.25 0.05 91.97 15:24:01 all 1.59 0.00 0.50 0.09 0.04 97.77 15:24:01 0 0.97 0.00 0.47 0.00 0.02 98.55 15:24:01 1 1.64 0.00 0.58 0.38 0.07 97.33 15:24:01 2 1.64 0.00 0.40 0.02 0.03 97.91 15:24:01 3 1.95 0.00 0.47 0.13 0.03 97.41 15:24:01 4 2.14 0.00 0.58 0.10 0.05 97.13 15:24:01 5 1.71 0.00 0.49 0.03 0.03 97.73 15:24:01 6 1.27 0.00 0.43 0.07 0.03 98.20 15:24:01 7 1.42 0.00 0.54 0.03 0.05 97.96 15:25:01 all 7.40 0.00 0.72 0.30 0.03 91.56 15:25:01 0 3.93 0.00 0.75 0.08 0.03 95.20 15:25:01 1 0.53 0.00 0.37 0.37 0.02 98.71 15:25:01 2 0.68 0.00 0.40 0.20 0.02 98.70 15:25:01 3 34.51 0.00 1.45 0.38 0.05 63.60 15:25:01 4 0.77 0.00 0.58 1.17 0.03 97.44 15:25:01 5 1.45 0.00 0.70 0.03 0.03 97.78 15:25:01 6 0.80 0.00 0.47 0.02 0.02 98.69 15:25:01 7 16.54 0.00 1.03 0.12 0.03 82.28 Average: all 11.08 0.00 2.17 4.02 0.05 82.68 Average: 0 10.42 0.00 2.18 2.15 0.04 85.21 Average: 1 11.64 0.00 2.41 5.23 0.06 80.65 Average: 2 11.41 0.00 2.04 1.44 0.04 85.07 Average: 3 14.07 0.00 2.32 2.82 0.05 80.74 Average: 4 9.72 0.00 2.11 3.95 0.05 84.17 Average: 5 11.53 0.00 2.17 6.86 0.06 79.38 Average: 6 9.46 0.00 1.86 2.92 0.05 85.71 Average: 7 10.42 0.00 2.24 6.80 0.05 80.49