09:39:17 Triggered by Gerrit: https://gerrit.onap.org/r/c/policy/docker/+/138370 09:39:17 Running as SYSTEM 09:39:17 [EnvInject] - Loading node environment variables. 09:39:17 Building remotely on prd-ubuntu1804-docker-8c-8g-21289 (ubuntu1804-docker-8c-8g) in workspace /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp 09:39:17 [ssh-agent] Looking for ssh-agent implementation... 09:39:17 [ssh-agent] Exec ssh-agent (binary ssh-agent on a remote machine) 09:39:17 $ ssh-agent 09:39:17 SSH_AUTH_SOCK=/tmp/ssh-ICM4a1N8fVMx/agent.2145 09:39:17 SSH_AGENT_PID=2147 09:39:17 [ssh-agent] Started. 09:39:17 Running ssh-add (command line suppressed) 09:39:17 Identity added: /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp@tmp/private_key_4766892636431158159.key (/w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp@tmp/private_key_4766892636431158159.key) 09:39:17 [ssh-agent] Using credentials onap-jobbuiler (Gerrit user) 09:39:17 The recommended git tool is: NONE 09:39:19 using credential onap-jenkins-ssh 09:39:19 Wiping out workspace first. 09:39:19 Cloning the remote Git repository 09:39:19 Cloning repository git://cloud.onap.org/mirror/policy/docker.git 09:39:19 > git init /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp # timeout=10 09:39:19 Fetching upstream changes from git://cloud.onap.org/mirror/policy/docker.git 09:39:19 > git --version # timeout=10 09:39:19 > git --version # 'git version 2.17.1' 09:39:19 using GIT_SSH to set credentials Gerrit user 09:39:19 Verifying host key using manually-configured host key entries 09:39:19 > git fetch --tags --progress -- git://cloud.onap.org/mirror/policy/docker.git +refs/heads/*:refs/remotes/origin/* # timeout=30 09:39:20 > git config remote.origin.url git://cloud.onap.org/mirror/policy/docker.git # timeout=10 09:39:20 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 09:39:20 > git config remote.origin.url git://cloud.onap.org/mirror/policy/docker.git # timeout=10 09:39:20 Fetching upstream changes from git://cloud.onap.org/mirror/policy/docker.git 09:39:20 using GIT_SSH to set credentials Gerrit user 09:39:20 Verifying host key using manually-configured host key entries 09:39:20 > git fetch --tags --progress -- git://cloud.onap.org/mirror/policy/docker.git refs/changes/70/138370/2 # timeout=30 09:39:20 > git rev-parse 98c473b1b99348ea19603eb6a0c4932cc295274d^{commit} # timeout=10 09:39:20 JENKINS-19022: warning: possible memory leak due to Git plugin usage; see: https://plugins.jenkins.io/git/#remove-git-plugin-buildsbybranch-builddata-script 09:39:20 Checking out Revision 98c473b1b99348ea19603eb6a0c4932cc295274d (refs/changes/70/138370/2) 09:39:20 > git config core.sparsecheckout # timeout=10 09:39:20 > git checkout -f 98c473b1b99348ea19603eb6a0c4932cc295274d # timeout=30 09:39:24 Commit message: "Fix helm plugin install failure" 09:39:24 > git rev-parse FETCH_HEAD^{commit} # timeout=10 09:39:24 > git rev-list --no-walk 54d234de0d9260f610425cd496a52265a4082441 # timeout=10 09:39:25 provisioning config files... 09:39:25 copy managed file [npmrc] to file:/home/jenkins/.npmrc 09:39:25 copy managed file [pipconf] to file:/home/jenkins/.config/pip/pip.conf 09:39:25 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash /tmp/jenkins17061869249085591111.sh 09:39:25 ---> python-tools-install.sh 09:39:25 Setup pyenv: 09:39:25 * system (set by /opt/pyenv/version) 09:39:25 * 3.8.13 (set by /opt/pyenv/version) 09:39:25 * 3.9.13 (set by /opt/pyenv/version) 09:39:25 * 3.10.6 (set by /opt/pyenv/version) 09:39:30 lf-activate-venv(): INFO: Creating python3 venv at /tmp/venv-O2Gi 09:39:30 lf-activate-venv(): INFO: Save venv in file: /tmp/.os_lf_venv 09:39:33 lf-activate-venv(): INFO: Installing: lftools 09:40:02 lf-activate-venv(): INFO: Adding /tmp/venv-O2Gi/bin to PATH 09:40:02 Generating Requirements File 09:40:22 Python 3.10.6 09:40:22 pip 24.1.1 from /tmp/venv-O2Gi/lib/python3.10/site-packages/pip (python 3.10) 09:40:23 appdirs==1.4.4 09:40:23 argcomplete==3.4.0 09:40:23 aspy.yaml==1.3.0 09:40:23 attrs==23.2.0 09:40:23 autopage==0.5.2 09:40:23 beautifulsoup4==4.12.3 09:40:23 boto3==1.34.139 09:40:23 botocore==1.34.139 09:40:23 bs4==0.0.2 09:40:23 cachetools==5.3.3 09:40:23 certifi==2024.7.4 09:40:23 cffi==1.16.0 09:40:23 cfgv==3.4.0 09:40:23 chardet==5.2.0 09:40:23 charset-normalizer==3.3.2 09:40:23 click==8.1.7 09:40:23 cliff==4.7.0 09:40:23 cmd2==2.4.3 09:40:23 cryptography==3.3.2 09:40:23 debtcollector==3.0.0 09:40:23 decorator==5.1.1 09:40:23 defusedxml==0.7.1 09:40:23 Deprecated==1.2.14 09:40:23 distlib==0.3.8 09:40:23 dnspython==2.6.1 09:40:23 docker==4.2.2 09:40:23 dogpile.cache==1.3.3 09:40:23 email_validator==2.2.0 09:40:23 filelock==3.15.4 09:40:23 future==1.0.0 09:40:23 gitdb==4.0.11 09:40:23 GitPython==3.1.43 09:40:23 google-auth==2.31.0 09:40:23 httplib2==0.22.0 09:40:23 identify==2.5.36 09:40:23 idna==3.7 09:40:23 importlib-resources==1.5.0 09:40:23 iso8601==2.1.0 09:40:23 Jinja2==3.1.4 09:40:23 jmespath==1.0.1 09:40:23 jsonpatch==1.33 09:40:23 jsonpointer==3.0.0 09:40:23 jsonschema==4.22.0 09:40:23 jsonschema-specifications==2023.12.1 09:40:23 keystoneauth1==5.6.0 09:40:23 kubernetes==30.1.0 09:40:23 lftools==0.37.10 09:40:23 lxml==5.2.2 09:40:23 MarkupSafe==2.1.5 09:40:23 msgpack==1.0.8 09:40:23 multi_key_dict==2.0.3 09:40:23 munch==4.0.0 09:40:23 netaddr==1.3.0 09:40:23 netifaces==0.11.0 09:40:23 niet==1.4.2 09:40:23 nodeenv==1.9.1 09:40:23 oauth2client==4.1.3 09:40:23 oauthlib==3.2.2 09:40:23 openstacksdk==3.2.0 09:40:23 os-client-config==2.1.0 09:40:23 os-service-types==1.7.0 09:40:23 osc-lib==3.0.1 09:40:23 oslo.config==9.4.0 09:40:23 oslo.context==5.5.0 09:40:23 oslo.i18n==6.3.0 09:40:23 oslo.log==6.0.0 09:40:23 oslo.serialization==5.4.0 09:40:23 oslo.utils==7.1.0 09:40:23 packaging==24.1 09:40:23 pbr==6.0.0 09:40:23 platformdirs==4.2.2 09:40:23 prettytable==3.10.0 09:40:23 pyasn1==0.6.0 09:40:23 pyasn1_modules==0.4.0 09:40:23 pycparser==2.22 09:40:23 pygerrit2==2.0.15 09:40:23 PyGithub==2.3.0 09:40:23 PyJWT==2.8.0 09:40:23 PyNaCl==1.5.0 09:40:23 pyparsing==2.4.7 09:40:23 pyperclip==1.9.0 09:40:23 pyrsistent==0.20.0 09:40:23 python-cinderclient==9.5.0 09:40:23 python-dateutil==2.9.0.post0 09:40:23 python-heatclient==3.5.0 09:40:23 python-jenkins==1.8.2 09:40:23 python-keystoneclient==5.4.0 09:40:23 python-magnumclient==4.5.0 09:40:23 python-novaclient==18.6.0 09:40:23 python-openstackclient==6.6.0 09:40:23 python-swiftclient==4.6.0 09:40:23 PyYAML==6.0.1 09:40:23 referencing==0.35.1 09:40:23 requests==2.32.3 09:40:23 requests-oauthlib==2.0.0 09:40:23 requestsexceptions==1.4.0 09:40:23 rfc3986==2.0.0 09:40:23 rpds-py==0.18.1 09:40:23 rsa==4.9 09:40:23 ruamel.yaml==0.18.6 09:40:23 ruamel.yaml.clib==0.2.8 09:40:23 s3transfer==0.10.2 09:40:23 simplejson==3.19.2 09:40:23 six==1.16.0 09:40:23 smmap==5.0.1 09:40:23 soupsieve==2.5 09:40:23 stevedore==5.2.0 09:40:23 tabulate==0.9.0 09:40:23 toml==0.10.2 09:40:23 tomlkit==0.12.5 09:40:23 tqdm==4.66.4 09:40:23 typing_extensions==4.12.2 09:40:23 tzdata==2024.1 09:40:23 urllib3==1.26.19 09:40:23 virtualenv==20.26.3 09:40:23 wcwidth==0.2.13 09:40:23 websocket-client==1.8.0 09:40:23 wrapt==1.16.0 09:40:23 xdg==6.0.0 09:40:23 xmltodict==0.13.0 09:40:23 yq==3.4.3 09:40:23 [EnvInject] - Injecting environment variables from a build step. 09:40:23 [EnvInject] - Injecting as environment variables the properties content 09:40:23 SET_JDK_VERSION=openjdk17 09:40:23 GIT_URL="git://cloud.onap.org/mirror" 09:40:23 09:40:23 [EnvInject] - Variables injected successfully. 09:40:23 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/sh /tmp/jenkins10779503130159208980.sh 09:40:23 ---> update-java-alternatives.sh 09:40:23 ---> Updating Java version 09:40:23 ---> Ubuntu/Debian system detected 09:40:23 update-alternatives: using /usr/lib/jvm/java-17-openjdk-amd64/bin/java to provide /usr/bin/java (java) in manual mode 09:40:23 update-alternatives: using /usr/lib/jvm/java-17-openjdk-amd64/bin/javac to provide /usr/bin/javac (javac) in manual mode 09:40:23 update-alternatives: using /usr/lib/jvm/java-17-openjdk-amd64 to provide /usr/lib/jvm/java-openjdk (java_sdk_openjdk) in manual mode 09:40:23 openjdk version "17.0.4" 2022-07-19 09:40:23 OpenJDK Runtime Environment (build 17.0.4+8-Ubuntu-118.04) 09:40:23 OpenJDK 64-Bit Server VM (build 17.0.4+8-Ubuntu-118.04, mixed mode, sharing) 09:40:23 JAVA_HOME=/usr/lib/jvm/java-17-openjdk-amd64 09:40:23 [EnvInject] - Injecting environment variables from a build step. 09:40:23 [EnvInject] - Injecting as environment variables the properties file path '/tmp/java.env' 09:40:23 [EnvInject] - Variables injected successfully. 09:40:23 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/sh -xe /tmp/jenkins8263590281927362503.sh 09:40:23 + /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp/csit/run-project-csit.sh apex-pdp 09:40:24 WARNING! Using --password via the CLI is insecure. Use --password-stdin. 09:40:24 WARNING! Your password will be stored unencrypted in /home/jenkins/.docker/config.json. 09:40:24 Configure a credential helper to remove this warning. See 09:40:24 https://docs.docker.com/engine/reference/commandline/login/#credentials-store 09:40:24 09:40:24 Login Succeeded 09:40:24 docker: 'compose' is not a docker command. 09:40:24 See 'docker --help' 09:40:24 Docker Compose Plugin not installed. Installing now... 09:40:24 % Total % Received % Xferd Average Speed Time Time Time Current 09:40:24 Dload Upload Total Spent Left Speed 09:40:24 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 09:40:24 0 60.0M 0 555k 0 0 2671k 0 0:00:23 --:--:-- 0:00:23 2671k 100 60.0M 100 60.0M 0 0 126M 0 --:--:-- --:--:-- --:--:-- 223M 09:40:24 Setting project configuration for: apex-pdp 09:40:24 Configuring docker compose... 09:40:26 Starting apex-pdp application with Grafana 09:40:26 time="2024-07-04T09:40:26Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:40:26 kafka Pulling 09:40:26 pap Pulling 09:40:26 grafana Pulling 09:40:26 prometheus Pulling 09:40:26 api Pulling 09:40:26 simulator Pulling 09:40:26 policy-db-migrator Pulling 09:40:26 apex-pdp Pulling 09:40:26 mariadb Pulling 09:40:26 zookeeper Pulling 09:40:27 31e352740f53 Pulling fs layer 09:40:27 21c7cf7066d0 Pulling fs layer 09:40:27 eb5e31f0ecf8 Pulling fs layer 09:40:27 4faab25371b2 Pulling fs layer 09:40:27 6b867d96d427 Pulling fs layer 09:40:27 93832cc54357 Pulling fs layer 09:40:27 4faab25371b2 Waiting 09:40:27 6b867d96d427 Waiting 09:40:27 93832cc54357 Waiting 09:40:27 31e352740f53 Downloading [> ] 48.06kB/3.398MB 09:40:27 eb5e31f0ecf8 Downloading [==================================================>] 305B/305B 09:40:27 eb5e31f0ecf8 Verifying Checksum 09:40:27 eb5e31f0ecf8 Download complete 09:40:27 31e352740f53 Pulling fs layer 09:40:27 e8bf24a82546 Pulling fs layer 09:40:27 154b803e2d93 Pulling fs layer 09:40:27 e4305231c991 Pulling fs layer 09:40:27 f469048fbe8d Pulling fs layer 09:40:27 c189e028fabb Pulling fs layer 09:40:27 31e352740f53 Downloading [> ] 48.06kB/3.398MB 09:40:27 c9bd119720e4 Pulling fs layer 09:40:27 e8bf24a82546 Waiting 09:40:27 e4305231c991 Waiting 09:40:27 c189e028fabb Waiting 09:40:27 c9bd119720e4 Waiting 09:40:27 31e352740f53 Pulling fs layer 09:40:27 21c7cf7066d0 Pulling fs layer 09:40:27 c3cc5e3d19ac Pulling fs layer 09:40:27 31e352740f53 Downloading [> ] 48.06kB/3.398MB 09:40:27 0d2280d71230 Pulling fs layer 09:40:27 984932e12fb0 Pulling fs layer 09:40:27 5687ac571232 Pulling fs layer 09:40:27 deac262509a5 Pulling fs layer 09:40:27 0d2280d71230 Waiting 09:40:27 984932e12fb0 Waiting 09:40:27 5687ac571232 Waiting 09:40:27 21c7cf7066d0 Downloading [> ] 539.6kB/73.93MB 09:40:27 21c7cf7066d0 Downloading [> ] 539.6kB/73.93MB 09:40:27 31e352740f53 Pulling fs layer 09:40:27 57703e441b07 Pulling fs layer 09:40:27 7138254c3790 Pulling fs layer 09:40:27 31e352740f53 Downloading [> ] 48.06kB/3.398MB 09:40:27 78f39bed0e83 Pulling fs layer 09:40:27 40796999d308 Pulling fs layer 09:40:27 14ddc757aae0 Pulling fs layer 09:40:27 57703e441b07 Waiting 09:40:27 ebe1cd824584 Pulling fs layer 09:40:27 d2893dc6732f Pulling fs layer 09:40:27 78f39bed0e83 Waiting 09:40:27 a23a963fcebe Pulling fs layer 09:40:27 7138254c3790 Waiting 09:40:27 40796999d308 Waiting 09:40:27 369dfa39565e Pulling fs layer 09:40:27 14ddc757aae0 Waiting 09:40:27 9146eb587aa8 Pulling fs layer 09:40:27 ebe1cd824584 Waiting 09:40:27 a120f6888c1f Pulling fs layer 09:40:27 d2893dc6732f Waiting 09:40:27 a23a963fcebe Waiting 09:40:27 9146eb587aa8 Waiting 09:40:27 369dfa39565e Waiting 09:40:27 a120f6888c1f Waiting 09:40:27 31e352740f53 Pulling fs layer 09:40:27 257d54e26411 Pulling fs layer 09:40:27 215302b53935 Pulling fs layer 09:40:27 eb2f448c7730 Pulling fs layer 09:40:27 c8ee90c58894 Pulling fs layer 09:40:27 e30cdb86c4f0 Pulling fs layer 09:40:27 c990b7e46fc8 Pulling fs layer 09:40:27 eb2f448c7730 Waiting 09:40:27 c8ee90c58894 Waiting 09:40:27 e30cdb86c4f0 Waiting 09:40:27 31e352740f53 Downloading [> ] 48.06kB/3.398MB 09:40:27 c990b7e46fc8 Waiting 09:40:27 257d54e26411 Waiting 09:40:27 215302b53935 Waiting 09:40:27 10ac4908093d Pulling fs layer 09:40:27 44779101e748 Pulling fs layer 09:40:27 a721db3e3f3d Pulling fs layer 09:40:27 1850a929b84a Pulling fs layer 09:40:27 397a918c7da3 Pulling fs layer 09:40:27 10ac4908093d Waiting 09:40:27 44779101e748 Waiting 09:40:27 a721db3e3f3d Waiting 09:40:27 1850a929b84a Waiting 09:40:27 806be17e856d Pulling fs layer 09:40:27 397a918c7da3 Waiting 09:40:27 634de6c90876 Pulling fs layer 09:40:27 806be17e856d Waiting 09:40:27 cd00854cfb1a Pulling fs layer 09:40:27 634de6c90876 Waiting 09:40:27 4faab25371b2 Downloading [> ] 539.6kB/158.6MB 09:40:27 9fa9226be034 Pulling fs layer 09:40:27 1617e25568b2 Pulling fs layer 09:40:27 3ecda1bfd07b Pulling fs layer 09:40:27 ac9f4de4b762 Pulling fs layer 09:40:27 ea63b2e6315f Pulling fs layer 09:40:27 fbd390d3bd00 Pulling fs layer 09:40:27 9b1ac15ef728 Pulling fs layer 09:40:27 8682f304eb80 Pulling fs layer 09:40:27 9fa9226be034 Waiting 09:40:27 5fbafe078afc Pulling fs layer 09:40:27 7fb53fd2ae10 Pulling fs layer 09:40:27 3ecda1bfd07b Waiting 09:40:27 592798bd3683 Pulling fs layer 09:40:27 ac9f4de4b762 Waiting 09:40:27 473fdc983780 Pulling fs layer 09:40:27 ea63b2e6315f Waiting 09:40:27 fbd390d3bd00 Waiting 09:40:27 473fdc983780 Waiting 09:40:27 592798bd3683 Waiting 09:40:27 7fb53fd2ae10 Waiting 09:40:27 8682f304eb80 Waiting 09:40:27 9b1ac15ef728 Waiting 09:40:27 1617e25568b2 Waiting 09:40:27 4abcf2066143 Pulling fs layer 09:40:27 c0e05c86127e Pulling fs layer 09:40:27 706651a94df6 Pulling fs layer 09:40:27 33e0a01314cc Pulling fs layer 09:40:27 f8b444c6ff40 Pulling fs layer 09:40:27 e6c38e6d3add Pulling fs layer 09:40:27 706651a94df6 Waiting 09:40:27 6ca01427385e Pulling fs layer 09:40:27 e35e8e85e24d Pulling fs layer 09:40:27 d0bef95bc6b2 Pulling fs layer 09:40:27 af860903a445 Pulling fs layer 09:40:27 4abcf2066143 Waiting 09:40:27 f8b444c6ff40 Waiting 09:40:27 e35e8e85e24d Waiting 09:40:27 c0e05c86127e Waiting 09:40:27 e6c38e6d3add Waiting 09:40:27 d0bef95bc6b2 Waiting 09:40:27 af860903a445 Waiting 09:40:27 6ca01427385e Waiting 09:40:27 31e352740f53 Verifying Checksum 09:40:27 31e352740f53 Verifying Checksum 09:40:27 31e352740f53 Verifying Checksum 09:40:27 31e352740f53 Download complete 09:40:27 31e352740f53 Download complete 09:40:27 31e352740f53 Download complete 09:40:27 31e352740f53 Download complete 09:40:27 31e352740f53 Download complete 09:40:27 31e352740f53 Extracting [> ] 65.54kB/3.398MB 09:40:27 31e352740f53 Extracting [> ] 65.54kB/3.398MB 09:40:27 31e352740f53 Extracting [> ] 65.54kB/3.398MB 09:40:27 31e352740f53 Extracting [> ] 65.54kB/3.398MB 09:40:27 31e352740f53 Extracting [> ] 65.54kB/3.398MB 09:40:27 6b867d96d427 Downloading [==================================================>] 1.153kB/1.153kB 09:40:27 6b867d96d427 Download complete 09:40:27 93832cc54357 Downloading [==================================================>] 1.127kB/1.127kB 09:40:27 93832cc54357 Verifying Checksum 09:40:27 93832cc54357 Download complete 09:40:27 21c7cf7066d0 Downloading [======> ] 10.27MB/73.93MB 09:40:27 21c7cf7066d0 Downloading [======> ] 10.27MB/73.93MB 09:40:27 e8bf24a82546 Downloading [> ] 539.6kB/180.3MB 09:40:27 4faab25371b2 Downloading [==> ] 8.65MB/158.6MB 09:40:27 31e352740f53 Extracting [=======================> ] 1.573MB/3.398MB 09:40:27 31e352740f53 Extracting [=======================> ] 1.573MB/3.398MB 09:40:27 31e352740f53 Extracting [=======================> ] 1.573MB/3.398MB 09:40:27 31e352740f53 Extracting [=======================> ] 1.573MB/3.398MB 09:40:27 31e352740f53 Extracting [=======================> ] 1.573MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 31e352740f53 Extracting [==================================================>] 3.398MB/3.398MB 09:40:27 e8bf24a82546 Downloading [==> ] 10.81MB/180.3MB 09:40:27 21c7cf7066d0 Downloading [===============> ] 22.71MB/73.93MB 09:40:27 21c7cf7066d0 Downloading [===============> ] 22.71MB/73.93MB 09:40:27 22ebf0e44c85 Pulling fs layer 09:40:27 00b33c871d26 Pulling fs layer 09:40:27 6b11e56702ad Pulling fs layer 09:40:27 53d69aa7d3fc Pulling fs layer 09:40:27 a3ab11953ef9 Pulling fs layer 09:40:27 91ef9543149d Pulling fs layer 09:40:27 2ec4f59af178 Pulling fs layer 09:40:27 8b7e81cd5ef1 Pulling fs layer 09:40:27 c52916c1316e Pulling fs layer 09:40:27 d93f69e96600 Pulling fs layer 09:40:27 bbb9d15c45a1 Pulling fs layer 09:40:27 22ebf0e44c85 Waiting 09:40:27 00b33c871d26 Waiting 09:40:27 6b11e56702ad Waiting 09:40:27 53d69aa7d3fc Waiting 09:40:27 a3ab11953ef9 Waiting 09:40:27 91ef9543149d Waiting 09:40:27 2ec4f59af178 Waiting 09:40:27 8b7e81cd5ef1 Waiting 09:40:27 c52916c1316e Waiting 09:40:27 d93f69e96600 Waiting 09:40:27 bbb9d15c45a1 Waiting 09:40:27 22ebf0e44c85 Pulling fs layer 09:40:27 00b33c871d26 Pulling fs layer 09:40:27 6b11e56702ad Pulling fs layer 09:40:27 53d69aa7d3fc Pulling fs layer 09:40:27 a3ab11953ef9 Pulling fs layer 09:40:27 91ef9543149d Pulling fs layer 09:40:27 2ec4f59af178 Pulling fs layer 09:40:27 8b7e81cd5ef1 Pulling fs layer 09:40:27 c52916c1316e Pulling fs layer 09:40:27 7a1cb9ad7f75 Pulling fs layer 09:40:27 0a92c7dea7af Pulling fs layer 09:40:27 53d69aa7d3fc Waiting 09:40:27 22ebf0e44c85 Waiting 09:40:27 00b33c871d26 Waiting 09:40:27 a3ab11953ef9 Waiting 09:40:27 6b11e56702ad Waiting 09:40:27 91ef9543149d Waiting 09:40:27 2ec4f59af178 Waiting 09:40:27 8b7e81cd5ef1 Waiting 09:40:27 c52916c1316e Waiting 09:40:27 0a92c7dea7af Waiting 09:40:27 7a1cb9ad7f75 Waiting 09:40:27 31e352740f53 Pull complete 09:40:27 31e352740f53 Pull complete 09:40:27 31e352740f53 Pull complete 09:40:27 31e352740f53 Pull complete 09:40:27 31e352740f53 Pull complete 09:40:27 4faab25371b2 Downloading [=======> ] 23.79MB/158.6MB 09:40:27 21c7cf7066d0 Downloading [=====================> ] 31.36MB/73.93MB 09:40:27 21c7cf7066d0 Downloading [=====================> ] 31.36MB/73.93MB 09:40:27 e8bf24a82546 Downloading [=====> ] 19.46MB/180.3MB 09:40:27 4faab25371b2 Downloading [==========> ] 34.6MB/158.6MB 09:40:27 21c7cf7066d0 Downloading [============================> ] 42.17MB/73.93MB 09:40:27 21c7cf7066d0 Downloading [============================> ] 42.17MB/73.93MB 09:40:27 e8bf24a82546 Downloading [========> ] 32.44MB/180.3MB 09:40:27 4faab25371b2 Downloading [==============> ] 46.5MB/158.6MB 09:40:27 21c7cf7066d0 Downloading [====================================> ] 53.53MB/73.93MB 09:40:27 21c7cf7066d0 Downloading [====================================> ] 53.53MB/73.93MB 09:40:27 e8bf24a82546 Downloading [============> ] 45.42MB/180.3MB 09:40:27 4faab25371b2 Downloading [===================> ] 60.55MB/158.6MB 09:40:27 21c7cf7066d0 Downloading [=============================================> ] 67.04MB/73.93MB 09:40:27 21c7cf7066d0 Downloading [=============================================> ] 67.04MB/73.93MB 09:40:27 e8bf24a82546 Downloading [===============> ] 57.31MB/180.3MB 09:40:27 4faab25371b2 Downloading [======================> ] 71.91MB/158.6MB 09:40:27 21c7cf7066d0 Verifying Checksum 09:40:27 21c7cf7066d0 Download complete 09:40:27 21c7cf7066d0 Verifying Checksum 09:40:27 21c7cf7066d0 Download complete 09:40:27 154b803e2d93 Downloading [=> ] 3.002kB/84.13kB 09:40:27 154b803e2d93 Downloading [==================================================>] 84.13kB/84.13kB 09:40:27 154b803e2d93 Verifying Checksum 09:40:27 154b803e2d93 Download complete 09:40:27 e4305231c991 Downloading [==================================================>] 92B/92B 09:40:27 e4305231c991 Verifying Checksum 09:40:27 e4305231c991 Download complete 09:40:27 e8bf24a82546 Downloading [==================> ] 65.96MB/180.3MB 09:40:27 4faab25371b2 Downloading [=========================> ] 82.18MB/158.6MB 09:40:27 f469048fbe8d Downloading [==================================================>] 92B/92B 09:40:27 f469048fbe8d Download complete 09:40:27 21c7cf7066d0 Extracting [> ] 557.1kB/73.93MB 09:40:27 21c7cf7066d0 Extracting [> ] 557.1kB/73.93MB 09:40:27 c189e028fabb Downloading [==================================================>] 300B/300B 09:40:27 c189e028fabb Verifying Checksum 09:40:27 c189e028fabb Download complete 09:40:27 e8bf24a82546 Downloading [======================> ] 79.48MB/180.3MB 09:40:27 c9bd119720e4 Downloading [> ] 539.6kB/246.3MB 09:40:27 4faab25371b2 Downloading [===============================> ] 98.4MB/158.6MB 09:40:27 21c7cf7066d0 Extracting [===> ] 5.014MB/73.93MB 09:40:27 21c7cf7066d0 Extracting [===> ] 5.014MB/73.93MB 09:40:28 e8bf24a82546 Downloading [==========================> ] 95.16MB/180.3MB 09:40:28 c9bd119720e4 Downloading [=> ] 8.109MB/246.3MB 09:40:28 4faab25371b2 Downloading [===================================> ] 111.9MB/158.6MB 09:40:28 21c7cf7066d0 Extracting [=======> ] 11.7MB/73.93MB 09:40:28 21c7cf7066d0 Extracting [=======> ] 11.7MB/73.93MB 09:40:28 e8bf24a82546 Downloading [=============================> ] 108.1MB/180.3MB 09:40:28 c9bd119720e4 Downloading [===> ] 19.46MB/246.3MB 09:40:28 4faab25371b2 Downloading [=======================================> ] 123.8MB/158.6MB 09:40:28 21c7cf7066d0 Extracting [============> ] 18.94MB/73.93MB 09:40:28 21c7cf7066d0 Extracting [============> ] 18.94MB/73.93MB 09:40:28 e8bf24a82546 Downloading [=================================> ] 119.5MB/180.3MB 09:40:28 4faab25371b2 Downloading [==========================================> ] 135.7MB/158.6MB 09:40:28 c9bd119720e4 Downloading [======> ] 31.36MB/246.3MB 09:40:28 21c7cf7066d0 Extracting [=================> ] 25.62MB/73.93MB 09:40:28 21c7cf7066d0 Extracting [=================> ] 25.62MB/73.93MB 09:40:28 e8bf24a82546 Downloading [====================================> ] 132.5MB/180.3MB 09:40:28 4faab25371b2 Downloading [===============================================> ] 149.2MB/158.6MB 09:40:28 c9bd119720e4 Downloading [========> ] 44.33MB/246.3MB 09:40:28 21c7cf7066d0 Extracting [=====================> ] 31.75MB/73.93MB 09:40:28 21c7cf7066d0 Extracting [=====================> ] 31.75MB/73.93MB 09:40:28 e8bf24a82546 Downloading [=======================================> ] 143.3MB/180.3MB 09:40:28 4faab25371b2 Verifying Checksum 09:40:28 4faab25371b2 Download complete 09:40:28 c9bd119720e4 Downloading [==========> ] 54.07MB/246.3MB 09:40:28 c3cc5e3d19ac Download complete 09:40:28 0d2280d71230 Downloading [=> ] 3.001kB/127.4kB 09:40:28 0d2280d71230 Downloading [==================================================>] 127.4kB/127.4kB 09:40:28 0d2280d71230 Verifying Checksum 09:40:28 0d2280d71230 Download complete 09:40:28 984932e12fb0 Downloading [==================================================>] 1.147kB/1.147kB 09:40:28 984932e12fb0 Verifying Checksum 09:40:28 984932e12fb0 Download complete 09:40:28 21c7cf7066d0 Extracting [=========================> ] 37.32MB/73.93MB 09:40:28 21c7cf7066d0 Extracting [=========================> ] 37.32MB/73.93MB 09:40:28 e8bf24a82546 Downloading [===========================================> ] 155.7MB/180.3MB 09:40:28 5687ac571232 Downloading [> ] 539.6kB/91.54MB 09:40:28 c9bd119720e4 Downloading [==============> ] 70.83MB/246.3MB 09:40:28 21c7cf7066d0 Extracting [=============================> ] 42.89MB/73.93MB 09:40:28 21c7cf7066d0 Extracting [=============================> ] 42.89MB/73.93MB 09:40:28 e8bf24a82546 Downloading [===============================================> ] 172.5MB/180.3MB 09:40:28 5687ac571232 Downloading [==> ] 5.406MB/91.54MB 09:40:28 c9bd119720e4 Downloading [=================> ] 84.34MB/246.3MB 09:40:28 e8bf24a82546 Verifying Checksum 09:40:28 e8bf24a82546 Download complete 09:40:28 deac262509a5 Downloading [==================================================>] 1.118kB/1.118kB 09:40:28 deac262509a5 Verifying Checksum 09:40:28 deac262509a5 Download complete 09:40:28 21c7cf7066d0 Extracting [================================> ] 47.35MB/73.93MB 09:40:28 21c7cf7066d0 Extracting [================================> ] 47.35MB/73.93MB 09:40:28 5687ac571232 Downloading [=======> ] 12.98MB/91.54MB 09:40:28 c9bd119720e4 Downloading [===================> ] 96.24MB/246.3MB 09:40:28 e8bf24a82546 Extracting [> ] 557.1kB/180.3MB 09:40:28 57703e441b07 Downloading [> ] 539.6kB/73.93MB 09:40:28 21c7cf7066d0 Extracting [===================================> ] 52.92MB/73.93MB 09:40:28 21c7cf7066d0 Extracting [===================================> ] 52.92MB/73.93MB 09:40:28 5687ac571232 Downloading [==========> ] 20MB/91.54MB 09:40:28 c9bd119720e4 Downloading [======================> ] 112.5MB/246.3MB 09:40:28 57703e441b07 Downloading [=> ] 2.702MB/73.93MB 09:40:28 e8bf24a82546 Extracting [> ] 3.342MB/180.3MB 09:40:28 21c7cf7066d0 Extracting [=====================================> ] 55.71MB/73.93MB 09:40:28 21c7cf7066d0 Extracting [=====================================> ] 55.71MB/73.93MB 09:40:28 5687ac571232 Downloading [================> ] 29.74MB/91.54MB 09:40:28 c9bd119720e4 Downloading [==========================> ] 129.2MB/246.3MB 09:40:29 e8bf24a82546 Extracting [===> ] 11.7MB/180.3MB 09:40:29 57703e441b07 Downloading [====> ] 6.487MB/73.93MB 09:40:29 21c7cf7066d0 Extracting [=========================================> ] 61.28MB/73.93MB 09:40:29 21c7cf7066d0 Extracting [=========================================> ] 61.28MB/73.93MB 09:40:29 5687ac571232 Downloading [======================> ] 40.55MB/91.54MB 09:40:29 c9bd119720e4 Downloading [=============================> ] 143.8MB/246.3MB 09:40:29 e8bf24a82546 Extracting [======> ] 23.95MB/180.3MB 09:40:29 57703e441b07 Downloading [======> ] 10.27MB/73.93MB 09:40:29 21c7cf7066d0 Extracting [==============================================> ] 68.52MB/73.93MB 09:40:29 21c7cf7066d0 Extracting [==============================================> ] 68.52MB/73.93MB 09:40:29 5687ac571232 Downloading [=========================> ] 47.58MB/91.54MB 09:40:29 c9bd119720e4 Downloading [==============================> ] 152.5MB/246.3MB 09:40:29 e8bf24a82546 Extracting [========> ] 29.52MB/180.3MB 09:40:29 57703e441b07 Downloading [=========> ] 14.6MB/73.93MB 09:40:29 5687ac571232 Downloading [================================> ] 58.93MB/91.54MB 09:40:29 c9bd119720e4 Downloading [=================================> ] 167.1MB/246.3MB 09:40:29 21c7cf7066d0 Extracting [==================================================>] 73.93MB/73.93MB 09:40:29 21c7cf7066d0 Extracting [==================================================>] 73.93MB/73.93MB 09:40:29 e8bf24a82546 Extracting [==========> ] 39.55MB/180.3MB 09:40:29 57703e441b07 Downloading [==============> ] 21.63MB/73.93MB 09:40:29 5687ac571232 Downloading [=====================================> ] 69.2MB/91.54MB 09:40:29 c9bd119720e4 Downloading [====================================> ] 177.3MB/246.3MB 09:40:29 e8bf24a82546 Extracting [==============> ] 51.81MB/180.3MB 09:40:29 57703e441b07 Downloading [===================> ] 28.11MB/73.93MB 09:40:29 5687ac571232 Downloading [============================================> ] 80.56MB/91.54MB 09:40:29 c9bd119720e4 Downloading [======================================> ] 189.2MB/246.3MB 09:40:29 e8bf24a82546 Extracting [==================> ] 67.4MB/180.3MB 09:40:29 57703e441b07 Downloading [=======================> ] 35.14MB/73.93MB 09:40:29 5687ac571232 Downloading [=================================================> ] 91.37MB/91.54MB 09:40:29 c9bd119720e4 Downloading [=======================================> ] 196.3MB/246.3MB 09:40:29 5687ac571232 Verifying Checksum 09:40:29 5687ac571232 Download complete 09:40:29 7138254c3790 Downloading [> ] 343kB/32.98MB 09:40:29 57703e441b07 Downloading [==============================> ] 44.87MB/73.93MB 09:40:29 e8bf24a82546 Extracting [====================> ] 74.65MB/180.3MB 09:40:29 c9bd119720e4 Downloading [===========================================> ] 211.9MB/246.3MB 09:40:29 7138254c3790 Downloading [=========> ] 6.192MB/32.98MB 09:40:29 57703e441b07 Downloading [========================================> ] 59.47MB/73.93MB 09:40:29 e8bf24a82546 Extracting [=======================> ] 83MB/180.3MB 09:40:29 c9bd119720e4 Downloading [==============================================> ] 229.8MB/246.3MB 09:40:29 57703e441b07 Verifying Checksum 09:40:29 57703e441b07 Download complete 09:40:29 7138254c3790 Downloading [========================> ] 16.17MB/32.98MB 09:40:29 21c7cf7066d0 Pull complete 09:40:29 21c7cf7066d0 Pull complete 09:40:29 e8bf24a82546 Extracting [========================> ] 89.69MB/180.3MB 09:40:29 eb5e31f0ecf8 Extracting [==================================================>] 305B/305B 09:40:29 c3cc5e3d19ac Extracting [==================================================>] 296B/296B 09:40:29 c3cc5e3d19ac Extracting [==================================================>] 296B/296B 09:40:29 eb5e31f0ecf8 Extracting [==================================================>] 305B/305B 09:40:29 78f39bed0e83 Downloading [==================================================>] 1.077kB/1.077kB 09:40:29 78f39bed0e83 Verifying Checksum 09:40:29 78f39bed0e83 Download complete 09:40:29 c9bd119720e4 Verifying Checksum 09:40:29 c9bd119720e4 Download complete 09:40:29 40796999d308 Downloading [============================> ] 3.003kB/5.325kB 09:40:29 40796999d308 Downloading [==================================================>] 5.325kB/5.325kB 09:40:29 40796999d308 Verifying Checksum 09:40:29 40796999d308 Download complete 09:40:29 14ddc757aae0 Downloading [============================> ] 3.003kB/5.314kB 09:40:29 14ddc757aae0 Download complete 09:40:29 ebe1cd824584 Downloading [==================================================>] 1.037kB/1.037kB 09:40:29 ebe1cd824584 Verifying Checksum 09:40:29 ebe1cd824584 Download complete 09:40:29 d2893dc6732f Downloading [==================================================>] 1.038kB/1.038kB 09:40:29 d2893dc6732f Verifying Checksum 09:40:29 d2893dc6732f Download complete 09:40:29 7138254c3790 Downloading [======================================> ] 25.12MB/32.98MB 09:40:29 369dfa39565e Downloading [==========> ] 3.002kB/13.79kB 09:40:29 369dfa39565e Download complete 09:40:29 a23a963fcebe Downloading [==========> ] 3.002kB/13.9kB 09:40:29 a23a963fcebe Downloading [==================================================>] 13.9kB/13.9kB 09:40:29 a23a963fcebe Verifying Checksum 09:40:29 a23a963fcebe Download complete 09:40:29 e8bf24a82546 Extracting [=========================> ] 93.59MB/180.3MB 09:40:29 9146eb587aa8 Downloading [==================================================>] 2.856kB/2.856kB 09:40:29 9146eb587aa8 Verifying Checksum 09:40:29 9146eb587aa8 Download complete 09:40:29 a120f6888c1f Download complete 09:40:29 215302b53935 Downloading [==================================================>] 293B/293B 09:40:29 215302b53935 Verifying Checksum 09:40:29 215302b53935 Download complete 09:40:29 257d54e26411 Downloading [> ] 539.6kB/73.93MB 09:40:30 7138254c3790 Verifying Checksum 09:40:30 7138254c3790 Download complete 09:40:30 c3cc5e3d19ac Pull complete 09:40:30 eb5e31f0ecf8 Pull complete 09:40:30 eb2f448c7730 Downloading [=> ] 3.001kB/127kB 09:40:30 eb2f448c7730 Downloading [==================================================>] 127kB/127kB 09:40:30 eb2f448c7730 Download complete 09:40:30 0d2280d71230 Extracting [============> ] 32.77kB/127.4kB 09:40:30 57703e441b07 Extracting [> ] 557.1kB/73.93MB 09:40:30 0d2280d71230 Extracting [==================================================>] 127.4kB/127.4kB 09:40:30 c8ee90c58894 Downloading [==================================================>] 1.329kB/1.329kB 09:40:30 c8ee90c58894 Verifying Checksum 09:40:30 c8ee90c58894 Download complete 09:40:30 c990b7e46fc8 Downloading [==================================================>] 1.299kB/1.299kB 09:40:30 c990b7e46fc8 Verifying Checksum 09:40:30 c990b7e46fc8 Download complete 09:40:30 e30cdb86c4f0 Downloading [> ] 539.6kB/98.32MB 09:40:30 10ac4908093d Downloading [> ] 310.2kB/30.43MB 09:40:30 e8bf24a82546 Extracting [==========================> ] 96.93MB/180.3MB 09:40:30 257d54e26411 Downloading [========> ] 12.43MB/73.93MB 09:40:30 57703e441b07 Extracting [===> ] 4.456MB/73.93MB 09:40:30 4faab25371b2 Extracting [> ] 557.1kB/158.6MB 09:40:30 e30cdb86c4f0 Downloading [===> ] 7.568MB/98.32MB 09:40:30 10ac4908093d Downloading [===========> ] 6.847MB/30.43MB 09:40:30 0d2280d71230 Pull complete 09:40:30 984932e12fb0 Extracting [==================================================>] 1.147kB/1.147kB 09:40:30 984932e12fb0 Extracting [==================================================>] 1.147kB/1.147kB 09:40:30 e8bf24a82546 Extracting [============================> ] 101.4MB/180.3MB 09:40:30 257d54e26411 Downloading [===============> ] 22.71MB/73.93MB 09:40:30 57703e441b07 Extracting [=====> ] 7.799MB/73.93MB 09:40:30 4faab25371b2 Extracting [===> ] 12.26MB/158.6MB 09:40:30 e30cdb86c4f0 Downloading [========> ] 16.22MB/98.32MB 09:40:30 10ac4908093d Downloading [=============================> ] 18.05MB/30.43MB 09:40:30 e8bf24a82546 Extracting [=============================> ] 106.4MB/180.3MB 09:40:30 257d54e26411 Downloading [=======================> ] 35.14MB/73.93MB 09:40:30 984932e12fb0 Pull complete 09:40:30 4faab25371b2 Extracting [=======> ] 24.51MB/158.6MB 09:40:30 e30cdb86c4f0 Downloading [==============> ] 28.11MB/98.32MB 09:40:30 57703e441b07 Extracting [=======> ] 11.7MB/73.93MB 09:40:30 10ac4908093d Downloading [=================================================> ] 30.19MB/30.43MB 09:40:30 10ac4908093d Download complete 09:40:30 44779101e748 Downloading [==================================================>] 1.744kB/1.744kB 09:40:30 44779101e748 Verifying Checksum 09:40:30 44779101e748 Download complete 09:40:30 a721db3e3f3d Downloading [> ] 64.45kB/5.526MB 09:40:30 257d54e26411 Downloading [===============================> ] 47.04MB/73.93MB 09:40:30 e8bf24a82546 Extracting [==============================> ] 110.9MB/180.3MB 09:40:30 57703e441b07 Extracting [=========> ] 14.48MB/73.93MB 09:40:30 4faab25371b2 Extracting [==========> ] 32.87MB/158.6MB 09:40:30 e30cdb86c4f0 Downloading [=====================> ] 43.25MB/98.32MB 09:40:30 10ac4908093d Extracting [> ] 327.7kB/30.43MB 09:40:30 a721db3e3f3d Verifying Checksum 09:40:30 a721db3e3f3d Download complete 09:40:30 5687ac571232 Extracting [> ] 557.1kB/91.54MB 09:40:30 1850a929b84a Downloading [==================================================>] 149B/149B 09:40:30 1850a929b84a Verifying Checksum 09:40:30 1850a929b84a Download complete 09:40:30 257d54e26411 Downloading [========================================> ] 60.01MB/73.93MB 09:40:30 397a918c7da3 Downloading [==================================================>] 327B/327B 09:40:30 397a918c7da3 Verifying Checksum 09:40:30 397a918c7da3 Download complete 09:40:30 e8bf24a82546 Extracting [===============================> ] 113.6MB/180.3MB 09:40:30 57703e441b07 Extracting [===========> ] 17.27MB/73.93MB 09:40:30 4faab25371b2 Extracting [============> ] 39.55MB/158.6MB 09:40:30 10ac4908093d Extracting [=====> ] 3.604MB/30.43MB 09:40:30 806be17e856d Downloading [> ] 539.6kB/89.72MB 09:40:30 e30cdb86c4f0 Downloading [===========================> ] 53.53MB/98.32MB 09:40:30 5687ac571232 Extracting [====> ] 8.913MB/91.54MB 09:40:30 257d54e26411 Downloading [================================================> ] 71.37MB/73.93MB 09:40:30 257d54e26411 Verifying Checksum 09:40:30 257d54e26411 Download complete 09:40:30 634de6c90876 Downloading [===========================================> ] 3.011kB/3.49kB 09:40:30 634de6c90876 Downloading [==================================================>] 3.49kB/3.49kB 09:40:30 634de6c90876 Verifying Checksum 09:40:30 634de6c90876 Download complete 09:40:30 e8bf24a82546 Extracting [================================> ] 117MB/180.3MB 09:40:30 57703e441b07 Extracting [=============> ] 20.05MB/73.93MB 09:40:30 cd00854cfb1a Downloading [=====================> ] 3.011kB/6.971kB 09:40:30 cd00854cfb1a Downloading [==================================================>] 6.971kB/6.971kB 09:40:30 cd00854cfb1a Verifying Checksum 09:40:30 cd00854cfb1a Download complete 09:40:30 4faab25371b2 Extracting [===============> ] 49.02MB/158.6MB 09:40:30 806be17e856d Downloading [====> ] 8.65MB/89.72MB 09:40:30 9fa9226be034 Downloading [> ] 15.3kB/783kB 09:40:30 e30cdb86c4f0 Downloading [==================================> ] 67.58MB/98.32MB 09:40:30 10ac4908093d Extracting [===========> ] 7.209MB/30.43MB 09:40:30 5687ac571232 Extracting [========> ] 15.04MB/91.54MB 09:40:30 9fa9226be034 Downloading [==================================================>] 783kB/783kB 09:40:30 9fa9226be034 Download complete 09:40:30 9fa9226be034 Extracting [==> ] 32.77kB/783kB 09:40:30 1617e25568b2 Downloading [=> ] 15.3kB/480.9kB 09:40:30 1617e25568b2 Downloading [==================================================>] 480.9kB/480.9kB 09:40:30 1617e25568b2 Verifying Checksum 09:40:30 1617e25568b2 Download complete 09:40:30 e8bf24a82546 Extracting [=================================> ] 119.8MB/180.3MB 09:40:30 3ecda1bfd07b Downloading [> ] 539.6kB/55.21MB 09:40:30 4faab25371b2 Extracting [=================> ] 56.26MB/158.6MB 09:40:30 57703e441b07 Extracting [===============> ] 23.4MB/73.93MB 09:40:30 257d54e26411 Extracting [> ] 557.1kB/73.93MB 09:40:30 806be17e856d Downloading [========> ] 15.68MB/89.72MB 09:40:30 e30cdb86c4f0 Downloading [========================================> ] 80.02MB/98.32MB 09:40:30 10ac4908093d Extracting [===============> ] 9.175MB/30.43MB 09:40:30 5687ac571232 Extracting [============> ] 22.28MB/91.54MB 09:40:30 9fa9226be034 Extracting [=======================> ] 360.4kB/783kB 09:40:30 9fa9226be034 Extracting [==================================================>] 783kB/783kB 09:40:30 3ecda1bfd07b Downloading [=====> ] 5.946MB/55.21MB 09:40:30 4faab25371b2 Extracting [====================> ] 64.06MB/158.6MB 09:40:30 257d54e26411 Extracting [==> ] 3.342MB/73.93MB 09:40:30 e8bf24a82546 Extracting [==================================> ] 123.1MB/180.3MB 09:40:30 e30cdb86c4f0 Downloading [===============================================> ] 92.99MB/98.32MB 09:40:30 57703e441b07 Extracting [=================> ] 25.62MB/73.93MB 09:40:30 806be17e856d Downloading [=============> ] 24.87MB/89.72MB 09:40:30 10ac4908093d Extracting [==================> ] 11.14MB/30.43MB 09:40:30 5687ac571232 Extracting [==============> ] 27.3MB/91.54MB 09:40:30 3ecda1bfd07b Downloading [==========> ] 11.35MB/55.21MB 09:40:30 9fa9226be034 Pull complete 09:40:30 4faab25371b2 Extracting [====================> ] 66.29MB/158.6MB 09:40:30 1617e25568b2 Extracting [===> ] 32.77kB/480.9kB 09:40:30 e30cdb86c4f0 Downloading [=================================================> ] 96.78MB/98.32MB 09:40:30 257d54e26411 Extracting [===> ] 5.014MB/73.93MB 09:40:30 e30cdb86c4f0 Verifying Checksum 09:40:30 e30cdb86c4f0 Download complete 09:40:31 57703e441b07 Extracting [===================> ] 29.52MB/73.93MB 09:40:31 806be17e856d Downloading [================> ] 30.28MB/89.72MB 09:40:31 e8bf24a82546 Extracting [==================================> ] 124.8MB/180.3MB 09:40:31 5687ac571232 Extracting [================> ] 30.64MB/91.54MB 09:40:31 10ac4908093d Extracting [=====================> ] 13.11MB/30.43MB 09:40:31 ac9f4de4b762 Downloading [> ] 506.8kB/50.13MB 09:40:31 3ecda1bfd07b Downloading [====================> ] 22.17MB/55.21MB 09:40:31 4faab25371b2 Extracting [=======================> ] 72.97MB/158.6MB 09:40:31 257d54e26411 Extracting [=====> ] 7.799MB/73.93MB 09:40:31 57703e441b07 Extracting [=====================> ] 31.75MB/73.93MB 09:40:31 806be17e856d Downloading [========================> ] 43.25MB/89.72MB 09:40:31 10ac4908093d Extracting [==========================> ] 16.06MB/30.43MB 09:40:31 5687ac571232 Extracting [====================> ] 37.32MB/91.54MB 09:40:31 e8bf24a82546 Extracting [===================================> ] 128.1MB/180.3MB 09:40:31 ac9f4de4b762 Downloading [=======> ] 7.617MB/50.13MB 09:40:31 1617e25568b2 Extracting [==================================> ] 327.7kB/480.9kB 09:40:31 3ecda1bfd07b Downloading [================================> ] 36.22MB/55.21MB 09:40:31 4faab25371b2 Extracting [========================> ] 77.43MB/158.6MB 09:40:31 806be17e856d Downloading [==============================> ] 55.15MB/89.72MB 09:40:31 1617e25568b2 Extracting [==================================================>] 480.9kB/480.9kB 09:40:31 10ac4908093d Extracting [===============================> ] 19.01MB/30.43MB 09:40:31 5687ac571232 Extracting [========================> ] 44.56MB/91.54MB 09:40:31 257d54e26411 Extracting [=======> ] 10.58MB/73.93MB 09:40:31 1617e25568b2 Extracting [==================================================>] 480.9kB/480.9kB 09:40:31 ac9f4de4b762 Downloading [===============> ] 15.24MB/50.13MB 09:40:31 e8bf24a82546 Extracting [====================================> ] 130.4MB/180.3MB 09:40:31 57703e441b07 Extracting [=======================> ] 34.54MB/73.93MB 09:40:31 3ecda1bfd07b Downloading [=========================================> ] 45.42MB/55.21MB 09:40:31 4faab25371b2 Extracting [=========================> ] 82.44MB/158.6MB 09:40:31 806be17e856d Downloading [====================================> ] 64.88MB/89.72MB 09:40:31 10ac4908093d Extracting [===================================> ] 21.63MB/30.43MB 09:40:31 ac9f4de4b762 Downloading [========================> ] 24.38MB/50.13MB 09:40:31 5687ac571232 Extracting [===========================> ] 50.14MB/91.54MB 09:40:31 1617e25568b2 Pull complete 09:40:31 257d54e26411 Extracting [=========> ] 13.93MB/73.93MB 09:40:31 3ecda1bfd07b Verifying Checksum 09:40:31 e8bf24a82546 Extracting [=====================================> ] 133.7MB/180.3MB 09:40:31 3ecda1bfd07b Download complete 09:40:31 57703e441b07 Extracting [========================> ] 36.77MB/73.93MB 09:40:31 ea63b2e6315f Downloading [==================================================>] 605B/605B 09:40:31 ea63b2e6315f Verifying Checksum 09:40:31 ea63b2e6315f Download complete 09:40:31 4faab25371b2 Extracting [===========================> ] 88.01MB/158.6MB 09:40:31 fbd390d3bd00 Downloading [==================================================>] 2.675kB/2.675kB 09:40:31 fbd390d3bd00 Verifying Checksum 09:40:31 fbd390d3bd00 Download complete 09:40:31 806be17e856d Downloading [=========================================> ] 74.61MB/89.72MB 09:40:31 9b1ac15ef728 Downloading [================================================> ] 3.011kB/3.087kB 09:40:31 9b1ac15ef728 Downloading [==================================================>] 3.087kB/3.087kB 09:40:31 9b1ac15ef728 Verifying Checksum 09:40:31 9b1ac15ef728 Download complete 09:40:31 10ac4908093d Extracting [=======================================> ] 24.25MB/30.43MB 09:40:31 8682f304eb80 Downloading [=====================================> ] 3.011kB/4.023kB 09:40:31 8682f304eb80 Downloading [==================================================>] 4.023kB/4.023kB 09:40:31 8682f304eb80 Verifying Checksum 09:40:31 8682f304eb80 Download complete 09:40:31 ac9f4de4b762 Downloading [=====================================> ] 37.58MB/50.13MB 09:40:31 5687ac571232 Extracting [===============================> ] 56.82MB/91.54MB 09:40:31 5fbafe078afc Downloading [==================================================>] 1.44kB/1.44kB 09:40:31 5fbafe078afc Verifying Checksum 09:40:31 5fbafe078afc Download complete 09:40:31 7fb53fd2ae10 Downloading [=> ] 3.009kB/138kB 09:40:31 7fb53fd2ae10 Downloading [==================================================>] 138kB/138kB 09:40:31 7fb53fd2ae10 Verifying Checksum 09:40:31 7fb53fd2ae10 Download complete 09:40:31 57703e441b07 Extracting [==========================> ] 38.99MB/73.93MB 09:40:31 e8bf24a82546 Extracting [======================================> ] 137MB/180.3MB 09:40:31 257d54e26411 Extracting [==========> ] 16.15MB/73.93MB 09:40:31 592798bd3683 Downloading [==================================================>] 100B/100B 09:40:31 592798bd3683 Verifying Checksum 09:40:31 592798bd3683 Download complete 09:40:31 4faab25371b2 Extracting [=============================> ] 94.7MB/158.6MB 09:40:31 806be17e856d Downloading [================================================> ] 86.51MB/89.72MB 09:40:31 3ecda1bfd07b Extracting [> ] 557.1kB/55.21MB 09:40:31 806be17e856d Verifying Checksum 09:40:31 806be17e856d Download complete 09:40:31 473fdc983780 Downloading [==================================================>] 721B/721B 09:40:31 473fdc983780 Verifying Checksum 09:40:31 473fdc983780 Download complete 09:40:31 ac9f4de4b762 Downloading [================================================> ] 48.25MB/50.13MB 09:40:31 5687ac571232 Extracting [==================================> ] 62.39MB/91.54MB 09:40:31 10ac4908093d Extracting [===========================================> ] 26.54MB/30.43MB 09:40:31 ac9f4de4b762 Verifying Checksum 09:40:31 ac9f4de4b762 Download complete 09:40:31 c0e05c86127e Downloading [==================================================>] 141B/141B 09:40:31 c0e05c86127e Verifying Checksum 09:40:31 c0e05c86127e Download complete 09:40:31 4abcf2066143 Downloading [> ] 48.06kB/3.409MB 09:40:31 706651a94df6 Downloading [> ] 31.68kB/3.162MB 09:40:31 33e0a01314cc Downloading [> ] 48.06kB/4.333MB 09:40:31 257d54e26411 Extracting [===========> ] 17.27MB/73.93MB 09:40:31 e8bf24a82546 Extracting [======================================> ] 138.7MB/180.3MB 09:40:31 57703e441b07 Extracting [============================> ] 42.34MB/73.93MB 09:40:31 4faab25371b2 Extracting [===============================> ] 101.4MB/158.6MB 09:40:31 3ecda1bfd07b Extracting [==> ] 2.228MB/55.21MB 09:40:31 706651a94df6 Verifying Checksum 09:40:31 706651a94df6 Download complete 09:40:31 4abcf2066143 Verifying Checksum 09:40:31 4abcf2066143 Download complete 09:40:31 5687ac571232 Extracting [====================================> ] 67.4MB/91.54MB 09:40:31 f8b444c6ff40 Downloading [===> ] 3.01kB/47.97kB 09:40:31 f8b444c6ff40 Downloading [==================================================>] 47.97kB/47.97kB 09:40:31 4abcf2066143 Extracting [> ] 65.54kB/3.409MB 09:40:31 33e0a01314cc Verifying Checksum 09:40:31 33e0a01314cc Download complete 09:40:31 f8b444c6ff40 Verifying Checksum 09:40:31 f8b444c6ff40 Download complete 09:40:31 10ac4908093d Extracting [==============================================> ] 28.51MB/30.43MB 09:40:31 e6c38e6d3add Downloading [======> ] 3.01kB/23.82kB 09:40:31 e6c38e6d3add Download complete 09:40:31 e8bf24a82546 Extracting [======================================> ] 140.4MB/180.3MB 09:40:31 e35e8e85e24d Downloading [> ] 506.8kB/50.55MB 09:40:31 d0bef95bc6b2 Downloading [============> ] 3.01kB/11.92kB 09:40:31 d0bef95bc6b2 Downloading [==================================================>] 11.92kB/11.92kB 09:40:31 d0bef95bc6b2 Verifying Checksum 09:40:31 d0bef95bc6b2 Download complete 09:40:31 257d54e26411 Extracting [=============> ] 20.05MB/73.93MB 09:40:31 6ca01427385e Downloading [> ] 539.6kB/61.48MB 09:40:31 af860903a445 Downloading [==================================================>] 1.226kB/1.226kB 09:40:31 4faab25371b2 Extracting [==================================> ] 109.2MB/158.6MB 09:40:31 af860903a445 Verifying Checksum 09:40:31 af860903a445 Download complete 09:40:31 57703e441b07 Extracting [==============================> ] 44.56MB/73.93MB 09:40:31 3ecda1bfd07b Extracting [====> ] 4.456MB/55.21MB 09:40:31 5687ac571232 Extracting [========================================> ] 73.53MB/91.54MB 09:40:31 10ac4908093d Extracting [================================================> ] 29.49MB/30.43MB 09:40:31 e35e8e85e24d Downloading [=====> ] 5.586MB/50.55MB 09:40:31 e8bf24a82546 Extracting [=======================================> ] 143.2MB/180.3MB 09:40:31 4abcf2066143 Extracting [=====> ] 393.2kB/3.409MB 09:40:31 6ca01427385e Downloading [=====> ] 7.028MB/61.48MB 09:40:31 4faab25371b2 Extracting [====================================> ] 116.4MB/158.6MB 09:40:31 257d54e26411 Extracting [===============> ] 23.4MB/73.93MB 09:40:31 57703e441b07 Extracting [===============================> ] 46.24MB/73.93MB 09:40:31 22ebf0e44c85 Downloading [> ] 376.1kB/37.02MB 09:40:31 22ebf0e44c85 Downloading [> ] 376.1kB/37.02MB 09:40:31 5687ac571232 Extracting [===========================================> ] 79.1MB/91.54MB 09:40:31 3ecda1bfd07b Extracting [======> ] 6.685MB/55.21MB 09:40:31 10ac4908093d Extracting [=================================================> ] 30.15MB/30.43MB 09:40:31 e35e8e85e24d Downloading [=============> ] 13.71MB/50.55MB 09:40:31 4abcf2066143 Extracting [===================================> ] 2.425MB/3.409MB 09:40:31 4faab25371b2 Extracting [======================================> ] 122.6MB/158.6MB 09:40:31 6ca01427385e Downloading [==========> ] 13.52MB/61.48MB 09:40:31 e8bf24a82546 Extracting [========================================> ] 145.4MB/180.3MB 09:40:31 257d54e26411 Extracting [=================> ] 25.62MB/73.93MB 09:40:31 4abcf2066143 Extracting [==================================================>] 3.409MB/3.409MB 09:40:31 22ebf0e44c85 Downloading [=======> ] 5.683MB/37.02MB 09:40:31 22ebf0e44c85 Downloading [=======> ] 5.683MB/37.02MB 09:40:31 5687ac571232 Extracting [=============================================> ] 83.56MB/91.54MB 09:40:31 57703e441b07 Extracting [================================> ] 48.46MB/73.93MB 09:40:32 3ecda1bfd07b Extracting [=======> ] 8.356MB/55.21MB 09:40:32 10ac4908093d Extracting [==================================================>] 30.43MB/30.43MB 09:40:32 e35e8e85e24d Downloading [===================> ] 19.81MB/50.55MB 09:40:32 4faab25371b2 Extracting [========================================> ] 127.6MB/158.6MB 09:40:32 22ebf0e44c85 Downloading [===========> ] 8.346MB/37.02MB 09:40:32 22ebf0e44c85 Downloading [===========> ] 8.346MB/37.02MB 09:40:32 6ca01427385e Downloading [====================> ] 25.41MB/61.48MB 09:40:32 5687ac571232 Extracting [===============================================> ] 86.34MB/91.54MB 09:40:32 57703e441b07 Extracting [=================================> ] 50.14MB/73.93MB 09:40:32 e8bf24a82546 Extracting [=========================================> ] 148.2MB/180.3MB 09:40:32 e35e8e85e24d Downloading [========================> ] 24.89MB/50.55MB 09:40:32 257d54e26411 Extracting [==================> ] 27.85MB/73.93MB 09:40:32 4faab25371b2 Extracting [========================================> ] 129.2MB/158.6MB 09:40:32 5687ac571232 Extracting [==================================================>] 91.54MB/91.54MB 09:40:32 22ebf0e44c85 Downloading [===================> ] 14.76MB/37.02MB 09:40:32 22ebf0e44c85 Downloading [===================> ] 14.76MB/37.02MB 09:40:32 6ca01427385e Downloading [==========================> ] 32.98MB/61.48MB 09:40:32 3ecda1bfd07b Extracting [========> ] 9.47MB/55.21MB 09:40:32 57703e441b07 Extracting [===================================> ] 52.92MB/73.93MB 09:40:32 e8bf24a82546 Extracting [=========================================> ] 151MB/180.3MB 09:40:32 e35e8e85e24d Downloading [==================================> ] 35.04MB/50.55MB 09:40:32 4faab25371b2 Extracting [==========================================> ] 134.3MB/158.6MB 09:40:32 4abcf2066143 Pull complete 09:40:32 257d54e26411 Extracting [====================> ] 30.08MB/73.93MB 09:40:32 c0e05c86127e Extracting [==================================================>] 141B/141B 09:40:32 10ac4908093d Pull complete 09:40:32 c0e05c86127e Extracting [==================================================>] 141B/141B 09:40:32 6ca01427385e Downloading [====================================> ] 44.87MB/61.48MB 09:40:32 22ebf0e44c85 Downloading [============================> ] 21.17MB/37.02MB 09:40:32 22ebf0e44c85 Downloading [============================> ] 21.17MB/37.02MB 09:40:32 3ecda1bfd07b Extracting [==========> ] 11.7MB/55.21MB 09:40:32 5687ac571232 Pull complete 09:40:32 44779101e748 Extracting [==================================================>] 1.744kB/1.744kB 09:40:32 deac262509a5 Extracting [==================================================>] 1.118kB/1.118kB 09:40:32 e8bf24a82546 Extracting [==========================================> ] 154.3MB/180.3MB 09:40:32 deac262509a5 Extracting [==================================================>] 1.118kB/1.118kB 09:40:32 57703e441b07 Extracting [=====================================> ] 55.15MB/73.93MB 09:40:32 44779101e748 Extracting [==================================================>] 1.744kB/1.744kB 09:40:32 4faab25371b2 Extracting [============================================> ] 142MB/158.6MB 09:40:32 e35e8e85e24d Downloading [================================================> ] 49.27MB/50.55MB 09:40:32 e35e8e85e24d Verifying Checksum 09:40:32 e35e8e85e24d Download complete 09:40:32 6ca01427385e Downloading [=========================================> ] 50.82MB/61.48MB 09:40:32 257d54e26411 Extracting [======================> ] 33.98MB/73.93MB 09:40:32 22ebf0e44c85 Downloading [========================================> ] 29.89MB/37.02MB 09:40:32 22ebf0e44c85 Downloading [========================================> ] 29.89MB/37.02MB 09:40:32 3ecda1bfd07b Extracting [============> ] 13.93MB/55.21MB 09:40:32 c0e05c86127e Pull complete 09:40:32 706651a94df6 Extracting [> ] 32.77kB/3.162MB 09:40:32 e8bf24a82546 Extracting [===========================================> ] 158.2MB/180.3MB 09:40:32 57703e441b07 Extracting [=======================================> ] 57.93MB/73.93MB 09:40:32 22ebf0e44c85 Verifying Checksum 09:40:32 22ebf0e44c85 Verifying Checksum 09:40:32 22ebf0e44c85 Download complete 09:40:32 22ebf0e44c85 Download complete 09:40:32 6ca01427385e Verifying Checksum 09:40:32 6ca01427385e Download complete 09:40:32 4faab25371b2 Extracting [==============================================> ] 148.2MB/158.6MB 09:40:32 00b33c871d26 Downloading [> ] 531.7kB/253.3MB 09:40:32 00b33c871d26 Downloading [> ] 531.7kB/253.3MB 09:40:32 257d54e26411 Extracting [========================> ] 35.65MB/73.93MB 09:40:32 44779101e748 Pull complete 09:40:32 deac262509a5 Pull complete 09:40:32 a721db3e3f3d Extracting [> ] 65.54kB/5.526MB 09:40:32 57703e441b07 Extracting [=======================================> ] 59.05MB/73.93MB 09:40:32 706651a94df6 Extracting [=====> ] 327.7kB/3.162MB 09:40:32 3ecda1bfd07b Extracting [===============> ] 16.71MB/55.21MB 09:40:32 e8bf24a82546 Extracting [============================================> ] 161.5MB/180.3MB 09:40:32 api Pulled 09:40:32 6b11e56702ad Downloading [> ] 77.47kB/7.707MB 09:40:32 6b11e56702ad Downloading [> ] 77.47kB/7.707MB 09:40:32 53d69aa7d3fc Downloading [=> ] 718B/19.96kB 09:40:32 53d69aa7d3fc Downloading [=> ] 718B/19.96kB 09:40:32 53d69aa7d3fc Verifying Checksum 09:40:32 53d69aa7d3fc Download complete 09:40:32 53d69aa7d3fc Verifying Checksum 09:40:32 53d69aa7d3fc Download complete 09:40:32 4faab25371b2 Extracting [===============================================> ] 151.5MB/158.6MB 09:40:32 00b33c871d26 Downloading [=> ] 8.547MB/253.3MB 09:40:32 00b33c871d26 Downloading [=> ] 8.547MB/253.3MB 09:40:32 22ebf0e44c85 Extracting [> ] 393.2kB/37.02MB 09:40:32 22ebf0e44c85 Extracting [> ] 393.2kB/37.02MB 09:40:32 257d54e26411 Extracting [=========================> ] 37.88MB/73.93MB 09:40:32 706651a94df6 Extracting [=================> ] 1.114MB/3.162MB 09:40:32 e8bf24a82546 Extracting [=============================================> ] 162.7MB/180.3MB 09:40:32 57703e441b07 Extracting [=========================================> ] 60.72MB/73.93MB 09:40:32 6b11e56702ad Downloading [========================> ] 3.701MB/7.707MB 09:40:32 6b11e56702ad Downloading [========================> ] 3.701MB/7.707MB 09:40:32 4faab25371b2 Extracting [=================================================> ] 156.5MB/158.6MB 09:40:32 a721db3e3f3d Extracting [==> ] 262.1kB/5.526MB 09:40:32 a3ab11953ef9 Downloading [> ] 407.8kB/39.52MB 09:40:32 a3ab11953ef9 Downloading [> ] 407.8kB/39.52MB 09:40:32 00b33c871d26 Downloading [==> ] 14.44MB/253.3MB 09:40:32 00b33c871d26 Downloading [==> ] 14.44MB/253.3MB 09:40:32 4faab25371b2 Extracting [==================================================>] 158.6MB/158.6MB 09:40:32 257d54e26411 Extracting [==========================> ] 39.55MB/73.93MB 09:40:32 22ebf0e44c85 Extracting [====> ] 3.539MB/37.02MB 09:40:32 22ebf0e44c85 Extracting [====> ] 3.539MB/37.02MB 09:40:32 3ecda1bfd07b Extracting [==================> ] 20.61MB/55.21MB 09:40:32 a3ab11953ef9 Downloading [=========> ] 7.285MB/39.52MB 09:40:32 a3ab11953ef9 Downloading [=========> ] 7.285MB/39.52MB 09:40:32 6b11e56702ad Downloading [==============================================> ] 7.183MB/7.707MB 09:40:32 6b11e56702ad Downloading [==============================================> ] 7.183MB/7.707MB 09:40:32 6b11e56702ad Verifying Checksum 09:40:32 6b11e56702ad Download complete 09:40:32 6b11e56702ad Verifying Checksum 09:40:32 6b11e56702ad Download complete 09:40:32 00b33c871d26 Downloading [====> ] 24.62MB/253.3MB 09:40:32 00b33c871d26 Downloading [====> ] 24.62MB/253.3MB 09:40:32 a721db3e3f3d Extracting [=============> ] 1.507MB/5.526MB 09:40:32 706651a94df6 Extracting [================================================> ] 3.047MB/3.162MB 09:40:32 e8bf24a82546 Extracting [=============================================> ] 164.3MB/180.3MB 09:40:32 57703e441b07 Extracting [==========================================> ] 63.5MB/73.93MB 09:40:32 4faab25371b2 Pull complete 09:40:32 6b867d96d427 Extracting [==================================================>] 1.153kB/1.153kB 09:40:32 22ebf0e44c85 Extracting [=====> ] 4.325MB/37.02MB 09:40:32 22ebf0e44c85 Extracting [=====> ] 4.325MB/37.02MB 09:40:32 257d54e26411 Extracting [===========================> ] 41.22MB/73.93MB 09:40:32 6b867d96d427 Extracting [==================================================>] 1.153kB/1.153kB 09:40:32 3ecda1bfd07b Extracting [===================> ] 21.17MB/55.21MB 09:40:32 706651a94df6 Extracting [==================================================>] 3.162MB/3.162MB 09:40:32 a3ab11953ef9 Downloading [===============> ] 12.5MB/39.52MB 09:40:32 a3ab11953ef9 Downloading [===============> ] 12.5MB/39.52MB 09:40:32 00b33c871d26 Downloading [=====> ] 28.36MB/253.3MB 09:40:32 00b33c871d26 Downloading [=====> ] 28.36MB/253.3MB 09:40:32 91ef9543149d Downloading [================================> ] 719B/1.101kB 09:40:32 91ef9543149d Downloading [================================> ] 719B/1.101kB 09:40:32 91ef9543149d Downloading [==================================================>] 1.101kB/1.101kB 09:40:32 91ef9543149d Downloading [==================================================>] 1.101kB/1.101kB 09:40:32 91ef9543149d Verifying Checksum 09:40:32 91ef9543149d Download complete 09:40:32 91ef9543149d Verifying Checksum 09:40:32 91ef9543149d Download complete 09:40:32 a721db3e3f3d Extracting [===================================> ] 3.932MB/5.526MB 09:40:32 e8bf24a82546 Extracting [==============================================> ] 166.6MB/180.3MB 09:40:32 57703e441b07 Extracting [============================================> ] 66.29MB/73.93MB 09:40:32 257d54e26411 Extracting [=============================> ] 43.45MB/73.93MB 09:40:32 22ebf0e44c85 Extracting [=========> ] 6.685MB/37.02MB 09:40:32 22ebf0e44c85 Extracting [=========> ] 6.685MB/37.02MB 09:40:32 3ecda1bfd07b Extracting [=====================> ] 23.4MB/55.21MB 09:40:33 a3ab11953ef9 Downloading [======================> ] 18.17MB/39.52MB 09:40:33 a3ab11953ef9 Downloading [======================> ] 18.17MB/39.52MB 09:40:33 00b33c871d26 Downloading [=======> ] 35.89MB/253.3MB 09:40:33 00b33c871d26 Downloading [=======> ] 35.89MB/253.3MB 09:40:33 6b867d96d427 Pull complete 09:40:33 e8bf24a82546 Extracting [===============================================> ] 169.9MB/180.3MB 09:40:33 93832cc54357 Extracting [==================================================>] 1.127kB/1.127kB 09:40:33 93832cc54357 Extracting [==================================================>] 1.127kB/1.127kB 09:40:33 2ec4f59af178 Downloading [========================================> ] 720B/881B 09:40:33 2ec4f59af178 Downloading [========================================> ] 720B/881B 09:40:33 2ec4f59af178 Download complete 09:40:33 2ec4f59af178 Download complete 09:40:33 706651a94df6 Pull complete 09:40:33 57703e441b07 Extracting [===============================================> ] 69.63MB/73.93MB 09:40:33 33e0a01314cc Extracting [> ] 65.54kB/4.333MB 09:40:33 a721db3e3f3d Extracting [=========================================> ] 4.588MB/5.526MB 09:40:33 3ecda1bfd07b Extracting [=======================> ] 26.18MB/55.21MB 09:40:33 257d54e26411 Extracting [==============================> ] 45.68MB/73.93MB 09:40:33 22ebf0e44c85 Extracting [=============> ] 9.83MB/37.02MB 09:40:33 22ebf0e44c85 Extracting [=============> ] 9.83MB/37.02MB 09:40:33 a3ab11953ef9 Downloading [=================================> ] 26.29MB/39.52MB 09:40:33 a3ab11953ef9 Downloading [=================================> ] 26.29MB/39.52MB 09:40:33 00b33c871d26 Downloading [=======> ] 39.62MB/253.3MB 09:40:33 00b33c871d26 Downloading [=======> ] 39.62MB/253.3MB 09:40:33 e8bf24a82546 Extracting [===============================================> ] 171.6MB/180.3MB 09:40:33 57703e441b07 Extracting [================================================> ] 72.42MB/73.93MB 09:40:33 8b7e81cd5ef1 Downloading [==================================================>] 131B/131B 09:40:33 8b7e81cd5ef1 Downloading [==================================================>] 131B/131B 09:40:33 8b7e81cd5ef1 Verifying Checksum 09:40:33 8b7e81cd5ef1 Verifying Checksum 09:40:33 8b7e81cd5ef1 Download complete 09:40:33 8b7e81cd5ef1 Download complete 09:40:33 33e0a01314cc Extracting [===> ] 262.1kB/4.333MB 09:40:33 257d54e26411 Extracting [================================> ] 47.91MB/73.93MB 09:40:33 a3ab11953ef9 Downloading [=======================================> ] 31.14MB/39.52MB 09:40:33 a3ab11953ef9 Downloading [=======================================> ] 31.14MB/39.52MB 09:40:33 93832cc54357 Pull complete 09:40:33 00b33c871d26 Downloading [========> ] 44.98MB/253.3MB 09:40:33 00b33c871d26 Downloading [========> ] 44.98MB/253.3MB 09:40:33 3ecda1bfd07b Extracting [=========================> ] 28.41MB/55.21MB 09:40:33 22ebf0e44c85 Extracting [================> ] 12.58MB/37.02MB 09:40:33 22ebf0e44c85 Extracting [================> ] 12.58MB/37.02MB 09:40:33 simulator Pulled 09:40:33 a721db3e3f3d Extracting [===========================================> ] 4.85MB/5.526MB 09:40:33 e8bf24a82546 Extracting [===============================================> ] 172.7MB/180.3MB 09:40:33 33e0a01314cc Extracting [==================> ] 1.638MB/4.333MB 09:40:33 57703e441b07 Extracting [==================================================>] 73.93MB/73.93MB 09:40:33 c52916c1316e Downloading [==================================================>] 171B/171B 09:40:33 c52916c1316e Downloading [==================================================>] 171B/171B 09:40:33 c52916c1316e Verifying Checksum 09:40:33 c52916c1316e Download complete 09:40:33 c52916c1316e Verifying Checksum 09:40:33 c52916c1316e Download complete 09:40:33 a721db3e3f3d Extracting [==================================================>] 5.526MB/5.526MB 09:40:33 a3ab11953ef9 Verifying Checksum 09:40:33 a3ab11953ef9 Download complete 09:40:33 a3ab11953ef9 Verifying Checksum 09:40:33 a3ab11953ef9 Download complete 09:40:33 00b33c871d26 Downloading [==========> ] 55.18MB/253.3MB 09:40:33 00b33c871d26 Downloading [==========> ] 55.18MB/253.3MB 09:40:33 257d54e26411 Extracting [==================================> ] 50.69MB/73.93MB 09:40:33 22ebf0e44c85 Extracting [====================> ] 15.34MB/37.02MB 09:40:33 22ebf0e44c85 Extracting [====================> ] 15.34MB/37.02MB 09:40:33 3ecda1bfd07b Extracting [===============================> ] 34.54MB/55.21MB 09:40:33 33e0a01314cc Extracting [==========================================> ] 3.67MB/4.333MB 09:40:33 57703e441b07 Pull complete 09:40:33 d93f69e96600 Downloading [> ] 534.8kB/115.2MB 09:40:33 e8bf24a82546 Extracting [================================================> ] 174.4MB/180.3MB 09:40:33 bbb9d15c45a1 Downloading [=========> ] 719B/3.633kB 09:40:33 bbb9d15c45a1 Download complete 09:40:33 3ecda1bfd07b Extracting [================================> ] 36.21MB/55.21MB 09:40:33 33e0a01314cc Extracting [==================================================>] 4.333MB/4.333MB 09:40:33 00b33c871d26 Downloading [===========> ] 60MB/253.3MB 09:40:33 00b33c871d26 Downloading [===========> ] 60MB/253.3MB 09:40:33 22ebf0e44c85 Extracting [======================> ] 16.91MB/37.02MB 09:40:33 22ebf0e44c85 Extracting [======================> ] 16.91MB/37.02MB 09:40:33 a721db3e3f3d Pull complete 09:40:33 257d54e26411 Extracting [===================================> ] 51.81MB/73.93MB 09:40:33 1850a929b84a Extracting [==================================================>] 149B/149B 09:40:33 1850a929b84a Extracting [==================================================>] 149B/149B 09:40:33 7138254c3790 Extracting [> ] 360.4kB/32.98MB 09:40:33 d93f69e96600 Downloading [===> ] 6.933MB/115.2MB 09:40:33 33e0a01314cc Pull complete 09:40:33 f8b444c6ff40 Extracting [==================================> ] 32.77kB/47.97kB 09:40:33 f8b444c6ff40 Extracting [==================================================>] 47.97kB/47.97kB 09:40:33 e8bf24a82546 Extracting [================================================> ] 175.5MB/180.3MB 09:40:33 00b33c871d26 Downloading [=============> ] 69.11MB/253.3MB 09:40:33 00b33c871d26 Downloading [=============> ] 69.11MB/253.3MB 09:40:33 3ecda1bfd07b Extracting [========================================> ] 44.56MB/55.21MB 09:40:33 22ebf0e44c85 Extracting [=========================> ] 18.87MB/37.02MB 09:40:33 22ebf0e44c85 Extracting [=========================> ] 18.87MB/37.02MB 09:40:33 7a1cb9ad7f75 Downloading [> ] 539.9kB/115.2MB 09:40:33 d93f69e96600 Downloading [======> ] 15.53MB/115.2MB 09:40:33 257d54e26411 Extracting [====================================> ] 54.03MB/73.93MB 09:40:33 7138254c3790 Extracting [===> ] 2.523MB/32.98MB 09:40:33 1850a929b84a Pull complete 09:40:33 397a918c7da3 Extracting [==================================================>] 327B/327B 09:40:33 397a918c7da3 Extracting [==================================================>] 327B/327B 09:40:33 00b33c871d26 Downloading [==============> ] 71.78MB/253.3MB 09:40:33 00b33c871d26 Downloading [==============> ] 71.78MB/253.3MB 09:40:33 e8bf24a82546 Extracting [================================================> ] 176.6MB/180.3MB 09:40:33 3ecda1bfd07b Extracting [============================================> ] 49.02MB/55.21MB 09:40:33 f8b444c6ff40 Pull complete 09:40:33 e6c38e6d3add Extracting [==================================================>] 23.82kB/23.82kB 09:40:33 e6c38e6d3add Extracting [==================================================>] 23.82kB/23.82kB 09:40:33 22ebf0e44c85 Extracting [============================> ] 20.84MB/37.02MB 09:40:33 22ebf0e44c85 Extracting [============================> ] 20.84MB/37.02MB 09:40:33 7a1cb9ad7f75 Downloading [===> ] 8.055MB/115.2MB 09:40:33 d93f69e96600 Downloading [==========> ] 24.66MB/115.2MB 09:40:33 7138254c3790 Extracting [======> ] 3.965MB/32.98MB 09:40:33 257d54e26411 Extracting [=====================================> ] 55.15MB/73.93MB 09:40:33 00b33c871d26 Downloading [================> ] 81.44MB/253.3MB 09:40:33 00b33c871d26 Downloading [================> ] 81.44MB/253.3MB 09:40:33 3ecda1bfd07b Extracting [================================================> ] 54.03MB/55.21MB 09:40:33 397a918c7da3 Pull complete 09:40:33 d93f69e96600 Downloading [=============> ] 31.06MB/115.2MB 09:40:33 7a1cb9ad7f75 Downloading [======> ] 15.04MB/115.2MB 09:40:33 22ebf0e44c85 Extracting [===============================> ] 23.2MB/37.02MB 09:40:33 22ebf0e44c85 Extracting [===============================> ] 23.2MB/37.02MB 09:40:33 e8bf24a82546 Extracting [=================================================> ] 178.3MB/180.3MB 09:40:33 7138254c3790 Extracting [========> ] 5.407MB/32.98MB 09:40:33 257d54e26411 Extracting [=======================================> ] 57.93MB/73.93MB 09:40:33 e6c38e6d3add Pull complete 09:40:33 00b33c871d26 Downloading [=================> ] 86.25MB/253.3MB 09:40:33 00b33c871d26 Downloading [=================> ] 86.25MB/253.3MB 09:40:33 d93f69e96600 Downloading [================> ] 39.11MB/115.2MB 09:40:33 7138254c3790 Extracting [==========> ] 6.849MB/32.98MB 09:40:33 7a1cb9ad7f75 Downloading [=========> ] 22.54MB/115.2MB 09:40:33 e8bf24a82546 Extracting [=================================================> ] 179.9MB/180.3MB 09:40:33 257d54e26411 Extracting [=======================================> ] 59.05MB/73.93MB 09:40:33 3ecda1bfd07b Extracting [=================================================> ] 55.15MB/55.21MB 09:40:33 22ebf0e44c85 Extracting [==================================> ] 25.56MB/37.02MB 09:40:33 22ebf0e44c85 Extracting [==================================> ] 25.56MB/37.02MB 09:40:33 806be17e856d Extracting [> ] 557.1kB/89.72MB 09:40:34 3ecda1bfd07b Extracting [==================================================>] 55.21MB/55.21MB 09:40:34 00b33c871d26 Downloading [===================> ] 97.52MB/253.3MB 09:40:34 00b33c871d26 Downloading [===================> ] 97.52MB/253.3MB 09:40:34 6ca01427385e Extracting [> ] 557.1kB/61.48MB 09:40:34 e8bf24a82546 Extracting [==================================================>] 180.3MB/180.3MB 09:40:34 7138254c3790 Extracting [=============> ] 8.651MB/32.98MB 09:40:34 d93f69e96600 Downloading [===================> ] 44.99MB/115.2MB 09:40:34 7a1cb9ad7f75 Downloading [===========> ] 26.29MB/115.2MB 09:40:34 257d54e26411 Extracting [=========================================> ] 60.72MB/73.93MB 09:40:34 3ecda1bfd07b Pull complete 09:40:34 806be17e856d Extracting [=> ] 2.228MB/89.72MB 09:40:34 22ebf0e44c85 Extracting [=======================================> ] 29.1MB/37.02MB 09:40:34 22ebf0e44c85 Extracting [=======================================> ] 29.1MB/37.02MB 09:40:34 00b33c871d26 Downloading [====================> ] 103.9MB/253.3MB 09:40:34 00b33c871d26 Downloading [====================> ] 103.9MB/253.3MB 09:40:34 6ca01427385e Extracting [==> ] 2.785MB/61.48MB 09:40:34 d93f69e96600 Downloading [=====================> ] 49.3MB/115.2MB 09:40:34 7138254c3790 Extracting [================> ] 10.81MB/32.98MB 09:40:34 7a1cb9ad7f75 Downloading [=============> ] 31.61MB/115.2MB 09:40:34 257d54e26411 Extracting [===========================================> ] 64.62MB/73.93MB 09:40:34 00b33c871d26 Downloading [=======================> ] 117.9MB/253.3MB 09:40:34 00b33c871d26 Downloading [=======================> ] 117.9MB/253.3MB 09:40:34 806be17e856d Extracting [==> ] 4.456MB/89.72MB 09:40:34 22ebf0e44c85 Extracting [==========================================> ] 31.46MB/37.02MB 09:40:34 22ebf0e44c85 Extracting [==========================================> ] 31.46MB/37.02MB 09:40:34 e8bf24a82546 Pull complete 09:40:34 154b803e2d93 Extracting [===================> ] 32.77kB/84.13kB 09:40:34 6ca01427385e Extracting [===> ] 3.899MB/61.48MB 09:40:34 154b803e2d93 Extracting [==================================================>] 84.13kB/84.13kB 09:40:34 d93f69e96600 Downloading [========================> ] 55.76MB/115.2MB 09:40:34 7138254c3790 Extracting [==================> ] 12.26MB/32.98MB 09:40:34 ac9f4de4b762 Extracting [> ] 524.3kB/50.13MB 09:40:34 257d54e26411 Extracting [=============================================> ] 66.85MB/73.93MB 09:40:34 7a1cb9ad7f75 Downloading [================> ] 36.96MB/115.2MB 09:40:34 806be17e856d Extracting [===> ] 6.128MB/89.72MB 09:40:34 00b33c871d26 Downloading [=========================> ] 129.6MB/253.3MB 09:40:34 00b33c871d26 Downloading [=========================> ] 129.6MB/253.3MB 09:40:34 d93f69e96600 Downloading [=============================> ] 68.09MB/115.2MB 09:40:34 6ca01427385e Extracting [====> ] 6.128MB/61.48MB 09:40:34 ac9f4de4b762 Extracting [==> ] 2.621MB/50.13MB 09:40:34 257d54e26411 Extracting [===============================================> ] 69.63MB/73.93MB 09:40:34 7a1cb9ad7f75 Downloading [==================> ] 41.78MB/115.2MB 09:40:34 7138254c3790 Extracting [======================> ] 14.78MB/32.98MB 09:40:34 22ebf0e44c85 Extracting [=============================================> ] 33.82MB/37.02MB 09:40:34 22ebf0e44c85 Extracting [=============================================> ] 33.82MB/37.02MB 09:40:34 00b33c871d26 Downloading [===========================> ] 139.7MB/253.3MB 09:40:34 00b33c871d26 Downloading [===========================> ] 139.7MB/253.3MB 09:40:34 806be17e856d Extracting [====> ] 8.913MB/89.72MB 09:40:34 257d54e26411 Extracting [================================================> ] 72.42MB/73.93MB 09:40:34 7138254c3790 Extracting [========================> ] 16.22MB/32.98MB 09:40:34 154b803e2d93 Pull complete 09:40:34 7a1cb9ad7f75 Downloading [===================> ] 45.01MB/115.2MB 09:40:34 6ca01427385e Extracting [======> ] 8.356MB/61.48MB 09:40:34 d93f69e96600 Downloading [================================> ] 74.49MB/115.2MB 09:40:34 e4305231c991 Extracting [==================================================>] 92B/92B 09:40:34 e4305231c991 Extracting [==================================================>] 92B/92B 09:40:34 ac9f4de4b762 Extracting [====> ] 4.194MB/50.13MB 09:40:34 22ebf0e44c85 Extracting [==============================================> ] 34.6MB/37.02MB 09:40:34 22ebf0e44c85 Extracting [==============================================> ] 34.6MB/37.02MB 09:40:34 00b33c871d26 Downloading [=============================> ] 149.4MB/253.3MB 09:40:34 00b33c871d26 Downloading [=============================> ] 149.4MB/253.3MB 09:40:34 7138254c3790 Extracting [============================> ] 18.74MB/32.98MB 09:40:34 d93f69e96600 Downloading [======================================> ] 87.88MB/115.2MB 09:40:34 7a1cb9ad7f75 Downloading [======================> ] 50.9MB/115.2MB 09:40:34 806be17e856d Extracting [=====> ] 10.58MB/89.72MB 09:40:34 e4305231c991 Pull complete 09:40:34 22ebf0e44c85 Extracting [================================================> ] 36.18MB/37.02MB 09:40:34 22ebf0e44c85 Extracting [================================================> ] 36.18MB/37.02MB 09:40:34 f469048fbe8d Extracting [==================================================>] 92B/92B 09:40:34 257d54e26411 Extracting [==================================================>] 73.93MB/73.93MB 09:40:34 6ca01427385e Extracting [=======> ] 9.47MB/61.48MB 09:40:34 f469048fbe8d Extracting [==================================================>] 92B/92B 09:40:34 7138254c3790 Extracting [==============================> ] 19.82MB/32.98MB 09:40:34 00b33c871d26 Downloading [==============================> ] 153.1MB/253.3MB 09:40:34 00b33c871d26 Downloading [==============================> ] 153.1MB/253.3MB 09:40:34 ac9f4de4b762 Extracting [======> ] 6.816MB/50.13MB 09:40:34 d93f69e96600 Downloading [========================================> ] 93.76MB/115.2MB 09:40:34 7a1cb9ad7f75 Downloading [=======================> ] 53.05MB/115.2MB 09:40:34 806be17e856d Extracting [======> ] 11.7MB/89.72MB 09:40:34 22ebf0e44c85 Extracting [==================================================>] 37.02MB/37.02MB 09:40:34 22ebf0e44c85 Extracting [==================================================>] 37.02MB/37.02MB 09:40:34 00b33c871d26 Downloading [================================> ] 163.3MB/253.3MB 09:40:34 00b33c871d26 Downloading [================================> ] 163.3MB/253.3MB 09:40:34 d93f69e96600 Downloading [==============================================> ] 106.1MB/115.2MB 09:40:34 7138254c3790 Extracting [==============================> ] 20.19MB/32.98MB 09:40:34 ac9f4de4b762 Extracting [========> ] 8.913MB/50.13MB 09:40:34 7a1cb9ad7f75 Downloading [========================> ] 56.81MB/115.2MB 09:40:34 806be17e856d Extracting [=======> ] 13.37MB/89.72MB 09:40:34 6ca01427385e Extracting [=========> ] 11.14MB/61.48MB 09:40:34 00b33c871d26 Downloading [=================================> ] 169.7MB/253.3MB 09:40:34 00b33c871d26 Downloading [=================================> ] 169.7MB/253.3MB 09:40:34 d93f69e96600 Downloading [===============================================> ] 110.4MB/115.2MB 09:40:34 ac9f4de4b762 Extracting [==========> ] 10.49MB/50.13MB 09:40:34 7138254c3790 Extracting [================================> ] 21.27MB/32.98MB 09:40:34 d93f69e96600 Verifying Checksum 09:40:34 d93f69e96600 Download complete 09:40:34 7a1cb9ad7f75 Downloading [==========================> ] 61.62MB/115.2MB 09:40:34 6ca01427385e Extracting [=========> ] 11.7MB/61.48MB 09:40:34 806be17e856d Extracting [=========> ] 16.15MB/89.72MB 09:40:35 00b33c871d26 Downloading [===================================> ] 181.5MB/253.3MB 09:40:35 00b33c871d26 Downloading [===================================> ] 181.5MB/253.3MB 09:40:35 7138254c3790 Extracting [==================================> ] 22.71MB/32.98MB 09:40:35 ac9f4de4b762 Extracting [=============> ] 13.11MB/50.13MB 09:40:35 0a92c7dea7af Downloading [==========> ] 719B/3.449kB 09:40:35 0a92c7dea7af Downloading [==================================================>] 3.449kB/3.449kB 09:40:35 0a92c7dea7af Verifying Checksum 09:40:35 0a92c7dea7af Download complete 09:40:35 7a1cb9ad7f75 Downloading [=============================> ] 67.53MB/115.2MB 09:40:35 22ebf0e44c85 Pull complete 09:40:35 22ebf0e44c85 Pull complete 09:40:35 f469048fbe8d Pull complete 09:40:35 257d54e26411 Pull complete 09:40:35 00b33c871d26 Downloading [====================================> ] 187.4MB/253.3MB 09:40:35 00b33c871d26 Downloading [====================================> ] 187.4MB/253.3MB 09:40:35 7a1cb9ad7f75 Downloading [=============================> ] 68.07MB/115.2MB 09:40:35 7138254c3790 Extracting [===================================> ] 23.43MB/32.98MB 09:40:35 6ca01427385e Extracting [=============> ] 16.15MB/61.48MB 09:40:35 ac9f4de4b762 Extracting [=============> ] 13.63MB/50.13MB 09:40:35 806be17e856d Extracting [==========> ] 19.5MB/89.72MB 09:40:35 00b33c871d26 Downloading [======================================> ] 194.3MB/253.3MB 09:40:35 00b33c871d26 Downloading [======================================> ] 194.3MB/253.3MB 09:40:35 7a1cb9ad7f75 Downloading [=================================> ] 76.1MB/115.2MB 09:40:35 806be17e856d Extracting [===========> ] 21.17MB/89.72MB 09:40:35 ac9f4de4b762 Extracting [===============> ] 15.2MB/50.13MB 09:40:35 6ca01427385e Extracting [===============> ] 18.94MB/61.48MB 09:40:35 7138254c3790 Extracting [====================================> ] 24.15MB/32.98MB 09:40:35 00b33c871d26 Downloading [======================================> ] 197.5MB/253.3MB 09:40:35 00b33c871d26 Downloading [======================================> ] 197.5MB/253.3MB 09:40:35 7a1cb9ad7f75 Downloading [=================================> ] 78.25MB/115.2MB 09:40:35 6ca01427385e Extracting [===================> ] 23.95MB/61.48MB 09:40:35 806be17e856d Extracting [=============> ] 23.4MB/89.72MB 09:40:35 ac9f4de4b762 Extracting [=================> ] 17.83MB/50.13MB 09:40:35 7138254c3790 Extracting [========================================> ] 27.03MB/32.98MB 09:40:35 00b33c871d26 Downloading [=======================================> ] 201.3MB/253.3MB 09:40:35 00b33c871d26 Downloading [=======================================> ] 201.3MB/253.3MB 09:40:35 7a1cb9ad7f75 Downloading [==================================> ] 79.85MB/115.2MB 09:40:35 6ca01427385e Extracting [======================> ] 27.3MB/61.48MB 09:40:35 806be17e856d Extracting [==============> ] 25.62MB/89.72MB 09:40:35 c189e028fabb Extracting [==================================================>] 300B/300B 09:40:35 c189e028fabb Extracting [==================================================>] 300B/300B 09:40:35 ac9f4de4b762 Extracting [====================> ] 20.97MB/50.13MB 09:40:35 7138254c3790 Extracting [===========================================> ] 28.84MB/32.98MB 09:40:35 00b33c871d26 Downloading [=========================================> ] 208.3MB/253.3MB 09:40:35 00b33c871d26 Downloading [=========================================> ] 208.3MB/253.3MB 09:40:35 7a1cb9ad7f75 Downloading [====================================> ] 83.04MB/115.2MB 09:40:35 6ca01427385e Extracting [=======================> ] 28.97MB/61.48MB 09:40:35 ac9f4de4b762 Extracting [=======================> ] 23.59MB/50.13MB 09:40:35 00b33c871d26 Downloading [===========================================> ] 219.5MB/253.3MB 09:40:35 00b33c871d26 Downloading [===========================================> ] 219.5MB/253.3MB 09:40:35 7138254c3790 Extracting [============================================> ] 29.56MB/32.98MB 09:40:35 806be17e856d Extracting [==============> ] 26.18MB/89.72MB 09:40:35 7a1cb9ad7f75 Downloading [=====================================> ] 85.72MB/115.2MB 09:40:35 215302b53935 Extracting [==================================================>] 293B/293B 09:40:35 215302b53935 Extracting [==================================================>] 293B/293B 09:40:35 6ca01427385e Extracting [=========================> ] 31.75MB/61.48MB 09:40:35 ac9f4de4b762 Extracting [=========================> ] 25.17MB/50.13MB 09:40:35 806be17e856d Extracting [===============> ] 27.85MB/89.72MB 09:40:35 00b33c871d26 Downloading [============================================> ] 227MB/253.3MB 09:40:35 00b33c871d26 Downloading [============================================> ] 227MB/253.3MB 09:40:35 7138254c3790 Extracting [==============================================> ] 30.64MB/32.98MB 09:40:35 7a1cb9ad7f75 Downloading [=======================================> ] 91.62MB/115.2MB 09:40:36 ac9f4de4b762 Extracting [=====================================> ] 37.75MB/50.13MB 09:40:36 7138254c3790 Extracting [==================================================>] 32.98MB/32.98MB 09:40:36 6ca01427385e Extracting [============================> ] 35.09MB/61.48MB 09:40:36 806be17e856d Extracting [================> ] 30.08MB/89.72MB 09:40:36 00b33c871d26 Downloading [===============================================> ] 238.8MB/253.3MB 09:40:36 00b33c871d26 Downloading [===============================================> ] 238.8MB/253.3MB 09:40:36 7a1cb9ad7f75 Downloading [==========================================> ] 98.05MB/115.2MB 09:40:36 ac9f4de4b762 Extracting [=========================================> ] 41.94MB/50.13MB 09:40:36 6ca01427385e Extracting [=============================> ] 36.77MB/61.48MB 09:40:36 7a1cb9ad7f75 Downloading [===========================================> ] 100.2MB/115.2MB 09:40:36 00b33c871d26 Downloading [================================================> ] 247.9MB/253.3MB 09:40:36 00b33c871d26 Downloading [================================================> ] 247.9MB/253.3MB 09:40:36 806be17e856d Extracting [==================> ] 32.31MB/89.72MB 09:40:36 ac9f4de4b762 Extracting [============================================> ] 44.56MB/50.13MB 09:40:36 00b33c871d26 Verifying Checksum 09:40:36 00b33c871d26 Download complete 09:40:36 00b33c871d26 Verifying Checksum 09:40:36 00b33c871d26 Download complete 09:40:36 6ca01427385e Extracting [==============================> ] 37.88MB/61.48MB 09:40:36 7a1cb9ad7f75 Downloading [============================================> ] 102.9MB/115.2MB 09:40:36 806be17e856d Extracting [==================> ] 33.42MB/89.72MB 09:40:36 ac9f4de4b762 Extracting [=================================================> ] 49.81MB/50.13MB 09:40:36 6ca01427385e Extracting [=================================> ] 40.67MB/61.48MB 09:40:36 7a1cb9ad7f75 Downloading [=================================================> ] 113.1MB/115.2MB 09:40:36 ac9f4de4b762 Extracting [==================================================>] 50.13MB/50.13MB 09:40:36 806be17e856d Extracting [====================> ] 36.77MB/89.72MB 09:40:36 7a1cb9ad7f75 Verifying Checksum 09:40:36 7a1cb9ad7f75 Download complete 09:40:36 6ca01427385e Extracting [===================================> ] 44.01MB/61.48MB 09:40:36 806be17e856d Extracting [======================> ] 40.11MB/89.72MB 09:40:36 00b33c871d26 Extracting [> ] 557.1kB/253.3MB 09:40:36 00b33c871d26 Extracting [> ] 557.1kB/253.3MB 09:40:36 6ca01427385e Extracting [======================================> ] 47.35MB/61.48MB 09:40:36 00b33c871d26 Extracting [=> ] 8.913MB/253.3MB 09:40:36 00b33c871d26 Extracting [=> ] 8.913MB/253.3MB 09:40:36 806be17e856d Extracting [=======================> ] 41.78MB/89.72MB 09:40:36 6ca01427385e Extracting [=======================================> ] 48.46MB/61.48MB 09:40:36 215302b53935 Pull complete 09:40:36 7138254c3790 Pull complete 09:40:36 c189e028fabb Pull complete 09:40:36 ac9f4de4b762 Pull complete 09:40:36 806be17e856d Extracting [=======================> ] 42.89MB/89.72MB 09:40:36 00b33c871d26 Extracting [==> ] 15.04MB/253.3MB 09:40:36 00b33c871d26 Extracting [==> ] 15.04MB/253.3MB 09:40:36 6ca01427385e Extracting [========================================> ] 50.14MB/61.48MB 09:40:36 806be17e856d Extracting [========================> ] 44.56MB/89.72MB 09:40:36 78f39bed0e83 Extracting [==================================================>] 1.077kB/1.077kB 09:40:36 78f39bed0e83 Extracting [==================================================>] 1.077kB/1.077kB 09:40:36 ea63b2e6315f Extracting [==================================================>] 605B/605B 09:40:36 ea63b2e6315f Extracting [==================================================>] 605B/605B 09:40:37 806be17e856d Extracting [=========================> ] 45.68MB/89.72MB 09:40:37 00b33c871d26 Extracting [====> ] 24.51MB/253.3MB 09:40:37 00b33c871d26 Extracting [====> ] 24.51MB/253.3MB 09:40:37 6ca01427385e Extracting [===========================================> ] 53.48MB/61.48MB 09:40:37 6ca01427385e Extracting [============================================> ] 55.15MB/61.48MB 09:40:37 806be17e856d Extracting [==========================> ] 47.35MB/89.72MB 09:40:37 00b33c871d26 Extracting [=====> ] 27.85MB/253.3MB 09:40:37 00b33c871d26 Extracting [=====> ] 27.85MB/253.3MB 09:40:37 6ca01427385e Extracting [==============================================> ] 57.38MB/61.48MB 09:40:37 806be17e856d Extracting [===========================> ] 50.14MB/89.72MB 09:40:37 00b33c871d26 Extracting [=====> ] 28.97MB/253.3MB 09:40:37 00b33c871d26 Extracting [=====> ] 28.97MB/253.3MB 09:40:37 6ca01427385e Extracting [================================================> ] 59.05MB/61.48MB 09:40:37 806be17e856d Extracting [==============================> ] 55.15MB/89.72MB 09:40:37 00b33c871d26 Extracting [======> ] 34.54MB/253.3MB 09:40:37 00b33c871d26 Extracting [======> ] 34.54MB/253.3MB 09:40:37 c9bd119720e4 Extracting [> ] 557.1kB/246.3MB 09:40:37 eb2f448c7730 Extracting [============> ] 32.77kB/127kB 09:40:37 eb2f448c7730 Extracting [==================================================>] 127kB/127kB 09:40:37 eb2f448c7730 Extracting [==================================================>] 127kB/127kB 09:40:37 6ca01427385e Extracting [================================================> ] 59.6MB/61.48MB 09:40:37 00b33c871d26 Extracting [========> ] 40.67MB/253.3MB 09:40:37 00b33c871d26 Extracting [========> ] 40.67MB/253.3MB 09:40:37 806be17e856d Extracting [================================> ] 57.93MB/89.72MB 09:40:37 c9bd119720e4 Extracting [> ] 1.114MB/246.3MB 09:40:37 6ca01427385e Extracting [==================================================>] 61.48MB/61.48MB 09:40:37 00b33c871d26 Extracting [=========> ] 49.02MB/253.3MB 09:40:37 00b33c871d26 Extracting [=========> ] 49.02MB/253.3MB 09:40:37 806be17e856d Extracting [=================================> ] 60.72MB/89.72MB 09:40:38 c9bd119720e4 Extracting [=> ] 9.47MB/246.3MB 09:40:38 00b33c871d26 Extracting [==========> ] 54.03MB/253.3MB 09:40:38 00b33c871d26 Extracting [==========> ] 54.03MB/253.3MB 09:40:38 c9bd119720e4 Extracting [==> ] 11.14MB/246.3MB 09:40:38 806be17e856d Extracting [====================================> ] 64.62MB/89.72MB 09:40:38 00b33c871d26 Extracting [============> ] 64.06MB/253.3MB 09:40:38 00b33c871d26 Extracting [============> ] 64.06MB/253.3MB 09:40:38 c9bd119720e4 Extracting [===> ] 17.27MB/246.3MB 09:40:38 806be17e856d Extracting [=====================================> ] 67.4MB/89.72MB 09:40:38 00b33c871d26 Extracting [=============> ] 67.96MB/253.3MB 09:40:38 00b33c871d26 Extracting [=============> ] 67.96MB/253.3MB 09:40:38 c9bd119720e4 Extracting [====> ] 22.28MB/246.3MB 09:40:38 806be17e856d Extracting [======================================> ] 69.63MB/89.72MB 09:40:38 00b33c871d26 Extracting [==============> ] 74.09MB/253.3MB 09:40:38 00b33c871d26 Extracting [==============> ] 74.09MB/253.3MB 09:40:38 c9bd119720e4 Extracting [======> ] 30.64MB/246.3MB 09:40:38 806be17e856d Extracting [========================================> ] 72.97MB/89.72MB 09:40:38 00b33c871d26 Extracting [================> ] 83MB/253.3MB 09:40:38 00b33c871d26 Extracting [================> ] 83MB/253.3MB 09:40:38 c9bd119720e4 Extracting [========> ] 40.11MB/246.3MB 09:40:38 806be17e856d Extracting [=========================================> ] 74.65MB/89.72MB 09:40:38 00b33c871d26 Extracting [=================> ] 88.57MB/253.3MB 09:40:38 00b33c871d26 Extracting [=================> ] 88.57MB/253.3MB 09:40:38 c9bd119720e4 Extracting [=========> ] 47.91MB/246.3MB 09:40:38 78f39bed0e83 Pull complete 09:40:38 806be17e856d Extracting [===========================================> ] 78.54MB/89.72MB 09:40:38 00b33c871d26 Extracting [===================> ] 99.16MB/253.3MB 09:40:38 00b33c871d26 Extracting [===================> ] 99.16MB/253.3MB 09:40:38 c9bd119720e4 Extracting [===========> ] 56.26MB/246.3MB 09:40:38 00b33c871d26 Extracting [====================> ] 105.3MB/253.3MB 09:40:38 00b33c871d26 Extracting [====================> ] 105.3MB/253.3MB 09:40:39 ea63b2e6315f Pull complete 09:40:39 806be17e856d Extracting [============================================> ] 80.22MB/89.72MB 09:40:39 c9bd119720e4 Extracting [============> ] 61.28MB/246.3MB 09:40:39 00b33c871d26 Extracting [=====================> ] 107MB/253.3MB 09:40:39 00b33c871d26 Extracting [=====================> ] 107MB/253.3MB 09:40:40 c9bd119720e4 Extracting [=============> ] 64.06MB/246.3MB 09:40:40 00b33c871d26 Extracting [=====================> ] 110.3MB/253.3MB 09:40:40 00b33c871d26 Extracting [=====================> ] 110.3MB/253.3MB 09:40:40 806be17e856d Extracting [==============================================> ] 83MB/89.72MB 09:40:40 c9bd119720e4 Extracting [==============> ] 72.42MB/246.3MB 09:40:40 00b33c871d26 Extracting [======================> ] 115.3MB/253.3MB 09:40:40 00b33c871d26 Extracting [======================> ] 115.3MB/253.3MB 09:40:40 806be17e856d Extracting [===============================================> ] 84.67MB/89.72MB 09:40:40 c9bd119720e4 Extracting [================> ] 83MB/246.3MB 09:40:40 00b33c871d26 Extracting [=======================> ] 119.2MB/253.3MB 09:40:40 00b33c871d26 Extracting [=======================> ] 119.2MB/253.3MB 09:40:40 806be17e856d Extracting [================================================> ] 86.9MB/89.72MB 09:40:40 c9bd119720e4 Extracting [==================> ] 91.91MB/246.3MB 09:40:40 00b33c871d26 Extracting [========================> ] 123.1MB/253.3MB 09:40:40 00b33c871d26 Extracting [========================> ] 123.1MB/253.3MB 09:40:40 c9bd119720e4 Extracting [====================> ] 101.9MB/246.3MB 09:40:40 00b33c871d26 Extracting [=========================> ] 128.1MB/253.3MB 09:40:40 00b33c871d26 Extracting [=========================> ] 128.1MB/253.3MB 09:40:40 806be17e856d Extracting [=================================================> ] 89.13MB/89.72MB 09:40:40 806be17e856d Extracting [==================================================>] 89.72MB/89.72MB 09:40:40 c9bd119720e4 Extracting [======================> ] 112.5MB/246.3MB 09:40:40 00b33c871d26 Extracting [=========================> ] 130.9MB/253.3MB 09:40:40 00b33c871d26 Extracting [=========================> ] 130.9MB/253.3MB 09:40:40 c9bd119720e4 Extracting [=========================> ] 123.7MB/246.3MB 09:40:40 00b33c871d26 Extracting [==========================> ] 135.9MB/253.3MB 09:40:40 00b33c871d26 Extracting [==========================> ] 135.9MB/253.3MB 09:40:40 40796999d308 Extracting [==================================================>] 5.325kB/5.325kB 09:40:40 40796999d308 Extracting [==================================================>] 5.325kB/5.325kB 09:40:40 c9bd119720e4 Extracting [============================> ] 138.7MB/246.3MB 09:40:40 00b33c871d26 Extracting [===========================> ] 140.9MB/253.3MB 09:40:40 00b33c871d26 Extracting [===========================> ] 140.9MB/253.3MB 09:40:41 c9bd119720e4 Extracting [=============================> ] 144.3MB/246.3MB 09:40:41 00b33c871d26 Extracting [===========================> ] 141.5MB/253.3MB 09:40:41 00b33c871d26 Extracting [===========================> ] 141.5MB/253.3MB 09:40:41 00b33c871d26 Extracting [============================> ] 144.3MB/253.3MB 09:40:41 00b33c871d26 Extracting [============================> ] 144.3MB/253.3MB 09:40:41 c9bd119720e4 Extracting [=============================> ] 144.8MB/246.3MB 09:40:41 00b33c871d26 Extracting [=============================> ] 148.7MB/253.3MB 09:40:41 00b33c871d26 Extracting [=============================> ] 148.7MB/253.3MB 09:40:41 c9bd119720e4 Extracting [================================> ] 158.2MB/246.3MB 09:40:41 00b33c871d26 Extracting [==============================> ] 152.1MB/253.3MB 09:40:41 00b33c871d26 Extracting [==============================> ] 152.1MB/253.3MB 09:40:41 c9bd119720e4 Extracting [=================================> ] 167.1MB/246.3MB 09:40:41 c9bd119720e4 Extracting [====================================> ] 181.6MB/246.3MB 09:40:41 00b33c871d26 Extracting [==============================> ] 155.4MB/253.3MB 09:40:41 00b33c871d26 Extracting [==============================> ] 155.4MB/253.3MB 09:40:42 c9bd119720e4 Extracting [=======================================> ] 192.7MB/246.3MB 09:40:42 00b33c871d26 Extracting [===============================> ] 159.9MB/253.3MB 09:40:42 00b33c871d26 Extracting [===============================> ] 159.9MB/253.3MB 09:40:42 c9bd119720e4 Extracting [=======================================> ] 193.9MB/246.3MB 09:40:42 00b33c871d26 Extracting [===============================> ] 160.4MB/253.3MB 09:40:42 00b33c871d26 Extracting [===============================> ] 160.4MB/253.3MB 09:40:42 c9bd119720e4 Extracting [========================================> ] 201.1MB/246.3MB 09:40:42 00b33c871d26 Extracting [================================> ] 164.9MB/253.3MB 09:40:42 00b33c871d26 Extracting [================================> ] 164.9MB/253.3MB 09:40:42 c9bd119720e4 Extracting [==========================================> ] 210MB/246.3MB 09:40:42 00b33c871d26 Extracting [=================================> ] 169.3MB/253.3MB 09:40:42 00b33c871d26 Extracting [=================================> ] 169.3MB/253.3MB 09:40:42 c9bd119720e4 Extracting [=============================================> ] 221.7MB/246.3MB 09:40:42 00b33c871d26 Extracting [=================================> ] 170.5MB/253.3MB 09:40:42 00b33c871d26 Extracting [=================================> ] 170.5MB/253.3MB 09:40:42 eb2f448c7730 Pull complete 09:40:42 6ca01427385e Pull complete 09:40:42 fbd390d3bd00 Extracting [==================================================>] 2.675kB/2.675kB 09:40:42 fbd390d3bd00 Extracting [==================================================>] 2.675kB/2.675kB 09:40:42 c9bd119720e4 Extracting [===============================================> ] 231.7MB/246.3MB 09:40:42 00b33c871d26 Extracting [=================================> ] 171.6MB/253.3MB 09:40:42 00b33c871d26 Extracting [=================================> ] 171.6MB/253.3MB 09:40:43 c9bd119720e4 Extracting [===============================================> ] 235.1MB/246.3MB 09:40:43 00b33c871d26 Extracting [=================================> ] 172.1MB/253.3MB 09:40:43 00b33c871d26 Extracting [=================================> ] 172.1MB/253.3MB 09:40:43 c9bd119720e4 Extracting [================================================> ] 239MB/246.3MB 09:40:43 00b33c871d26 Extracting [==================================> ] 173.2MB/253.3MB 09:40:43 00b33c871d26 Extracting [==================================> ] 173.2MB/253.3MB 09:40:43 c9bd119720e4 Extracting [==================================================>] 246.3MB/246.3MB 09:40:43 00b33c871d26 Extracting [==================================> ] 174.9MB/253.3MB 09:40:43 00b33c871d26 Extracting [==================================> ] 174.9MB/253.3MB 09:40:43 00b33c871d26 Extracting [===================================> ] 178.8MB/253.3MB 09:40:43 00b33c871d26 Extracting [===================================> ] 178.8MB/253.3MB 09:40:43 00b33c871d26 Extracting [====================================> ] 185.5MB/253.3MB 09:40:43 00b33c871d26 Extracting [====================================> ] 185.5MB/253.3MB 09:40:43 00b33c871d26 Extracting [====================================> ] 187.2MB/253.3MB 09:40:43 00b33c871d26 Extracting [====================================> ] 187.2MB/253.3MB 09:40:43 00b33c871d26 Extracting [=====================================> ] 188.3MB/253.3MB 09:40:43 00b33c871d26 Extracting [=====================================> ] 188.3MB/253.3MB 09:40:44 00b33c871d26 Extracting [=====================================> ] 189.4MB/253.3MB 09:40:44 00b33c871d26 Extracting [=====================================> ] 189.4MB/253.3MB 09:40:44 00b33c871d26 Extracting [======================================> ] 192.7MB/253.3MB 09:40:44 00b33c871d26 Extracting [======================================> ] 192.7MB/253.3MB 09:40:45 00b33c871d26 Extracting [======================================> ] 193.9MB/253.3MB 09:40:45 00b33c871d26 Extracting [======================================> ] 193.9MB/253.3MB 09:40:45 c8ee90c58894 Extracting [==================================================>] 1.329kB/1.329kB 09:40:45 c8ee90c58894 Extracting [==================================================>] 1.329kB/1.329kB 09:40:45 00b33c871d26 Extracting [======================================> ] 196.1MB/253.3MB 09:40:45 00b33c871d26 Extracting [======================================> ] 196.1MB/253.3MB 09:40:45 00b33c871d26 Extracting [=======================================> ] 197.8MB/253.3MB 09:40:45 00b33c871d26 Extracting [=======================================> ] 197.8MB/253.3MB 09:40:45 00b33c871d26 Extracting [=======================================> ] 199.4MB/253.3MB 09:40:45 00b33c871d26 Extracting [=======================================> ] 199.4MB/253.3MB 09:40:45 00b33c871d26 Extracting [=======================================> ] 200.5MB/253.3MB 09:40:45 00b33c871d26 Extracting [=======================================> ] 200.5MB/253.3MB 09:40:45 806be17e856d Pull complete 09:40:45 40796999d308 Pull complete 09:40:45 fbd390d3bd00 Pull complete 09:40:45 c9bd119720e4 Pull complete 09:40:45 14ddc757aae0 Extracting [==================================================>] 5.314kB/5.314kB 09:40:45 14ddc757aae0 Extracting [==================================================>] 5.314kB/5.314kB 09:40:45 c8ee90c58894 Pull complete 09:40:45 634de6c90876 Extracting [==================================================>] 3.49kB/3.49kB 09:40:45 634de6c90876 Extracting [==================================================>] 3.49kB/3.49kB 09:40:45 9b1ac15ef728 Extracting [==================================================>] 3.087kB/3.087kB 09:40:45 9b1ac15ef728 Extracting [==================================================>] 3.087kB/3.087kB 09:40:45 00b33c871d26 Extracting [========================================> ] 202.8MB/253.3MB 09:40:45 00b33c871d26 Extracting [========================================> ] 202.8MB/253.3MB 09:40:45 apex-pdp Pulled 09:40:46 e35e8e85e24d Extracting [> ] 524.3kB/50.55MB 09:40:46 e30cdb86c4f0 Extracting [> ] 557.1kB/98.32MB 09:40:46 00b33c871d26 Extracting [========================================> ] 204.4MB/253.3MB 09:40:46 00b33c871d26 Extracting [========================================> ] 204.4MB/253.3MB 09:40:46 e30cdb86c4f0 Extracting [======> ] 12.81MB/98.32MB 09:40:46 00b33c871d26 Extracting [========================================> ] 205.6MB/253.3MB 09:40:46 00b33c871d26 Extracting [========================================> ] 205.6MB/253.3MB 09:40:46 e30cdb86c4f0 Extracting [===========> ] 23.4MB/98.32MB 09:40:46 00b33c871d26 Extracting [========================================> ] 207.2MB/253.3MB 09:40:46 00b33c871d26 Extracting [========================================> ] 207.2MB/253.3MB 09:40:46 e35e8e85e24d Extracting [=> ] 1.573MB/50.55MB 09:40:46 e30cdb86c4f0 Extracting [===============> ] 30.64MB/98.32MB 09:40:46 e30cdb86c4f0 Extracting [================> ] 32.87MB/98.32MB 09:40:46 00b33c871d26 Extracting [=========================================> ] 209.5MB/253.3MB 09:40:46 00b33c871d26 Extracting [=========================================> ] 209.5MB/253.3MB 09:40:46 e30cdb86c4f0 Extracting [====================> ] 40.11MB/98.32MB 09:40:46 e30cdb86c4f0 Extracting [=========================> ] 50.69MB/98.32MB 09:40:46 e30cdb86c4f0 Extracting [================================> ] 63.5MB/98.32MB 09:40:47 e30cdb86c4f0 Extracting [=======================================> ] 77.43MB/98.32MB 09:40:47 e30cdb86c4f0 Extracting [===============================================> ] 93.03MB/98.32MB 09:40:47 e30cdb86c4f0 Extracting [==================================================>] 98.32MB/98.32MB 09:40:47 00b33c871d26 Extracting [=========================================> ] 211.1MB/253.3MB 09:40:47 00b33c871d26 Extracting [=========================================> ] 211.1MB/253.3MB 09:40:47 e35e8e85e24d Extracting [==> ] 2.097MB/50.55MB 09:40:47 00b33c871d26 Extracting [==========================================> ] 213.9MB/253.3MB 09:40:47 00b33c871d26 Extracting [==========================================> ] 213.9MB/253.3MB 09:40:47 9b1ac15ef728 Pull complete 09:40:47 634de6c90876 Pull complete 09:40:47 14ddc757aae0 Pull complete 09:40:47 00b33c871d26 Extracting [==========================================> ] 214.5MB/253.3MB 09:40:47 00b33c871d26 Extracting [==========================================> ] 214.5MB/253.3MB 09:40:47 00b33c871d26 Extracting [==========================================> ] 216.1MB/253.3MB 09:40:47 00b33c871d26 Extracting [==========================================> ] 216.1MB/253.3MB 09:40:48 00b33c871d26 Extracting [==========================================> ] 216.7MB/253.3MB 09:40:48 00b33c871d26 Extracting [==========================================> ] 216.7MB/253.3MB 09:40:48 e35e8e85e24d Extracting [===> ] 3.67MB/50.55MB 09:40:48 e30cdb86c4f0 Pull complete 09:40:48 8682f304eb80 Extracting [==================================================>] 4.023kB/4.023kB 09:40:48 8682f304eb80 Extracting [==================================================>] 4.023kB/4.023kB 09:40:48 00b33c871d26 Extracting [==========================================> ] 217.8MB/253.3MB 09:40:48 00b33c871d26 Extracting [==========================================> ] 217.8MB/253.3MB 09:40:48 ebe1cd824584 Extracting [==================================================>] 1.037kB/1.037kB 09:40:48 ebe1cd824584 Extracting [==================================================>] 1.037kB/1.037kB 09:40:48 00b33c871d26 Extracting [===========================================> ] 218.4MB/253.3MB 09:40:48 00b33c871d26 Extracting [===========================================> ] 218.4MB/253.3MB 09:40:48 e35e8e85e24d Extracting [====> ] 4.194MB/50.55MB 09:40:48 e35e8e85e24d Extracting [====> ] 4.719MB/50.55MB 09:40:48 00b33c871d26 Extracting [===========================================> ] 218.9MB/253.3MB 09:40:48 00b33c871d26 Extracting [===========================================> ] 218.9MB/253.3MB 09:40:48 00b33c871d26 Extracting [===========================================> ] 219.5MB/253.3MB 09:40:48 00b33c871d26 Extracting [===========================================> ] 219.5MB/253.3MB 09:40:48 e35e8e85e24d Extracting [======> ] 6.816MB/50.55MB 09:40:49 00b33c871d26 Extracting [===========================================> ] 220.6MB/253.3MB 09:40:49 00b33c871d26 Extracting [===========================================> ] 220.6MB/253.3MB 09:40:49 e35e8e85e24d Extracting [=======> ] 7.864MB/50.55MB 09:40:49 c990b7e46fc8 Extracting [==================================================>] 1.299kB/1.299kB 09:40:49 c990b7e46fc8 Extracting [==================================================>] 1.299kB/1.299kB 09:40:49 cd00854cfb1a Extracting [==================================================>] 6.971kB/6.971kB 09:40:49 cd00854cfb1a Extracting [==================================================>] 6.971kB/6.971kB 09:40:49 00b33c871d26 Extracting [===========================================> ] 222.3MB/253.3MB 09:40:49 00b33c871d26 Extracting [===========================================> ] 222.3MB/253.3MB 09:40:49 e35e8e85e24d Extracting [========> ] 8.913MB/50.55MB 09:40:49 e35e8e85e24d Extracting [==========> ] 10.49MB/50.55MB 09:40:49 00b33c871d26 Extracting [============================================> ] 224.5MB/253.3MB 09:40:49 00b33c871d26 Extracting [============================================> ] 224.5MB/253.3MB 09:40:49 00b33c871d26 Extracting [============================================> ] 227.8MB/253.3MB 09:40:49 00b33c871d26 Extracting [============================================> ] 227.8MB/253.3MB 09:40:49 e35e8e85e24d Extracting [===========> ] 11.53MB/50.55MB 09:40:49 00b33c871d26 Extracting [=============================================> ] 229.5MB/253.3MB 09:40:49 00b33c871d26 Extracting [=============================================> ] 229.5MB/253.3MB 09:40:49 e35e8e85e24d Extracting [===============> ] 15.73MB/50.55MB 09:40:49 00b33c871d26 Extracting [=============================================> ] 232.3MB/253.3MB 09:40:49 00b33c871d26 Extracting [=============================================> ] 232.3MB/253.3MB 09:40:49 e35e8e85e24d Extracting [==================> ] 18.87MB/50.55MB 09:40:50 00b33c871d26 Extracting [==============================================> ] 235.1MB/253.3MB 09:40:50 00b33c871d26 Extracting [==============================================> ] 235.1MB/253.3MB 09:40:50 e35e8e85e24d Extracting [=====================> ] 22.02MB/50.55MB 09:40:50 00b33c871d26 Extracting [===============================================> ] 240.6MB/253.3MB 09:40:50 00b33c871d26 Extracting [===============================================> ] 240.6MB/253.3MB 09:40:50 e35e8e85e24d Extracting [========================> ] 24.64MB/50.55MB 09:40:50 8682f304eb80 Pull complete 09:40:50 e35e8e85e24d Extracting [==========================> ] 26.74MB/50.55MB 09:40:50 00b33c871d26 Extracting [===============================================> ] 241.2MB/253.3MB 09:40:50 00b33c871d26 Extracting [===============================================> ] 241.2MB/253.3MB 09:40:50 e35e8e85e24d Extracting [============================> ] 28.84MB/50.55MB 09:40:50 00b33c871d26 Extracting [=================================================> ] 249MB/253.3MB 09:40:50 00b33c871d26 Extracting [=================================================> ] 249MB/253.3MB 09:40:51 00b33c871d26 Extracting [=================================================> ] 250.7MB/253.3MB 09:40:51 00b33c871d26 Extracting [=================================================> ] 250.7MB/253.3MB 09:40:51 e35e8e85e24d Extracting [==============================> ] 30.93MB/50.55MB 09:40:51 ebe1cd824584 Pull complete 09:40:51 5fbafe078afc Extracting [==================================================>] 1.44kB/1.44kB 09:40:51 5fbafe078afc Extracting [==================================================>] 1.44kB/1.44kB 09:40:51 00b33c871d26 Extracting [=================================================> ] 251.2MB/253.3MB 09:40:51 00b33c871d26 Extracting [=================================================> ] 251.2MB/253.3MB 09:40:51 e35e8e85e24d Extracting [===============================> ] 31.46MB/50.55MB 09:40:51 00b33c871d26 Extracting [==================================================>] 253.3MB/253.3MB 09:40:51 00b33c871d26 Extracting [==================================================>] 253.3MB/253.3MB 09:40:51 e35e8e85e24d Extracting [=================================> ] 34.08MB/50.55MB 09:40:51 e35e8e85e24d Extracting [====================================> ] 36.7MB/50.55MB 09:40:52 e35e8e85e24d Extracting [=======================================> ] 40.37MB/50.55MB 09:40:52 e35e8e85e24d Extracting [==========================================> ] 42.99MB/50.55MB 09:40:52 d2893dc6732f Extracting [==================================================>] 1.038kB/1.038kB 09:40:52 d2893dc6732f Extracting [==================================================>] 1.038kB/1.038kB 09:40:53 e35e8e85e24d Extracting [================================================> ] 48.76MB/50.55MB 09:40:53 cd00854cfb1a Pull complete 09:40:53 c990b7e46fc8 Pull complete 09:40:53 e35e8e85e24d Extracting [==================================================>] 50.55MB/50.55MB 09:40:55 00b33c871d26 Pull complete 09:40:55 00b33c871d26 Pull complete 09:40:55 5fbafe078afc Pull complete 09:40:56 d2893dc6732f Pull complete 09:40:56 e35e8e85e24d Pull complete 09:40:58 6b11e56702ad Extracting [> ] 98.3kB/7.707MB 09:40:58 6b11e56702ad Extracting [> ] 98.3kB/7.707MB 09:40:58 6b11e56702ad Extracting [====================================> ] 5.702MB/7.707MB 09:40:58 6b11e56702ad Extracting [====================================> ] 5.702MB/7.707MB 09:40:58 6b11e56702ad Extracting [================================================> ] 7.471MB/7.707MB 09:40:58 6b11e56702ad Extracting [================================================> ] 7.471MB/7.707MB 09:40:58 6b11e56702ad Extracting [==================================================>] 7.707MB/7.707MB 09:40:58 6b11e56702ad Extracting [==================================================>] 7.707MB/7.707MB 09:40:59 7fb53fd2ae10 Extracting [===========> ] 32.77kB/138kB 09:40:59 7fb53fd2ae10 Extracting [==================================================>] 138kB/138kB 09:40:59 7fb53fd2ae10 Extracting [==================================================>] 138kB/138kB 09:40:59 mariadb Pulled 09:41:02 6b11e56702ad Pull complete 09:41:02 6b11e56702ad Pull complete 09:41:02 a23a963fcebe Extracting [==================================================>] 13.9kB/13.9kB 09:41:02 a23a963fcebe Extracting [==================================================>] 13.9kB/13.9kB 09:41:02 pap Pulled 09:41:03 d0bef95bc6b2 Extracting [==================================================>] 11.92kB/11.92kB 09:41:03 d0bef95bc6b2 Extracting [==================================================>] 11.92kB/11.92kB 09:41:03 7fb53fd2ae10 Pull complete 09:41:04 53d69aa7d3fc Extracting [==================================================>] 19.96kB/19.96kB 09:41:04 53d69aa7d3fc Extracting [==================================================>] 19.96kB/19.96kB 09:41:04 53d69aa7d3fc Extracting [==================================================>] 19.96kB/19.96kB 09:41:05 a23a963fcebe Pull complete 09:41:05 592798bd3683 Extracting [==================================================>] 100B/100B 09:41:05 592798bd3683 Extracting [==================================================>] 100B/100B 09:41:06 d0bef95bc6b2 Pull complete 09:41:07 369dfa39565e Extracting [==================================================>] 13.79kB/13.79kB 09:41:07 369dfa39565e Extracting [==================================================>] 13.79kB/13.79kB 09:41:09 53d69aa7d3fc Pull complete 09:41:09 53d69aa7d3fc Pull complete 09:41:09 af860903a445 Extracting [==================================================>] 1.226kB/1.226kB 09:41:09 af860903a445 Extracting [==================================================>] 1.226kB/1.226kB 09:41:10 592798bd3683 Pull complete 09:41:11 a3ab11953ef9 Extracting [> ] 426kB/39.52MB 09:41:11 a3ab11953ef9 Extracting [> ] 426kB/39.52MB 09:41:11 a3ab11953ef9 Extracting [============> ] 9.798MB/39.52MB 09:41:11 a3ab11953ef9 Extracting [============> ] 9.798MB/39.52MB 09:41:11 a3ab11953ef9 Extracting [======================> ] 17.47MB/39.52MB 09:41:11 a3ab11953ef9 Extracting [======================> ] 17.47MB/39.52MB 09:41:12 a3ab11953ef9 Extracting [=================================> ] 26.84MB/39.52MB 09:41:12 a3ab11953ef9 Extracting [=================================> ] 26.84MB/39.52MB 09:41:12 a3ab11953ef9 Extracting [=================================================> ] 38.76MB/39.52MB 09:41:12 a3ab11953ef9 Extracting [=================================================> ] 38.76MB/39.52MB 09:41:12 a3ab11953ef9 Extracting [==================================================>] 39.52MB/39.52MB 09:41:12 a3ab11953ef9 Extracting [==================================================>] 39.52MB/39.52MB 09:41:14 369dfa39565e Pull complete 09:41:14 473fdc983780 Extracting [==================================================>] 721B/721B 09:41:14 473fdc983780 Extracting [==================================================>] 721B/721B 09:41:17 af860903a445 Pull complete 09:41:18 9146eb587aa8 Extracting [==================================================>] 2.856kB/2.856kB 09:41:18 9146eb587aa8 Extracting [==================================================>] 2.856kB/2.856kB 09:41:21 a3ab11953ef9 Pull complete 09:41:21 a3ab11953ef9 Pull complete 09:41:23 473fdc983780 Pull complete 09:41:23 91ef9543149d Extracting [==================================================>] 1.101kB/1.101kB 09:41:23 91ef9543149d Extracting [==================================================>] 1.101kB/1.101kB 09:41:23 91ef9543149d Extracting [==================================================>] 1.101kB/1.101kB 09:41:23 91ef9543149d Extracting [==================================================>] 1.101kB/1.101kB 09:41:23 9146eb587aa8 Pull complete 09:41:25 a120f6888c1f Extracting [==================================================>] 2.864kB/2.864kB 09:41:25 a120f6888c1f Extracting [==================================================>] 2.864kB/2.864kB 09:41:26 91ef9543149d Pull complete 09:41:26 91ef9543149d Pull complete 09:41:26 grafana Pulled 09:41:27 2ec4f59af178 Extracting [==================================================>] 881B/881B 09:41:27 2ec4f59af178 Extracting [==================================================>] 881B/881B 09:41:27 2ec4f59af178 Extracting [==================================================>] 881B/881B 09:41:27 2ec4f59af178 Extracting [==================================================>] 881B/881B 09:41:34 prometheus Pulled 09:41:36 a120f6888c1f Pull complete 09:41:36 2ec4f59af178 Pull complete 09:41:36 2ec4f59af178 Pull complete 09:41:36 8b7e81cd5ef1 Extracting [==================================================>] 131B/131B 09:41:36 8b7e81cd5ef1 Extracting [==================================================>] 131B/131B 09:41:36 8b7e81cd5ef1 Extracting [==================================================>] 131B/131B 09:41:36 8b7e81cd5ef1 Extracting [==================================================>] 131B/131B 09:41:37 policy-db-migrator Pulled 09:41:39 8b7e81cd5ef1 Pull complete 09:41:39 8b7e81cd5ef1 Pull complete 09:41:41 c52916c1316e Extracting [==================================================>] 171B/171B 09:41:41 c52916c1316e Extracting [==================================================>] 171B/171B 09:41:41 c52916c1316e Extracting [==================================================>] 171B/171B 09:41:41 c52916c1316e Extracting [==================================================>] 171B/171B 09:41:44 c52916c1316e Pull complete 09:41:44 c52916c1316e Pull complete 09:41:44 d93f69e96600 Extracting [> ] 557.1kB/115.2MB 09:41:44 7a1cb9ad7f75 Extracting [> ] 557.1kB/115.2MB 09:41:44 d93f69e96600 Extracting [======> ] 13.93MB/115.2MB 09:41:44 7a1cb9ad7f75 Extracting [====> ] 10.58MB/115.2MB 09:41:44 d93f69e96600 Extracting [==========> ] 25.07MB/115.2MB 09:41:44 7a1cb9ad7f75 Extracting [==========> ] 25.07MB/115.2MB 09:41:44 d93f69e96600 Extracting [==============> ] 33.98MB/115.2MB 09:41:44 7a1cb9ad7f75 Extracting [================> ] 37.32MB/115.2MB 09:41:44 d93f69e96600 Extracting [====================> ] 47.35MB/115.2MB 09:41:44 7a1cb9ad7f75 Extracting [=======================> ] 54.03MB/115.2MB 09:41:44 d93f69e96600 Extracting [===========================> ] 62.39MB/115.2MB 09:41:45 7a1cb9ad7f75 Extracting [=============================> ] 69.07MB/115.2MB 09:41:45 d93f69e96600 Extracting [==================================> ] 79.1MB/115.2MB 09:41:45 7a1cb9ad7f75 Extracting [====================================> ] 83.56MB/115.2MB 09:41:45 d93f69e96600 Extracting [========================================> ] 93.03MB/115.2MB 09:41:45 7a1cb9ad7f75 Extracting [==========================================> ] 98.04MB/115.2MB 09:41:45 d93f69e96600 Extracting [==============================================> ] 107MB/115.2MB 09:41:45 7a1cb9ad7f75 Extracting [================================================> ] 110.9MB/115.2MB 09:41:45 7a1cb9ad7f75 Extracting [==================================================>] 115.2MB/115.2MB 09:41:45 d93f69e96600 Extracting [=================================================> ] 113.6MB/115.2MB 09:41:45 d93f69e96600 Extracting [==================================================>] 115.2MB/115.2MB 09:41:45 7a1cb9ad7f75 Pull complete 09:41:45 d93f69e96600 Pull complete 09:41:45 0a92c7dea7af Extracting [==================================================>] 3.449kB/3.449kB 09:41:45 0a92c7dea7af Extracting [==================================================>] 3.449kB/3.449kB 09:41:45 bbb9d15c45a1 Extracting [==================================================>] 3.633kB/3.633kB 09:41:45 bbb9d15c45a1 Extracting [==================================================>] 3.633kB/3.633kB 09:41:45 bbb9d15c45a1 Pull complete 09:41:45 0a92c7dea7af Pull complete 09:41:46 kafka Pulled 09:41:46 zookeeper Pulled 09:41:46 Network compose_default Creating 09:41:46 Network compose_default Created 09:41:46 Container zookeeper Creating 09:41:46 Container prometheus Creating 09:41:46 Container simulator Creating 09:41:46 Container mariadb Creating 09:41:49 Container mariadb Created 09:41:49 Container zookeeper Created 09:41:49 Container policy-db-migrator Creating 09:41:49 Container kafka Creating 09:41:49 Container prometheus Created 09:41:49 Container grafana Creating 09:41:49 Container simulator Created 09:41:50 Container grafana Created 09:41:50 Container policy-db-migrator Created 09:41:50 Container policy-api Creating 09:41:50 Container kafka Created 09:41:50 Container policy-api Created 09:41:50 Container policy-pap Creating 09:41:50 Container policy-pap Created 09:41:50 Container policy-apex-pdp Creating 09:41:50 Container policy-apex-pdp Created 09:41:50 Container mariadb Starting 09:41:50 Container zookeeper Starting 09:41:50 Container prometheus Starting 09:41:50 Container simulator Starting 09:41:52 Container zookeeper Started 09:41:52 Container kafka Starting 09:41:53 Container mariadb Started 09:41:53 Container policy-db-migrator Starting 09:41:53 Container policy-db-migrator Started 09:41:53 Container policy-api Starting 09:41:54 Container policy-api Started 09:41:55 Container kafka Started 09:41:55 Container policy-pap Starting 09:41:56 Container simulator Started 09:41:57 Container prometheus Started 09:41:57 Container grafana Starting 09:42:00 Container grafana Started 09:42:00 Container policy-pap Started 09:42:00 Container policy-apex-pdp Starting 09:42:00 Container policy-apex-pdp Started 09:42:00 Prometheus server: http://localhost:30259 09:42:00 Grafana server: http://localhost:30269 09:42:10 Waiting for REST to come up on localhost port 30003... 09:42:10 NAMES STATUS 09:42:10 policy-apex-pdp Up 10 seconds 09:42:10 policy-pap Up 10 seconds 09:42:10 policy-api Up 15 seconds 09:42:10 kafka Up 15 seconds 09:42:10 policy-db-migrator Up 16 seconds 09:42:10 grafana Up 10 seconds 09:42:10 zookeeper Up 18 seconds 09:42:10 simulator Up 14 seconds 09:42:10 mariadb Up 17 seconds 09:42:10 prometheus Up 13 seconds 09:42:15 NAMES STATUS 09:42:15 policy-apex-pdp Up 15 seconds 09:42:15 policy-pap Up 15 seconds 09:42:15 policy-api Up 20 seconds 09:42:15 kafka Up 20 seconds 09:42:15 grafana Up 15 seconds 09:42:15 zookeeper Up 23 seconds 09:42:15 simulator Up 19 seconds 09:42:15 mariadb Up 22 seconds 09:42:15 prometheus Up 18 seconds 09:42:20 NAMES STATUS 09:42:20 policy-apex-pdp Up 20 seconds 09:42:20 policy-pap Up 21 seconds 09:42:20 policy-api Up 25 seconds 09:42:20 kafka Up 25 seconds 09:42:20 grafana Up 21 seconds 09:42:20 zookeeper Up 28 seconds 09:42:20 simulator Up 24 seconds 09:42:20 mariadb Up 27 seconds 09:42:20 prometheus Up 23 seconds 09:42:25 NAMES STATUS 09:42:25 policy-apex-pdp Up 25 seconds 09:42:25 policy-pap Up 26 seconds 09:42:25 policy-api Up 31 seconds 09:42:25 kafka Up 30 seconds 09:42:25 grafana Up 26 seconds 09:42:25 zookeeper Up 33 seconds 09:42:25 simulator Up 29 seconds 09:42:25 mariadb Up 32 seconds 09:42:25 prometheus Up 28 seconds 09:42:30 NAMES STATUS 09:42:30 policy-apex-pdp Up 30 seconds 09:42:30 policy-pap Up 31 seconds 09:42:30 policy-api Up 36 seconds 09:42:30 kafka Up 35 seconds 09:42:30 grafana Up 31 seconds 09:42:30 zookeeper Up 38 seconds 09:42:30 simulator Up 34 seconds 09:42:30 mariadb Up 37 seconds 09:42:30 prometheus Up 33 seconds 09:42:35 NAMES STATUS 09:42:35 policy-apex-pdp Up 35 seconds 09:42:35 policy-pap Up 36 seconds 09:42:35 policy-api Up 41 seconds 09:42:35 kafka Up 40 seconds 09:42:35 grafana Up 36 seconds 09:42:35 zookeeper Up 43 seconds 09:42:35 simulator Up 39 seconds 09:42:35 mariadb Up 42 seconds 09:42:35 prometheus Up 38 seconds 09:42:40 NAMES STATUS 09:42:40 policy-apex-pdp Up 40 seconds 09:42:40 policy-pap Up 41 seconds 09:42:40 policy-api Up 46 seconds 09:42:40 kafka Up 45 seconds 09:42:40 grafana Up 41 seconds 09:42:40 zookeeper Up 48 seconds 09:42:40 simulator Up 44 seconds 09:42:40 mariadb Up 47 seconds 09:42:40 prometheus Up 43 seconds 09:42:41 Waiting for REST to come up on localhost port 30001... 09:42:41 NAMES STATUS 09:42:41 policy-apex-pdp Up 40 seconds 09:42:41 policy-pap Up 41 seconds 09:42:41 policy-api Up 46 seconds 09:42:41 kafka Up 45 seconds 09:42:41 grafana Up 41 seconds 09:42:41 zookeeper Up 48 seconds 09:42:41 simulator Up 44 seconds 09:42:41 mariadb Up 47 seconds 09:42:41 prometheus Up 43 seconds 09:43:01 Build docker image for robot framework 09:43:01 Error: No such image: policy-csit-robot 09:43:01 Cloning into '/w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp/csit/resources/tests/models'... 09:43:02 Build robot framework docker image 09:43:02 Sending build context to Docker daemon 16.49MB 09:43:02 Step 1/9 : FROM nexus3.onap.org:10001/library/python:3.10-slim-bullseye 09:43:02 3.10-slim-bullseye: Pulling from library/python 09:43:02 76956b537f14: Pulling fs layer 09:43:02 f75f1b8a4051: Pulling fs layer 09:43:02 f9adc358e0b8: Pulling fs layer 09:43:02 f66e101ef41f: Pulling fs layer 09:43:02 b913137adf9e: Pulling fs layer 09:43:02 f66e101ef41f: Waiting 09:43:02 f75f1b8a4051: Download complete 09:43:02 f66e101ef41f: Download complete 09:43:02 b913137adf9e: Verifying Checksum 09:43:02 b913137adf9e: Download complete 09:43:02 f9adc358e0b8: Verifying Checksum 09:43:02 f9adc358e0b8: Download complete 09:43:03 76956b537f14: Verifying Checksum 09:43:03 76956b537f14: Download complete 09:43:04 76956b537f14: Pull complete 09:43:04 f75f1b8a4051: Pull complete 09:43:04 f9adc358e0b8: Pull complete 09:43:04 f66e101ef41f: Pull complete 09:43:05 b913137adf9e: Pull complete 09:43:05 Digest: sha256:46193e24d7f1f03f4e2f9e21e1a5f8361ac29c83db447b4d5355fae9445943b0 09:43:05 Status: Downloaded newer image for nexus3.onap.org:10001/library/python:3.10-slim-bullseye 09:43:05 ---> 08150e0479fc 09:43:05 Step 2/9 : ARG CSIT_SCRIPT=${CSIT_SCRIPT} 09:43:07 ---> Running in 6f9b847d93a6 09:43:07 Removing intermediate container 6f9b847d93a6 09:43:07 ---> c7b3686e60f5 09:43:07 Step 3/9 : ARG ROBOT_FILE=${ROBOT_FILE} 09:43:08 ---> Running in b76f00cd71b9 09:43:08 Removing intermediate container b76f00cd71b9 09:43:08 ---> b8b7a6a037d8 09:43:08 Step 4/9 : ENV ROBOT_WORKSPACE=/opt/robotworkspace ROBOT_FILE=$ROBOT_FILE TEST_ENV=$TEST_ENV 09:43:08 ---> Running in 46c5e1c0bcf3 09:43:08 Removing intermediate container 46c5e1c0bcf3 09:43:08 ---> e5c84f1876be 09:43:08 Step 5/9 : RUN python3 -m pip -qq install --upgrade pip && python3 -m pip -qq install --upgrade --extra-index-url="https://nexus3.onap.org/repository/PyPi.staging/simple" 'robotframework-onap==0.6.0.*' --pre && python3 -m pip -qq install --upgrade confluent-kafka && python3 -m pip freeze 09:43:08 ---> Running in fe2ce6c1d680 09:43:22 bcrypt==4.1.3 09:43:22 certifi==2024.7.4 09:43:22 cffi==1.17.0rc1 09:43:22 charset-normalizer==3.3.2 09:43:22 confluent-kafka==2.4.0 09:43:22 cryptography==42.0.8 09:43:22 decorator==5.1.1 09:43:22 deepdiff==7.0.1 09:43:22 dnspython==2.6.1 09:43:22 future==1.0.0 09:43:22 idna==3.7 09:43:22 Jinja2==3.1.4 09:43:22 jsonpath-rw==1.4.0 09:43:22 kafka-python==2.0.2 09:43:22 MarkupSafe==2.1.5 09:43:22 more-itertools==5.0.0 09:43:22 ordered-set==4.1.0 09:43:22 paramiko==3.4.0 09:43:22 pbr==6.0.0 09:43:22 ply==3.11 09:43:22 protobuf==5.27.2 09:43:22 pycparser==2.22 09:43:22 PyNaCl==1.5.0 09:43:22 PyYAML==6.0.2rc1 09:43:22 requests==2.32.3 09:43:22 robotframework==7.0.1 09:43:22 robotframework-onap==0.6.0.dev105 09:43:22 robotframework-requests==1.0a11 09:43:22 robotlibcore-temp==1.0.2 09:43:22 six==1.16.0 09:43:22 urllib3==2.2.2 09:43:27 Removing intermediate container fe2ce6c1d680 09:43:27 ---> 19c6e87509b1 09:43:27 Step 6/9 : RUN mkdir -p ${ROBOT_WORKSPACE} 09:43:27 ---> Running in 554a8b2b1bd0 09:43:28 Removing intermediate container 554a8b2b1bd0 09:43:28 ---> 5cb913fe3c7c 09:43:28 Step 7/9 : COPY scripts/run-test.sh tests/ ${ROBOT_WORKSPACE}/ 09:43:30 ---> e2153ab89116 09:43:30 Step 8/9 : WORKDIR ${ROBOT_WORKSPACE} 09:43:30 ---> Running in bc235c2d82d6 09:43:31 Removing intermediate container bc235c2d82d6 09:43:31 ---> 5aec88b37feb 09:43:31 Step 9/9 : CMD ["sh", "-c", "./run-test.sh" ] 09:43:31 ---> Running in a11b12712588 09:43:31 Removing intermediate container a11b12712588 09:43:31 ---> 4afb00ba0ed7 09:43:31 Successfully built 4afb00ba0ed7 09:43:31 Successfully tagged policy-csit-robot:latest 09:43:34 top - 09:43:34 up 5 min, 0 users, load average: 2.45, 2.06, 0.91 09:43:34 Tasks: 208 total, 1 running, 131 sleeping, 0 stopped, 0 zombie 09:43:34 %Cpu(s): 11.6 us, 2.7 sy, 0.0 ni, 75.6 id, 9.9 wa, 0.0 hi, 0.1 si, 0.1 st 09:43:34 09:43:34 total used free shared buff/cache available 09:43:34 Mem: 31G 2.7G 22G 1.3M 6.5G 28G 09:43:34 Swap: 1.0G 0B 1.0G 09:43:34 09:43:34 NAMES STATUS 09:43:34 policy-apex-pdp Up About a minute 09:43:34 policy-pap Up About a minute 09:43:34 policy-api Up About a minute 09:43:34 kafka Up About a minute 09:43:34 grafana Up About a minute 09:43:34 zookeeper Up About a minute 09:43:34 simulator Up About a minute 09:43:34 mariadb Up About a minute 09:43:34 prometheus Up About a minute 09:43:34 09:43:36 CONTAINER ID NAME CPU % MEM USAGE / LIMIT MEM % NET I/O BLOCK I/O PIDS 09:43:36 78aaf3c409fe policy-apex-pdp 0.38% 183.1MiB / 31.41GiB 0.57% 37.7kB / 54.9kB 0B / 0B 50 09:43:36 882188c4979b policy-pap 0.80% 494.3MiB / 31.41GiB 1.54% 129kB / 155kB 0B / 149MB 64 09:43:36 6c4b8e865b32 policy-api 0.10% 474.2MiB / 31.41GiB 1.47% 990kB / 674kB 0B / 0B 54 09:43:36 81de5e5f37ca kafka 1.93% 390MiB / 31.41GiB 1.21% 163kB / 156kB 0B / 561kB 85 09:43:36 4a040a44f63b grafana 0.05% 57.52MiB / 31.41GiB 0.18% 24.6kB / 4.67kB 0B / 26.7MB 19 09:43:36 ba21d07ef1db zookeeper 0.10% 99.83MiB / 31.41GiB 0.31% 58.3kB / 50.2kB 4.1kB / 369kB 60 09:43:36 e280d07f2941 simulator 0.06% 124.4MiB / 31.41GiB 0.39% 1.38kB / 0B 28.7kB / 0B 78 09:43:36 d7e787a4da9c mariadb 0.02% 102.8MiB / 31.41GiB 0.32% 970kB / 1.22MB 11MB / 71.9MB 29 09:43:36 800b11c39ebc prometheus 0.24% 20.87MiB / 31.41GiB 0.06% 67.7kB / 3.36kB 0B / 0B 13 09:43:36 09:43:36 time="2024-07-04T09:43:36Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:43:37 Container policy-csit Creating 09:43:37 Container policy-csit Created 09:43:37 Attaching to policy-csit 09:43:38 policy-csit | Invoking the robot tests from: apex-pdp-test.robot apex-slas.robot 09:43:38 policy-csit | Run Robot test 09:43:38 policy-csit | ROBOT_VARIABLES=-v DATA:/opt/robotworkspace/models/models-examples/src/main/resources/policies 09:43:38 policy-csit | -v NODETEMPLATES:/opt/robotworkspace/models/models-examples/src/main/resources/nodetemplates 09:43:38 policy-csit | -v POLICY_API_IP:policy-api:6969 09:43:38 policy-csit | -v POLICY_RUNTIME_ACM_IP:policy-clamp-runtime-acm:6969 09:43:38 policy-csit | -v POLICY_PARTICIPANT_SIM_IP:policy-clamp-ac-sim-ppnt:6969 09:43:38 policy-csit | -v POLICY_PAP_IP:policy-pap:6969 09:43:38 policy-csit | -v APEX_IP:policy-apex-pdp:6969 09:43:38 policy-csit | -v APEX_EVENTS_IP:policy-apex-pdp:23324 09:43:38 policy-csit | -v KAFKA_IP:kafka:9092 09:43:38 policy-csit | -v PROMETHEUS_IP:prometheus:9090 09:43:38 policy-csit | -v POLICY_PDPX_IP:policy-xacml-pdp:6969 09:43:38 policy-csit | -v POLICY_DROOLS_IP:policy-drools-pdp:9696 09:43:38 policy-csit | -v DROOLS_IP:policy-drools-apps:6969 09:43:38 policy-csit | -v DROOLS_IP_2:policy-drools-apps:9696 09:43:38 policy-csit | -v TEMP_FOLDER:/tmp/distribution 09:43:38 policy-csit | -v DISTRIBUTION_IP:policy-distribution:6969 09:43:38 policy-csit | -v TEST_ENV: 09:43:38 policy-csit | -v JAEGER_IP:jaeger:16686 09:43:38 policy-csit | Starting Robot test suites ... 09:43:38 policy-csit | ============================================================================== 09:43:38 policy-csit | Apex-Pdp-Test & Apex-Slas 09:43:38 policy-csit | ============================================================================== 09:43:38 policy-csit | Apex-Pdp-Test & Apex-Slas.Apex-Pdp-Test 09:43:38 policy-csit | ============================================================================== 09:43:38 policy-csit | Healthcheck :: Runs Apex PDP Health check | PASS | 09:43:38 policy-csit | ------------------------------------------------------------------------------ 09:43:39 policy-csit | ExecuteApexSampleDomainPolicy | FAIL | 09:43:39 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:43:39 policy-csit | ------------------------------------------------------------------------------ 09:43:40 policy-csit | ExecuteApexTestPnfPolicy | FAIL | 09:43:40 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:43:40 policy-csit | ------------------------------------------------------------------------------ 09:43:40 policy-csit | ExecuteApexTestPnfPolicyWithMetadataSet | FAIL | 09:43:40 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:43:40 policy-csit | ------------------------------------------------------------------------------ 09:43:40 policy-csit | Metrics :: Verify policy-apex-pdp is exporting prometheus metrics | FAIL | 09:43:40 policy-csit | '# HELP jvm_classes_currently_loaded The number of classes that are currently loaded in the JVM 09:43:40 policy-csit | # TYPE jvm_classes_currently_loaded gauge 09:43:40 policy-csit | jvm_classes_currently_loaded 7533.0 09:43:40 policy-csit | # HELP jvm_classes_loaded_total The total number of classes that have been loaded since the JVM has started execution 09:43:40 policy-csit | # TYPE jvm_classes_loaded_total counter 09:43:40 policy-csit | jvm_classes_loaded_total 7533.0 09:43:40 policy-csit | # HELP jvm_classes_unloaded_total The total number of classes that have been unloaded since the JVM has started execution 09:43:40 policy-csit | # TYPE jvm_classes_unloaded_total counter 09:43:40 policy-csit | jvm_classes_unloaded_total 0.0 09:43:40 policy-csit | # HELP process_cpu_seconds_total Total user and system CPU time spent in seconds. 09:43:40 policy-csit | # TYPE process_cpu_seconds_total counter 09:43:40 policy-csit | process_cpu_seconds_total 7.72 09:43:40 policy-csit | # HELP process_start_time_seconds Start time of the process since unix epoch in seconds. 09:43:40 policy-csit | # TYPE process_start_time_seconds gauge 09:43:40 policy-csit | process_start_time_seconds 1.720086157998E9 09:43:40 policy-csit | [ Message content over the limit has been removed. ] 09:43:40 policy-csit | # TYPE pdpa_policy_deployments_total counter 09:43:40 policy-csit | # HELP jvm_memory_pool_allocated_bytes_created Total bytes allocated in a given JVM memory pool. Only updated after GC, not continuously. 09:43:40 policy-csit | # TYPE jvm_memory_pool_allocated_bytes_created gauge 09:43:40 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="CodeHeap 'profiled nmethods'",} 1.720086159595E9 09:43:40 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="G1 Old Gen",} 1.720086159613E9 09:43:40 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="G1 Eden Space",} 1.720086159613E9 09:43:40 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="CodeHeap 'non-profiled nmethods'",} 1.720086159613E9 09:43:40 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="G1 Survivor Space",} 1.720086159613E9 09:43:40 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="Compressed Class Space",} 1.720086159613E9 09:43:40 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="Metaspace",} 1.720086159613E9 09:43:40 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="CodeHeap 'non-nmethods'",} 1.720086159613E9 09:43:40 policy-csit | ' does not contain 'pdpa_policy_deployments_total{operation="deploy",status="TOTAL",} 3.0' 09:43:40 policy-csit | ------------------------------------------------------------------------------ 09:43:40 policy-csit | Apex-Pdp-Test & Apex-Slas.Apex-Pdp-Test | FAIL | 09:43:40 policy-csit | 5 tests, 1 passed, 4 failed 09:43:40 policy-csit | ============================================================================== 09:43:40 policy-csit | Apex-Pdp-Test & Apex-Slas.Apex-Slas 09:43:40 policy-csit | ============================================================================== 09:43:40 policy-csit | Healthcheck :: Runs Apex PDP Health check | PASS | 09:43:40 policy-csit | ------------------------------------------------------------------------------ 09:43:41 policy-csit | ValidatePolicyExecutionAndEventRateLowComplexity :: Validate that ... | FAIL | 09:43:41 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:43:41 policy-csit | ------------------------------------------------------------------------------ 09:43:41 policy-csit | ValidatePolicyExecutionAndEventRateModerateComplexity :: Validate ... | FAIL | 09:43:41 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:43:41 policy-csit | ------------------------------------------------------------------------------ 09:43:42 policy-csit | ValidatePolicyExecutionAndEventRateHighComplexity :: Validate that... | FAIL | 09:43:42 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:43:42 policy-csit | ------------------------------------------------------------------------------ 09:44:42 policy-csit | WaitForPrometheusServer :: Sleep time to wait for Prometheus serve... | PASS | 09:44:42 policy-csit | ------------------------------------------------------------------------------ 09:44:42 policy-csit | ValidatePolicyExecutionTimes :: Validate policy execution times us... | FAIL | 09:44:42 policy-csit | Resolving variable '${resp['data']['result'][0]['value'][1]}' failed: IndexError: list index out of range 09:44:42 policy-csit | ------------------------------------------------------------------------------ 09:44:42 policy-csit | Apex-Pdp-Test & Apex-Slas.Apex-Slas | FAIL | 09:44:42 policy-csit | 6 tests, 2 passed, 4 failed 09:44:42 policy-csit | ============================================================================== 09:44:42 policy-csit | Apex-Pdp-Test & Apex-Slas | FAIL | 09:44:42 policy-csit | 11 tests, 3 passed, 8 failed 09:44:42 policy-csit | ============================================================================== 09:44:42 policy-csit | Output: /tmp/results/output.xml 09:44:42 policy-csit | Log: /tmp/results/log.html 09:44:42 policy-csit | Report: /tmp/results/report.html 09:44:42 policy-csit | RESULT: 8 09:44:42 policy-csit exited with code 8 09:44:42 NAMES STATUS 09:44:42 policy-apex-pdp Up 2 minutes 09:44:42 policy-pap Up 2 minutes 09:44:42 policy-api Up 2 minutes 09:44:42 kafka Up 2 minutes 09:44:42 grafana Up 2 minutes 09:44:42 zookeeper Up 2 minutes 09:44:42 simulator Up 2 minutes 09:44:42 mariadb Up 2 minutes 09:44:42 prometheus Up 2 minutes 09:44:42 Shut down started! 09:44:44 Collecting logs from docker compose containers... 09:44:44 time="2024-07-04T09:44:44Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:44 time="2024-07-04T09:44:44Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:44 time="2024-07-04T09:44:44Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:45 time="2024-07-04T09:44:45Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:45 time="2024-07-04T09:44:45Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:45 time="2024-07-04T09:44:45Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:46 time="2024-07-04T09:44:46Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:46 time="2024-07-04T09:44:46Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:46 time="2024-07-04T09:44:46Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:47 time="2024-07-04T09:44:47Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:47 time="2024-07-04T09:44:47Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:47 time="2024-07-04T09:44:47Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:47 ======== Logs from grafana ======== 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.769755729Z level=info msg="Starting Grafana" version=11.1.0 commit=5b85c4c2fcf5d32d4f68aaef345c53096359b2f1 branch=HEAD compiled=2024-07-04T09:41:59Z 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770005843Z level=info msg="Config loaded from" file=/usr/share/grafana/conf/defaults.ini 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770012383Z level=info msg="Config loaded from" file=/etc/grafana/grafana.ini 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770015973Z level=info msg="Config overridden from command line" arg="default.paths.data=/var/lib/grafana" 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770018743Z level=info msg="Config overridden from command line" arg="default.paths.logs=/var/log/grafana" 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770024083Z level=info msg="Config overridden from command line" arg="default.paths.plugins=/var/lib/grafana/plugins" 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770026883Z level=info msg="Config overridden from command line" arg="default.paths.provisioning=/etc/grafana/provisioning" 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770029973Z level=info msg="Config overridden from command line" arg="default.log.mode=console" 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770033263Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_DATA=/var/lib/grafana" 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770036073Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_LOGS=/var/log/grafana" 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770038653Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_PLUGINS=/var/lib/grafana/plugins" 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770041233Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_PROVISIONING=/etc/grafana/provisioning" 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770043873Z level=info msg=Target target=[all] 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770049063Z level=info msg="Path Home" path=/usr/share/grafana 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770051593Z level=info msg="Path Data" path=/var/lib/grafana 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770054043Z level=info msg="Path Logs" path=/var/log/grafana 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770060143Z level=info msg="Path Plugins" path=/var/lib/grafana/plugins 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770062824Z level=info msg="Path Provisioning" path=/etc/grafana/provisioning 09:44:47 grafana | logger=settings t=2024-07-04T09:41:59.770065494Z level=info msg="App mode production" 09:44:47 grafana | logger=featuremgmt t=2024-07-04T09:41:59.770327388Z level=info msg=FeatureToggles recordedQueriesMulti=true cloudWatchCrossAccountQuerying=true awsAsyncQueryCaching=true correlations=true transformationsRedesign=true alertingSimplifiedRouting=true lokiQueryHints=true annotationPermissionUpdate=true lokiMetricDataplane=true kubernetesPlaylists=true cloudWatchNewLabelParsing=true logsExploreTableVisualisation=true logRowsPopoverMenu=true logsInfiniteScrolling=true nestedFolders=true topnav=true managedPluginsInstall=true alertingInsights=true prometheusConfigOverhaulAuth=true panelMonitoring=true dashgpt=true exploreContentOutline=true prometheusMetricEncyclopedia=true recoveryThreshold=true dataplaneFrontendFallback=true exploreMetrics=true lokiStructuredMetadata=true logsContextDatasourceUi=true ssoSettingsApi=true lokiQuerySplitting=true angularDeprecationUI=true prometheusDataplane=true alertingNoDataErrorExecution=true influxdbBackendMigration=true publicDashboards=true awsDatasourcesNewFormStyling=true betterPageScrolling=true 09:44:47 grafana | logger=sqlstore t=2024-07-04T09:41:59.770370248Z level=info msg="Connecting to DB" dbtype=sqlite3 09:44:47 grafana | logger=sqlstore t=2024-07-04T09:41:59.770382648Z level=info msg="Creating SQLite database file" path=/var/lib/grafana/grafana.db 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.771717119Z level=info msg="Locking database" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.771729099Z level=info msg="Starting DB migrations" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.772235337Z level=info msg="Executing migration" id="create migration_log table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.77311325Z level=info msg="Migration successfully executed" id="create migration_log table" duration=877.543µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.810567008Z level=info msg="Executing migration" id="create user table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.81132783Z level=info msg="Migration successfully executed" id="create user table" duration=760.262µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.816467409Z level=info msg="Executing migration" id="add unique index user.login" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.817935922Z level=info msg="Migration successfully executed" id="add unique index user.login" duration=1.469373ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.822684965Z level=info msg="Executing migration" id="add unique index user.email" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.824195198Z level=info msg="Migration successfully executed" id="add unique index user.email" duration=1.513313ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.827607081Z level=info msg="Executing migration" id="drop index UQE_user_login - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.82884458Z level=info msg="Migration successfully executed" id="drop index UQE_user_login - v1" duration=1.229909ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.834113461Z level=info msg="Executing migration" id="drop index UQE_user_email - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.834838232Z level=info msg="Migration successfully executed" id="drop index UQE_user_email - v1" duration=725.371µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.841697948Z level=info msg="Executing migration" id="Rename table user to user_v1 - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.845796571Z level=info msg="Migration successfully executed" id="Rename table user to user_v1 - v1" duration=4.098373ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.851741843Z level=info msg="Executing migration" id="create user table v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.852542945Z level=info msg="Migration successfully executed" id="create user table v2" duration=801.272µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.856640028Z level=info msg="Executing migration" id="create index UQE_user_login - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.85740094Z level=info msg="Migration successfully executed" id="create index UQE_user_login - v2" duration=762.042µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.884946855Z level=info msg="Executing migration" id="create index UQE_user_email - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.885976221Z level=info msg="Migration successfully executed" id="create index UQE_user_email - v2" duration=1.033196ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.891337524Z level=info msg="Executing migration" id="copy data_source v1 to v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.891703569Z level=info msg="Migration successfully executed" id="copy data_source v1 to v2" duration=366.496µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.912757994Z level=info msg="Executing migration" id="Drop old table user_v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.914196236Z level=info msg="Migration successfully executed" id="Drop old table user_v1" duration=1.440693ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.920656095Z level=info msg="Executing migration" id="Add column help_flags1 to user table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.923469809Z level=info msg="Migration successfully executed" id="Add column help_flags1 to user table" duration=2.812864ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.929526182Z level=info msg="Executing migration" id="Update user table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.929571093Z level=info msg="Migration successfully executed" id="Update user table charset" duration=47.751µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.936005242Z level=info msg="Executing migration" id="Add last_seen_at column to user" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.936924057Z level=info msg="Migration successfully executed" id="Add last_seen_at column to user" duration=919.035µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.941810812Z level=info msg="Executing migration" id="Add missing user data" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.941984304Z level=info msg="Migration successfully executed" id="Add missing user data" duration=173.602µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.945530049Z level=info msg="Executing migration" id="Add is_disabled column to user" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.946366112Z level=info msg="Migration successfully executed" id="Add is_disabled column to user" duration=835.923µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.949558451Z level=info msg="Executing migration" id="Add index user.login/user.email" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.950072209Z level=info msg="Migration successfully executed" id="Add index user.login/user.email" duration=513.788µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.956147193Z level=info msg="Executing migration" id="Add is_service_account column to user" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.956989156Z level=info msg="Migration successfully executed" id="Add is_service_account column to user" duration=839.643µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.964990689Z level=info msg="Executing migration" id="Update is_service_account column to nullable" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.970793109Z level=info msg="Migration successfully executed" id="Update is_service_account column to nullable" duration=5.79748ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.976265133Z level=info msg="Executing migration" id="Add uid column to user" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.977071605Z level=info msg="Migration successfully executed" id="Add uid column to user" duration=806.592µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.987178281Z level=info msg="Executing migration" id="Update uid column values for users" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.987799421Z level=info msg="Migration successfully executed" id="Update uid column values for users" duration=621.71µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.993235215Z level=info msg="Executing migration" id="Add unique index user_uid" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.994621246Z level=info msg="Migration successfully executed" id="Add unique index user_uid" duration=1.387181ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.999303808Z level=info msg="Executing migration" id="update login field with orgid to allow for multiple service accounts with same name across orgs" 09:44:47 grafana | logger=migrator t=2024-07-04T09:41:59.999607703Z level=info msg="Migration successfully executed" id="update login field with orgid to allow for multiple service accounts with same name across orgs" duration=302.375µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.003089667Z level=info msg="Executing migration" id="update login and email fields to lowercase" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.003649935Z level=info msg="Migration successfully executed" id="update login and email fields to lowercase" duration=563.048µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.007494714Z level=info msg="Executing migration" id="update login and email fields to lowercase2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.007779307Z level=info msg="Migration successfully executed" id="update login and email fields to lowercase2" duration=285.063µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.010905238Z level=info msg="Executing migration" id="create temp user table v1-7" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.01177913Z level=info msg="Migration successfully executed" id="create temp user table v1-7" duration=873.752µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.017671706Z level=info msg="Executing migration" id="create index IDX_temp_user_email - v1-7" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.018365935Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_email - v1-7" duration=693.939µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.023145767Z level=info msg="Executing migration" id="create index IDX_temp_user_org_id - v1-7" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.023823796Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_org_id - v1-7" duration=678.279µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.031156912Z level=info msg="Executing migration" id="create index IDX_temp_user_code - v1-7" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.03182665Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_code - v1-7" duration=669.248µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.036371179Z level=info msg="Executing migration" id="create index IDX_temp_user_status - v1-7" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.037024218Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_status - v1-7" duration=655.259µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.043049786Z level=info msg="Executing migration" id="Update temp_user table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.043076646Z level=info msg="Migration successfully executed" id="Update temp_user table charset" duration=28.19µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.047260811Z level=info msg="Executing migration" id="drop index IDX_temp_user_email - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.047918809Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_email - v1" duration=664.658µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.051570947Z level=info msg="Executing migration" id="drop index IDX_temp_user_org_id - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.052241266Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_org_id - v1" duration=635.788µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.056194627Z level=info msg="Executing migration" id="drop index IDX_temp_user_code - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.056834285Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_code - v1" duration=639.618µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.059691492Z level=info msg="Executing migration" id="drop index IDX_temp_user_status - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.06031313Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_status - v1" duration=621.518µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.064334833Z level=info msg="Executing migration" id="Rename table temp_user to temp_user_tmp_qwerty - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.067467123Z level=info msg="Migration successfully executed" id="Rename table temp_user to temp_user_tmp_qwerty - v1" duration=3.13206ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.074411384Z level=info msg="Executing migration" id="create temp_user v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.075925253Z level=info msg="Migration successfully executed" id="create temp_user v2" duration=1.51642ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.079919035Z level=info msg="Executing migration" id="create index IDX_temp_user_email - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.081020879Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_email - v2" duration=1.101624ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.083890926Z level=info msg="Executing migration" id="create index IDX_temp_user_org_id - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.084577385Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_org_id - v2" duration=685.989µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.090233428Z level=info msg="Executing migration" id="create index IDX_temp_user_code - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.091028409Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_code - v2" duration=794.721µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.093319498Z level=info msg="Executing migration" id="create index IDX_temp_user_status - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.094063868Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_status - v2" duration=744.27µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.097061257Z level=info msg="Executing migration" id="copy temp_user v1 to v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.097520803Z level=info msg="Migration successfully executed" id="copy temp_user v1 to v2" duration=458.966µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.100299499Z level=info msg="Executing migration" id="drop temp_user_tmp_qwerty" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.100908226Z level=info msg="Migration successfully executed" id="drop temp_user_tmp_qwerty" duration=606.857µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.106728002Z level=info msg="Executing migration" id="Set created for temp users that will otherwise prematurely expire" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.1073621Z level=info msg="Migration successfully executed" id="Set created for temp users that will otherwise prematurely expire" duration=625.368µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.110653712Z level=info msg="Executing migration" id="create star table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.111392612Z level=info msg="Migration successfully executed" id="create star table" duration=738.28µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.114234349Z level=info msg="Executing migration" id="add unique index star.user_id_dashboard_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.115004349Z level=info msg="Migration successfully executed" id="add unique index star.user_id_dashboard_id" duration=769.59µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.119567478Z level=info msg="Executing migration" id="create org table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.120382648Z level=info msg="Migration successfully executed" id="create org table v1" duration=809.02µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.123335336Z level=info msg="Executing migration" id="create index UQE_org_name - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.123998165Z level=info msg="Migration successfully executed" id="create index UQE_org_name - v1" duration=662.779µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.127412649Z level=info msg="Executing migration" id="create org_user table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.128114788Z level=info msg="Migration successfully executed" id="create org_user table v1" duration=701.639µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.182183916Z level=info msg="Executing migration" id="create index IDX_org_user_org_id - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.183516614Z level=info msg="Migration successfully executed" id="create index IDX_org_user_org_id - v1" duration=1.333068ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.19024608Z level=info msg="Executing migration" id="create index UQE_org_user_org_id_user_id - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.191316594Z level=info msg="Migration successfully executed" id="create index UQE_org_user_org_id_user_id - v1" duration=1.070984ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.194461535Z level=info msg="Executing migration" id="create index IDX_org_user_user_id - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.195229245Z level=info msg="Migration successfully executed" id="create index IDX_org_user_user_id - v1" duration=767.52µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.198147753Z level=info msg="Executing migration" id="Update org table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.198179623Z level=info msg="Migration successfully executed" id="Update org table charset" duration=25.73µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.200411412Z level=info msg="Executing migration" id="Update org_user table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.200434762Z level=info msg="Migration successfully executed" id="Update org_user table charset" duration=24.26µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.206337459Z level=info msg="Executing migration" id="Migrate all Read Only Viewers to Viewers" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.206773714Z level=info msg="Migration successfully executed" id="Migrate all Read Only Viewers to Viewers" duration=442.686µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.215562678Z level=info msg="Executing migration" id="create dashboard table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.216473539Z level=info msg="Migration successfully executed" id="create dashboard table" duration=910.831µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.223345818Z level=info msg="Executing migration" id="add index dashboard.account_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.224467652Z level=info msg="Migration successfully executed" id="add index dashboard.account_id" duration=1.125074ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.230298858Z level=info msg="Executing migration" id="add unique index dashboard_account_id_slug" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.231365192Z level=info msg="Migration successfully executed" id="add unique index dashboard_account_id_slug" duration=1.065734ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.241087387Z level=info msg="Executing migration" id="create dashboard_tag table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.243565159Z level=info msg="Migration successfully executed" id="create dashboard_tag table" duration=2.475262ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.255452453Z level=info msg="Executing migration" id="add unique index dashboard_tag.dasboard_id_term" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.256644078Z level=info msg="Migration successfully executed" id="add unique index dashboard_tag.dasboard_id_term" duration=1.194975ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.263572538Z level=info msg="Executing migration" id="drop index UQE_dashboard_tag_dashboard_id_term - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.26454865Z level=info msg="Migration successfully executed" id="drop index UQE_dashboard_tag_dashboard_id_term - v1" duration=972.232µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.275629032Z level=info msg="Executing migration" id="Rename table dashboard to dashboard_v1 - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.280224312Z level=info msg="Migration successfully executed" id="Rename table dashboard to dashboard_v1 - v1" duration=4.59295ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.288025513Z level=info msg="Executing migration" id="create dashboard v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.288984995Z level=info msg="Migration successfully executed" id="create dashboard v2" duration=961.543µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.295844084Z level=info msg="Executing migration" id="create index IDX_dashboard_org_id - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.297270072Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_org_id - v2" duration=1.437729ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.308228944Z level=info msg="Executing migration" id="create index UQE_dashboard_org_id_slug - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.308999514Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_org_id_slug - v2" duration=772.74µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.313628493Z level=info msg="Executing migration" id="copy dashboard v1 to v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.313890657Z level=info msg="Migration successfully executed" id="copy dashboard v1 to v2" duration=262.634µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.322282825Z level=info msg="Executing migration" id="drop table dashboard_v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.323139546Z level=info msg="Migration successfully executed" id="drop table dashboard_v1" duration=856.491µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.327225339Z level=info msg="Executing migration" id="alter dashboard.data to mediumtext v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.32732543Z level=info msg="Migration successfully executed" id="alter dashboard.data to mediumtext v1" duration=101.411µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.33197286Z level=info msg="Executing migration" id="Add column updated_by in dashboard - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.334786386Z level=info msg="Migration successfully executed" id="Add column updated_by in dashboard - v2" duration=2.813646ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.34046954Z level=info msg="Executing migration" id="Add column created_by in dashboard - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.342366965Z level=info msg="Migration successfully executed" id="Add column created_by in dashboard - v2" duration=1.898675ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.346851582Z level=info msg="Executing migration" id="Add column gnetId in dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.348565114Z level=info msg="Migration successfully executed" id="Add column gnetId in dashboard" duration=1.713312ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.353203525Z level=info msg="Executing migration" id="Add index for gnetId in dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.354265518Z level=info msg="Migration successfully executed" id="Add index for gnetId in dashboard" duration=1.061443ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.361953518Z level=info msg="Executing migration" id="Add column plugin_id in dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.366321344Z level=info msg="Migration successfully executed" id="Add column plugin_id in dashboard" duration=4.372526ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.375200309Z level=info msg="Executing migration" id="Add index for plugin_id in dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.376596657Z level=info msg="Migration successfully executed" id="Add index for plugin_id in dashboard" duration=1.396228ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.389182159Z level=info msg="Executing migration" id="Add index for dashboard_id in dashboard_tag" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.39160349Z level=info msg="Migration successfully executed" id="Add index for dashboard_id in dashboard_tag" duration=2.421741ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.397623378Z level=info msg="Executing migration" id="Update dashboard table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.397671409Z level=info msg="Migration successfully executed" id="Update dashboard table charset" duration=52.441µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.411494267Z level=info msg="Executing migration" id="Update dashboard_tag table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.411546208Z level=info msg="Migration successfully executed" id="Update dashboard_tag table charset" duration=58.051µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.455429065Z level=info msg="Executing migration" id="Add column folder_id in dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.458042349Z level=info msg="Migration successfully executed" id="Add column folder_id in dashboard" duration=2.618144ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.465185751Z level=info msg="Executing migration" id="Add column isFolder in dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.467342209Z level=info msg="Migration successfully executed" id="Add column isFolder in dashboard" duration=2.156348ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.471837987Z level=info msg="Executing migration" id="Add column has_acl in dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.474116186Z level=info msg="Migration successfully executed" id="Add column has_acl in dashboard" duration=2.271979ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.481741365Z level=info msg="Executing migration" id="Add column uid in dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.484069085Z level=info msg="Migration successfully executed" id="Add column uid in dashboard" duration=2.31884ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.493300464Z level=info msg="Executing migration" id="Update uid column values in dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.494044034Z level=info msg="Migration successfully executed" id="Update uid column values in dashboard" duration=746.03µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.499225521Z level=info msg="Executing migration" id="Add unique index dashboard_org_id_uid" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.500389346Z level=info msg="Migration successfully executed" id="Add unique index dashboard_org_id_uid" duration=1.165705ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.540049438Z level=info msg="Executing migration" id="Remove unique index org_id_slug" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.54328107Z level=info msg="Migration successfully executed" id="Remove unique index org_id_slug" duration=3.235942ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.558347584Z level=info msg="Executing migration" id="Update dashboard title length" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.558413385Z level=info msg="Migration successfully executed" id="Update dashboard title length" duration=70.721µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.564575985Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_title_folder_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.565505587Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_title_folder_id" duration=935.362µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.571555715Z level=info msg="Executing migration" id="create dashboard_provisioning" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.572534348Z level=info msg="Migration successfully executed" id="create dashboard_provisioning" duration=982.023µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.579595519Z level=info msg="Executing migration" id="Rename table dashboard_provisioning to dashboard_provisioning_tmp_qwerty - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.589808651Z level=info msg="Migration successfully executed" id="Rename table dashboard_provisioning to dashboard_provisioning_tmp_qwerty - v1" duration=10.210072ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.59590618Z level=info msg="Executing migration" id="create dashboard_provisioning v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.596506657Z level=info msg="Migration successfully executed" id="create dashboard_provisioning v2" duration=600.477µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.600654921Z level=info msg="Executing migration" id="create index IDX_dashboard_provisioning_dashboard_id - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.601702564Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_provisioning_dashboard_id - v2" duration=1.046973ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.611510791Z level=info msg="Executing migration" id="create index IDX_dashboard_provisioning_dashboard_id_name - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.612770187Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_provisioning_dashboard_id_name - v2" duration=1.262616ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.622890008Z level=info msg="Executing migration" id="copy dashboard_provisioning v1 to v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.623495486Z level=info msg="Migration successfully executed" id="copy dashboard_provisioning v1 to v2" duration=607.008µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.629155539Z level=info msg="Executing migration" id="drop dashboard_provisioning_tmp_qwerty" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.629814438Z level=info msg="Migration successfully executed" id="drop dashboard_provisioning_tmp_qwerty" duration=658.969µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.637048211Z level=info msg="Executing migration" id="Add check_sum column" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.640794059Z level=info msg="Migration successfully executed" id="Add check_sum column" duration=3.746848ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.645426349Z level=info msg="Executing migration" id="Add index for dashboard_title" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.646452902Z level=info msg="Migration successfully executed" id="Add index for dashboard_title" duration=1.026483ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.653865318Z level=info msg="Executing migration" id="delete tags for deleted dashboards" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.654244593Z level=info msg="Migration successfully executed" id="delete tags for deleted dashboards" duration=378.665µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.663503053Z level=info msg="Executing migration" id="delete stars for deleted dashboards" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.663706045Z level=info msg="Migration successfully executed" id="delete stars for deleted dashboards" duration=211.402µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.670357551Z level=info msg="Executing migration" id="Add index for dashboard_is_folder" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.671203122Z level=info msg="Migration successfully executed" id="Add index for dashboard_is_folder" duration=836.401µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.675995174Z level=info msg="Executing migration" id="Add isPublic for dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.678670529Z level=info msg="Migration successfully executed" id="Add isPublic for dashboard" duration=2.674995ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.686186956Z level=info msg="Executing migration" id="Add deleted for dashboard" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.688381114Z level=info msg="Migration successfully executed" id="Add deleted for dashboard" duration=2.194448ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.699143353Z level=info msg="Executing migration" id="Add index for deleted" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.699874053Z level=info msg="Migration successfully executed" id="Add index for deleted" duration=730.389µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.709409246Z level=info msg="Executing migration" id="create data_source table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.711093357Z level=info msg="Migration successfully executed" id="create data_source table" duration=1.693271ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.715965521Z level=info msg="Executing migration" id="add index data_source.account_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.717923726Z level=info msg="Migration successfully executed" id="add index data_source.account_id" duration=1.958336ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.724746284Z level=info msg="Executing migration" id="add unique index data_source.account_id_name" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.726104361Z level=info msg="Migration successfully executed" id="add unique index data_source.account_id_name" duration=1.357177ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.733968723Z level=info msg="Executing migration" id="drop index IDX_data_source_account_id - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.735256429Z level=info msg="Migration successfully executed" id="drop index IDX_data_source_account_id - v1" duration=1.287566ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.738865506Z level=info msg="Executing migration" id="drop index UQE_data_source_account_id_name - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.739981241Z level=info msg="Migration successfully executed" id="drop index UQE_data_source_account_id_name - v1" duration=1.116225ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.747139463Z level=info msg="Executing migration" id="Rename table data_source to data_source_v1 - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.753778149Z level=info msg="Migration successfully executed" id="Rename table data_source to data_source_v1 - v1" duration=6.638066ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.759895768Z level=info msg="Executing migration" id="create data_source table v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.760946872Z level=info msg="Migration successfully executed" id="create data_source table v2" duration=1.050724ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.799430349Z level=info msg="Executing migration" id="create index IDX_data_source_org_id - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.800378981Z level=info msg="Migration successfully executed" id="create index IDX_data_source_org_id - v2" duration=948.433µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.807123398Z level=info msg="Executing migration" id="create index UQE_data_source_org_id_name - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.808564157Z level=info msg="Migration successfully executed" id="create index UQE_data_source_org_id_name - v2" duration=1.439789ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.812242584Z level=info msg="Executing migration" id="Drop old table data_source_v1 #2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.814112098Z level=info msg="Migration successfully executed" id="Drop old table data_source_v1 #2" duration=1.882484ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.824103437Z level=info msg="Executing migration" id="Add column with_credentials" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.828559585Z level=info msg="Migration successfully executed" id="Add column with_credentials" duration=4.447088ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.837366928Z level=info msg="Executing migration" id="Add secure json data column" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.841510482Z level=info msg="Migration successfully executed" id="Add secure json data column" duration=4.147294ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.847145535Z level=info msg="Executing migration" id="Update data_source table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.847378688Z level=info msg="Migration successfully executed" id="Update data_source table charset" duration=234.593µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.862425702Z level=info msg="Executing migration" id="Update initial version to 1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.862924589Z level=info msg="Migration successfully executed" id="Update initial version to 1" duration=502.257µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.904317943Z level=info msg="Executing migration" id="Add read_only data column" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.908414596Z level=info msg="Migration successfully executed" id="Add read_only data column" duration=4.091373ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.914501715Z level=info msg="Executing migration" id="Migrate logging ds to loki ds" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.91488836Z level=info msg="Migration successfully executed" id="Migrate logging ds to loki ds" duration=386.645µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.923003955Z level=info msg="Executing migration" id="Update json_data with nulls" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.923321369Z level=info msg="Migration successfully executed" id="Update json_data with nulls" duration=317.184µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.934772627Z level=info msg="Executing migration" id="Add uid column" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.939219034Z level=info msg="Migration successfully executed" id="Add uid column" duration=4.447097ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.944596184Z level=info msg="Executing migration" id="Update uid value" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.9451257Z level=info msg="Migration successfully executed" id="Update uid value" duration=528.866µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.948644536Z level=info msg="Executing migration" id="Add unique index datasource_org_id_uid" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.94972948Z level=info msg="Migration successfully executed" id="Add unique index datasource_org_id_uid" duration=1.076064ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.961764375Z level=info msg="Executing migration" id="add unique index datasource_org_id_is_default" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.963283625Z level=info msg="Migration successfully executed" id="add unique index datasource_org_id_is_default" duration=1.51998ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.972610966Z level=info msg="Executing migration" id="Add is_prunable column" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.977355677Z level=info msg="Migration successfully executed" id="Add is_prunable column" duration=4.744361ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.983838051Z level=info msg="Executing migration" id="Add api_version column" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.98610988Z level=info msg="Migration successfully executed" id="Add api_version column" duration=2.272279ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.990668179Z level=info msg="Executing migration" id="create api_key table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.99154027Z level=info msg="Migration successfully executed" id="create api_key table" duration=864.821µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:00.998345128Z level=info msg="Executing migration" id="add index api_key.account_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.00085867Z level=info msg="Migration successfully executed" id="add index api_key.account_id" duration=2.513342ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.008860754Z level=info msg="Executing migration" id="add index api_key.key" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.010630697Z level=info msg="Migration successfully executed" id="add index api_key.key" duration=1.772493ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.0170752Z level=info msg="Executing migration" id="add index api_key.account_id_name" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.018024522Z level=info msg="Migration successfully executed" id="add index api_key.account_id_name" duration=949.212µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.020840739Z level=info msg="Executing migration" id="drop index IDX_api_key_account_id - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.021473837Z level=info msg="Migration successfully executed" id="drop index IDX_api_key_account_id - v1" duration=633.058µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.027001398Z level=info msg="Executing migration" id="drop index UQE_api_key_key - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.028481347Z level=info msg="Migration successfully executed" id="drop index UQE_api_key_key - v1" duration=1.478739ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.034723748Z level=info msg="Executing migration" id="drop index UQE_api_key_account_id_name - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.035730391Z level=info msg="Migration successfully executed" id="drop index UQE_api_key_account_id_name - v1" duration=1.008013ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.040558143Z level=info msg="Executing migration" id="Rename table api_key to api_key_v1 - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.049427648Z level=info msg="Migration successfully executed" id="Rename table api_key to api_key_v1 - v1" duration=8.877735ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.053719903Z level=info msg="Executing migration" id="create api_key table v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.054973569Z level=info msg="Migration successfully executed" id="create api_key table v2" duration=1.253666ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.062558287Z level=info msg="Executing migration" id="create index IDX_api_key_org_id - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.064014026Z level=info msg="Migration successfully executed" id="create index IDX_api_key_org_id - v2" duration=1.458529ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.070287357Z level=info msg="Executing migration" id="create index UQE_api_key_key - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.071380621Z level=info msg="Migration successfully executed" id="create index UQE_api_key_key - v2" duration=1.096044ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.076507437Z level=info msg="Executing migration" id="create index UQE_api_key_org_id_name - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.07748052Z level=info msg="Migration successfully executed" id="create index UQE_api_key_org_id_name - v2" duration=974.123µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.082344093Z level=info msg="Executing migration" id="copy api_key v1 to v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.082858389Z level=info msg="Migration successfully executed" id="copy api_key v1 to v2" duration=515.147µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.090579089Z level=info msg="Executing migration" id="Drop old table api_key_v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.091292168Z level=info msg="Migration successfully executed" id="Drop old table api_key_v1" duration=713.239µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.100363655Z level=info msg="Executing migration" id="Update api_key table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.100410566Z level=info msg="Migration successfully executed" id="Update api_key table charset" duration=55.131µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.106755378Z level=info msg="Executing migration" id="Add expires to api_key table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.112541473Z level=info msg="Migration successfully executed" id="Add expires to api_key table" duration=5.785934ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.200763902Z level=info msg="Executing migration" id="Add service account foreign key" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.206192722Z level=info msg="Migration successfully executed" id="Add service account foreign key" duration=5.42818ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.215424991Z level=info msg="Executing migration" id="set service account foreign key to nil if 0" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.215748375Z level=info msg="Migration successfully executed" id="set service account foreign key to nil if 0" duration=323.244µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.224534608Z level=info msg="Executing migration" id="Add last_used_at to api_key table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.229056187Z level=info msg="Migration successfully executed" id="Add last_used_at to api_key table" duration=4.520719ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.236902738Z level=info msg="Executing migration" id="Add is_revoked column to api_key table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.239672254Z level=info msg="Migration successfully executed" id="Add is_revoked column to api_key table" duration=2.768376ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.243598375Z level=info msg="Executing migration" id="create dashboard_snapshot table v4" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.244494286Z level=info msg="Migration successfully executed" id="create dashboard_snapshot table v4" duration=895.781µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.280830175Z level=info msg="Executing migration" id="drop table dashboard_snapshot_v4 #1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.282040461Z level=info msg="Migration successfully executed" id="drop table dashboard_snapshot_v4 #1" duration=1.212936ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.287313939Z level=info msg="Executing migration" id="create dashboard_snapshot table v5 #2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.288864159Z level=info msg="Migration successfully executed" id="create dashboard_snapshot table v5 #2" duration=1.55021ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.292689989Z level=info msg="Executing migration" id="create index UQE_dashboard_snapshot_key - v5" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.293717042Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_snapshot_key - v5" duration=1.023124ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.29670951Z level=info msg="Executing migration" id="create index UQE_dashboard_snapshot_delete_key - v5" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.297620852Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_snapshot_delete_key - v5" duration=911.502µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.303019952Z level=info msg="Executing migration" id="create index IDX_dashboard_snapshot_user_id - v5" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.303927614Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_snapshot_user_id - v5" duration=907.602µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.307371778Z level=info msg="Executing migration" id="alter dashboard_snapshot to mediumtext v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.307441879Z level=info msg="Migration successfully executed" id="alter dashboard_snapshot to mediumtext v2" duration=70.981µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.31060939Z level=info msg="Executing migration" id="Update dashboard_snapshot table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.31065566Z level=info msg="Migration successfully executed" id="Update dashboard_snapshot table charset" duration=48.52µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.314673392Z level=info msg="Executing migration" id="Add column external_delete_url to dashboard_snapshots table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.319394303Z level=info msg="Migration successfully executed" id="Add column external_delete_url to dashboard_snapshots table" duration=4.720261ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.325204198Z level=info msg="Executing migration" id="Add encrypted dashboard json column" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.328289708Z level=info msg="Migration successfully executed" id="Add encrypted dashboard json column" duration=3.08423ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.331662622Z level=info msg="Executing migration" id="Change dashboard_encrypted column to MEDIUMBLOB" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.331737273Z level=info msg="Migration successfully executed" id="Change dashboard_encrypted column to MEDIUMBLOB" duration=68.22µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.33538725Z level=info msg="Executing migration" id="create quota table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.336403453Z level=info msg="Migration successfully executed" id="create quota table v1" duration=1.016673ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.341543839Z level=info msg="Executing migration" id="create index UQE_quota_org_id_user_id_target - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.343082119Z level=info msg="Migration successfully executed" id="create index UQE_quota_org_id_user_id_target - v1" duration=1.53832ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.346607505Z level=info msg="Executing migration" id="Update quota table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.346648955Z level=info msg="Migration successfully executed" id="Update quota table charset" duration=42.85µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.350243321Z level=info msg="Executing migration" id="create plugin_setting table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.351216604Z level=info msg="Migration successfully executed" id="create plugin_setting table" duration=973.083µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.354471426Z level=info msg="Executing migration" id="create index UQE_plugin_setting_org_id_plugin_id - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.355595851Z level=info msg="Migration successfully executed" id="create index UQE_plugin_setting_org_id_plugin_id - v1" duration=1.124595ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.360082109Z level=info msg="Executing migration" id="Add column plugin_version to plugin_settings" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.363844967Z level=info msg="Migration successfully executed" id="Add column plugin_version to plugin_settings" duration=3.759428ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.368067811Z level=info msg="Executing migration" id="Update plugin_setting table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.368122472Z level=info msg="Migration successfully executed" id="Update plugin_setting table charset" duration=55.951µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.371323154Z level=info msg="Executing migration" id="create session table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.372498879Z level=info msg="Migration successfully executed" id="create session table" duration=1.175265ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.377176229Z level=info msg="Executing migration" id="Drop old table playlist table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.37726041Z level=info msg="Migration successfully executed" id="Drop old table playlist table" duration=84.571µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.38033486Z level=info msg="Executing migration" id="Drop old table playlist_item table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.380407271Z level=info msg="Migration successfully executed" id="Drop old table playlist_item table" duration=72.811µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.383741334Z level=info msg="Executing migration" id="create playlist table v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.384660446Z level=info msg="Migration successfully executed" id="create playlist table v2" duration=918.982µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.392372645Z level=info msg="Executing migration" id="create playlist item table v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.393810214Z level=info msg="Migration successfully executed" id="create playlist item table v2" duration=1.439749ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.398102689Z level=info msg="Executing migration" id="Update playlist table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.39813192Z level=info msg="Migration successfully executed" id="Update playlist table charset" duration=36.121µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.40043896Z level=info msg="Executing migration" id="Update playlist_item table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.40046467Z level=info msg="Migration successfully executed" id="Update playlist_item table charset" duration=26.98µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.403421308Z level=info msg="Executing migration" id="Add playlist column created_at" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.407020445Z level=info msg="Migration successfully executed" id="Add playlist column created_at" duration=3.605097ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.411924728Z level=info msg="Executing migration" id="Add playlist column updated_at" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.41521145Z level=info msg="Migration successfully executed" id="Add playlist column updated_at" duration=3.286642ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.428842016Z level=info msg="Executing migration" id="drop preferences table v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.429278442Z level=info msg="Migration successfully executed" id="drop preferences table v2" duration=438.626µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.439370442Z level=info msg="Executing migration" id="drop preferences table v3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.439697406Z level=info msg="Migration successfully executed" id="drop preferences table v3" duration=326.284µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.450216352Z level=info msg="Executing migration" id="create preferences table v3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.451353247Z level=info msg="Migration successfully executed" id="create preferences table v3" duration=1.139895ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.460308173Z level=info msg="Executing migration" id="Update preferences table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.460349833Z level=info msg="Migration successfully executed" id="Update preferences table charset" duration=42.01µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.469937427Z level=info msg="Executing migration" id="Add column team_id in preferences" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.474794519Z level=info msg="Migration successfully executed" id="Add column team_id in preferences" duration=4.851232ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.484321312Z level=info msg="Executing migration" id="Update team_id column values in preferences" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.484649257Z level=info msg="Migration successfully executed" id="Update team_id column values in preferences" duration=332.265µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.494157639Z level=info msg="Executing migration" id="Add column week_start in preferences" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.499324806Z level=info msg="Migration successfully executed" id="Add column week_start in preferences" duration=5.166937ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.530355417Z level=info msg="Executing migration" id="Add column preferences.json_data" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.536577027Z level=info msg="Migration successfully executed" id="Add column preferences.json_data" duration=6.22236ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.566072728Z level=info msg="Executing migration" id="alter preferences.json_data to mediumtext v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.566325611Z level=info msg="Migration successfully executed" id="alter preferences.json_data to mediumtext v1" duration=252.323µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.570614517Z level=info msg="Executing migration" id="Add preferences index org_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.572109746Z level=info msg="Migration successfully executed" id="Add preferences index org_id" duration=1.495069ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.57628136Z level=info msg="Executing migration" id="Add preferences index user_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.577185731Z level=info msg="Migration successfully executed" id="Add preferences index user_id" duration=904.531µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.582651012Z level=info msg="Executing migration" id="create alert table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.584602967Z level=info msg="Migration successfully executed" id="create alert table v1" duration=1.949165ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.589265357Z level=info msg="Executing migration" id="add index alert org_id & id " 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.590720246Z level=info msg="Migration successfully executed" id="add index alert org_id & id " duration=1.454389ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.59641357Z level=info msg="Executing migration" id="add index alert state" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.597408193Z level=info msg="Migration successfully executed" id="add index alert state" duration=994.383µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.638484693Z level=info msg="Executing migration" id="add index alert dashboard_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.640401818Z level=info msg="Migration successfully executed" id="add index alert dashboard_id" duration=1.919694ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.646044281Z level=info msg="Executing migration" id="Create alert_rule_tag table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.647188325Z level=info msg="Migration successfully executed" id="Create alert_rule_tag table v1" duration=1.148354ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.651597772Z level=info msg="Executing migration" id="Add unique index alert_rule_tag.alert_id_tag_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.652396452Z level=info msg="Migration successfully executed" id="Add unique index alert_rule_tag.alert_id_tag_id" duration=798.18µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.655963459Z level=info msg="Executing migration" id="drop index UQE_alert_rule_tag_alert_id_tag_id - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.656754359Z level=info msg="Migration successfully executed" id="drop index UQE_alert_rule_tag_alert_id_tag_id - v1" duration=790.56µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.662339801Z level=info msg="Executing migration" id="Rename table alert_rule_tag to alert_rule_tag_v1 - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.672180088Z level=info msg="Migration successfully executed" id="Rename table alert_rule_tag to alert_rule_tag_v1 - v1" duration=9.838577ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.676910279Z level=info msg="Executing migration" id="Create alert_rule_tag table v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.677612998Z level=info msg="Migration successfully executed" id="Create alert_rule_tag table v2" duration=702.979µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.682104746Z level=info msg="Executing migration" id="create index UQE_alert_rule_tag_alert_id_tag_id - Add unique index alert_rule_tag.alert_id_tag_id V2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.683094499Z level=info msg="Migration successfully executed" id="create index UQE_alert_rule_tag_alert_id_tag_id - Add unique index alert_rule_tag.alert_id_tag_id V2" duration=989.653µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.714035088Z level=info msg="Executing migration" id="copy alert_rule_tag v1 to v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.714432163Z level=info msg="Migration successfully executed" id="copy alert_rule_tag v1 to v2" duration=398.995µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.717847567Z level=info msg="Executing migration" id="drop table alert_rule_tag_v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.718388075Z level=info msg="Migration successfully executed" id="drop table alert_rule_tag_v1" duration=540.237µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.721659347Z level=info msg="Executing migration" id="create alert_notification table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.722314085Z level=info msg="Migration successfully executed" id="create alert_notification table v1" duration=660.948µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.726677632Z level=info msg="Executing migration" id="Add column is_default" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.729508238Z level=info msg="Migration successfully executed" id="Add column is_default" duration=2.829997ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.733916765Z level=info msg="Executing migration" id="Add column frequency" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.736739862Z level=info msg="Migration successfully executed" id="Add column frequency" duration=2.822446ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.740300167Z level=info msg="Executing migration" id="Add column send_reminder" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.743186165Z level=info msg="Migration successfully executed" id="Add column send_reminder" duration=2.885558ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.748489493Z level=info msg="Executing migration" id="Add column disable_resolve_message" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.751252279Z level=info msg="Migration successfully executed" id="Add column disable_resolve_message" duration=2.762276ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.7551915Z level=info msg="Executing migration" id="add index alert_notification org_id & name" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.75677613Z level=info msg="Migration successfully executed" id="add index alert_notification org_id & name" duration=1.58413ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.761135147Z level=info msg="Executing migration" id="Update alert table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.761174137Z level=info msg="Migration successfully executed" id="Update alert table charset" duration=41.31µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.766116521Z level=info msg="Executing migration" id="Update alert_notification table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.766154751Z level=info msg="Migration successfully executed" id="Update alert_notification table charset" duration=38.58µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.76997297Z level=info msg="Executing migration" id="create notification_journal table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.770876802Z level=info msg="Migration successfully executed" id="create notification_journal table v1" duration=903.972µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.776705447Z level=info msg="Executing migration" id="add index notification_journal org_id & alert_id & notifier_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.778225237Z level=info msg="Migration successfully executed" id="add index notification_journal org_id & alert_id & notifier_id" duration=1.52106ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.782633714Z level=info msg="Executing migration" id="drop alert_notification_journal" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.783543166Z level=info msg="Migration successfully executed" id="drop alert_notification_journal" duration=908.192µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.788288687Z level=info msg="Executing migration" id="create alert_notification_state table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.789741546Z level=info msg="Migration successfully executed" id="create alert_notification_state table v1" duration=1.451919ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.79393055Z level=info msg="Executing migration" id="add index alert_notification_state org_id & alert_id & notifier_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.795682633Z level=info msg="Migration successfully executed" id="add index alert_notification_state org_id & alert_id & notifier_id" duration=1.754643ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.800397503Z level=info msg="Executing migration" id="Add for to alert table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.805665271Z level=info msg="Migration successfully executed" id="Add for to alert table" duration=5.267248ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.809517811Z level=info msg="Executing migration" id="Add column uid in alert_notification" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.813896208Z level=info msg="Migration successfully executed" id="Add column uid in alert_notification" duration=4.378697ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.817545805Z level=info msg="Executing migration" id="Update uid column values in alert_notification" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.817778538Z level=info msg="Migration successfully executed" id="Update uid column values in alert_notification" duration=233.133µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.823268319Z level=info msg="Executing migration" id="Add unique index alert_notification_org_id_uid" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.824244171Z level=info msg="Migration successfully executed" id="Add unique index alert_notification_org_id_uid" duration=975.032µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.833244227Z level=info msg="Executing migration" id="Remove unique index org_id_name" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.835316814Z level=info msg="Migration successfully executed" id="Remove unique index org_id_name" duration=2.072057ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.840893906Z level=info msg="Executing migration" id="Add column secure_settings in alert_notification" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.848020088Z level=info msg="Migration successfully executed" id="Add column secure_settings in alert_notification" duration=7.123982ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.852955152Z level=info msg="Executing migration" id="alter alert.settings to mediumtext" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.853121964Z level=info msg="Migration successfully executed" id="alter alert.settings to mediumtext" duration=168.622µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.857971707Z level=info msg="Executing migration" id="Add non-unique index alert_notification_state_alert_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.859655159Z level=info msg="Migration successfully executed" id="Add non-unique index alert_notification_state_alert_id" duration=1.682962ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.864831575Z level=info msg="Executing migration" id="Add non-unique index alert_rule_tag_alert_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.865884739Z level=info msg="Migration successfully executed" id="Add non-unique index alert_rule_tag_alert_id" duration=1.045154ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.869785039Z level=info msg="Executing migration" id="Drop old annotation table v4" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.870017272Z level=info msg="Migration successfully executed" id="Drop old annotation table v4" duration=232.643µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.873256754Z level=info msg="Executing migration" id="create annotation table v5" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.874250507Z level=info msg="Migration successfully executed" id="create annotation table v5" duration=993.883µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.880370976Z level=info msg="Executing migration" id="add index annotation 0 v3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.881625182Z level=info msg="Migration successfully executed" id="add index annotation 0 v3" duration=1.253946ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.884907504Z level=info msg="Executing migration" id="add index annotation 1 v3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.886031349Z level=info msg="Migration successfully executed" id="add index annotation 1 v3" duration=1.123645ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.888971977Z level=info msg="Executing migration" id="add index annotation 2 v3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.889926429Z level=info msg="Migration successfully executed" id="add index annotation 2 v3" duration=953.672µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.941719118Z level=info msg="Executing migration" id="add index annotation 3 v3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.944151089Z level=info msg="Migration successfully executed" id="add index annotation 3 v3" duration=2.433761ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.947866577Z level=info msg="Executing migration" id="add index annotation 4 v3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.948957701Z level=info msg="Migration successfully executed" id="add index annotation 4 v3" duration=1.091884ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.953634292Z level=info msg="Executing migration" id="Update annotation table charset" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.953663052Z level=info msg="Migration successfully executed" id="Update annotation table charset" duration=29.98µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.956450468Z level=info msg="Executing migration" id="Add column region_id to annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.96127373Z level=info msg="Migration successfully executed" id="Add column region_id to annotation table" duration=4.822722ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.964148168Z level=info msg="Executing migration" id="Drop category_id index" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.964999219Z level=info msg="Migration successfully executed" id="Drop category_id index" duration=851.231µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.967875006Z level=info msg="Executing migration" id="Add column tags to annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.973225555Z level=info msg="Migration successfully executed" id="Add column tags to annotation table" duration=5.348909ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.978051827Z level=info msg="Executing migration" id="Create annotation_tag table v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.978860548Z level=info msg="Migration successfully executed" id="Create annotation_tag table v2" duration=810.221µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.992404823Z level=info msg="Executing migration" id="Add unique index annotation_tag.annotation_id_tag_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.993540487Z level=info msg="Migration successfully executed" id="Add unique index annotation_tag.annotation_id_tag_id" duration=1.137924ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.997500228Z level=info msg="Executing migration" id="drop index UQE_annotation_tag_annotation_id_tag_id - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:01.99838226Z level=info msg="Migration successfully executed" id="drop index UQE_annotation_tag_annotation_id_tag_id - v2" duration=882.572µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.003429035Z level=info msg="Executing migration" id="Rename table annotation_tag to annotation_tag_v2 - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.01470734Z level=info msg="Migration successfully executed" id="Rename table annotation_tag to annotation_tag_v2 - v2" duration=11.281525ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.017637558Z level=info msg="Executing migration" id="Create annotation_tag table v3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.018162655Z level=info msg="Migration successfully executed" id="Create annotation_tag table v3" duration=524.907µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.02316364Z level=info msg="Executing migration" id="create index UQE_annotation_tag_annotation_id_tag_id - Add unique index annotation_tag.annotation_id_tag_id V3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.02399385Z level=info msg="Migration successfully executed" id="create index UQE_annotation_tag_annotation_id_tag_id - Add unique index annotation_tag.annotation_id_tag_id V3" duration=831.02µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.029875266Z level=info msg="Executing migration" id="copy annotation_tag v2 to v3" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.030296561Z level=info msg="Migration successfully executed" id="copy annotation_tag v2 to v3" duration=426.035µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.03330463Z level=info msg="Executing migration" id="drop table annotation_tag_v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.033975799Z level=info msg="Migration successfully executed" id="drop table annotation_tag_v2" duration=669.569µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.037353373Z level=info msg="Executing migration" id="Update alert annotations and set TEXT to empty" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.03793596Z level=info msg="Migration successfully executed" id="Update alert annotations and set TEXT to empty" duration=582.417µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.043016616Z level=info msg="Executing migration" id="Add created time to annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.049923025Z level=info msg="Migration successfully executed" id="Add created time to annotation table" duration=6.906289ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.053018525Z level=info msg="Executing migration" id="Add updated time to annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.056025474Z level=info msg="Migration successfully executed" id="Add updated time to annotation table" duration=3.006699ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.059702361Z level=info msg="Executing migration" id="Add index for created in annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.060777045Z level=info msg="Migration successfully executed" id="Add index for created in annotation table" duration=1.080014ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.065265693Z level=info msg="Executing migration" id="Add index for updated in annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.067507382Z level=info msg="Migration successfully executed" id="Add index for updated in annotation table" duration=2.242879ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.070731993Z level=info msg="Executing migration" id="Convert existing annotations from seconds to milliseconds" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.071090428Z level=info msg="Migration successfully executed" id="Convert existing annotations from seconds to milliseconds" duration=358.615µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.076598039Z level=info msg="Executing migration" id="Add epoch_end column" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.081822396Z level=info msg="Migration successfully executed" id="Add epoch_end column" duration=5.224557ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.084940357Z level=info msg="Executing migration" id="Add index for epoch_end" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.086256214Z level=info msg="Migration successfully executed" id="Add index for epoch_end" duration=1.315657ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.090717151Z level=info msg="Executing migration" id="Make epoch_end the same as epoch" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.091050075Z level=info msg="Migration successfully executed" id="Make epoch_end the same as epoch" duration=332.834µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.094358978Z level=info msg="Executing migration" id="Move region to single row" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.094991966Z level=info msg="Migration successfully executed" id="Move region to single row" duration=632.458µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.099083259Z level=info msg="Executing migration" id="Remove index org_id_epoch from annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.100190604Z level=info msg="Migration successfully executed" id="Remove index org_id_epoch from annotation table" duration=1.107285ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.104090814Z level=info msg="Executing migration" id="Remove index org_id_dashboard_id_panel_id_epoch from annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.105110447Z level=info msg="Migration successfully executed" id="Remove index org_id_dashboard_id_panel_id_epoch from annotation table" duration=1.019573ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.108309128Z level=info msg="Executing migration" id="Add index for org_id_dashboard_id_epoch_end_epoch on annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.109430173Z level=info msg="Migration successfully executed" id="Add index for org_id_dashboard_id_epoch_end_epoch on annotation table" duration=1.121205ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.112525233Z level=info msg="Executing migration" id="Add index for org_id_epoch_end_epoch on annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.113567026Z level=info msg="Migration successfully executed" id="Add index for org_id_epoch_end_epoch on annotation table" duration=1.034383ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.116852069Z level=info msg="Executing migration" id="Remove index org_id_epoch_epoch_end from annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.118201366Z level=info msg="Migration successfully executed" id="Remove index org_id_epoch_epoch_end from annotation table" duration=1.350047ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.121002782Z level=info msg="Executing migration" id="Add index for alert_id on annotation table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.122151727Z level=info msg="Migration successfully executed" id="Add index for alert_id on annotation table" duration=1.141015ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.12623542Z level=info msg="Executing migration" id="Increase tags column to length 4096" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.126391152Z level=info msg="Migration successfully executed" id="Increase tags column to length 4096" duration=155.562µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.130835799Z level=info msg="Executing migration" id="create test_data table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.131845382Z level=info msg="Migration successfully executed" id="create test_data table" duration=1.009333ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.135351537Z level=info msg="Executing migration" id="create dashboard_version table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.136747375Z level=info msg="Migration successfully executed" id="create dashboard_version table v1" duration=1.395268ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.140557584Z level=info msg="Executing migration" id="add index dashboard_version.dashboard_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.141842991Z level=info msg="Migration successfully executed" id="add index dashboard_version.dashboard_id" duration=1.286237ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.14481591Z level=info msg="Executing migration" id="add unique index dashboard_version.dashboard_id and dashboard_version.version" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.146080116Z level=info msg="Migration successfully executed" id="add unique index dashboard_version.dashboard_id and dashboard_version.version" duration=1.264426ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.149153165Z level=info msg="Executing migration" id="Set dashboard version to 1 where 0" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.14948384Z level=info msg="Migration successfully executed" id="Set dashboard version to 1 where 0" duration=331.155µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.152693061Z level=info msg="Executing migration" id="save existing dashboard data in dashboard_version table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.153157227Z level=info msg="Migration successfully executed" id="save existing dashboard data in dashboard_version table v1" duration=464.396µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.156065664Z level=info msg="Executing migration" id="alter dashboard_version.data to mediumtext v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.156228757Z level=info msg="Migration successfully executed" id="alter dashboard_version.data to mediumtext v1" duration=162.692µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.1596124Z level=info msg="Executing migration" id="create team table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.160508222Z level=info msg="Migration successfully executed" id="create team table" duration=895.982µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.16344684Z level=info msg="Executing migration" id="add index team.org_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.164471553Z level=info msg="Migration successfully executed" id="add index team.org_id" duration=1.024733ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.168740438Z level=info msg="Executing migration" id="add unique index team_org_id_name" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.169996094Z level=info msg="Migration successfully executed" id="add unique index team_org_id_name" duration=1.255256ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.174161318Z level=info msg="Executing migration" id="Add column uid in team" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.179564998Z level=info msg="Migration successfully executed" id="Add column uid in team" duration=5.40326ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.182719589Z level=info msg="Executing migration" id="Update uid column values in team" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.182912741Z level=info msg="Migration successfully executed" id="Update uid column values in team" duration=192.652µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.186334145Z level=info msg="Executing migration" id="Add unique index team_org_id_uid" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.187107045Z level=info msg="Migration successfully executed" id="Add unique index team_org_id_uid" duration=772.83µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.190102954Z level=info msg="Executing migration" id="create team member table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.190805613Z level=info msg="Migration successfully executed" id="create team member table" duration=702.669µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.194143856Z level=info msg="Executing migration" id="add index team_member.org_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.196142232Z level=info msg="Migration successfully executed" id="add index team_member.org_id" duration=1.998336ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.199634577Z level=info msg="Executing migration" id="add unique index team_member_org_id_team_id_user_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.200783192Z level=info msg="Migration successfully executed" id="add unique index team_member_org_id_team_id_user_id" duration=1.148505ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.203882951Z level=info msg="Executing migration" id="add index team_member.team_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.204949905Z level=info msg="Migration successfully executed" id="add index team_member.team_id" duration=1.067024ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.208099356Z level=info msg="Executing migration" id="Add column email to team table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.21614657Z level=info msg="Migration successfully executed" id="Add column email to team table" duration=8.050714ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.219826267Z level=info msg="Executing migration" id="Add column external to team_member table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.224419217Z level=info msg="Migration successfully executed" id="Add column external to team_member table" duration=4.59346ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.227675879Z level=info msg="Executing migration" id="Add column permission to team_member table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.232160266Z level=info msg="Migration successfully executed" id="Add column permission to team_member table" duration=4.484117ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.237588196Z level=info msg="Executing migration" id="create dashboard acl table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.238557059Z level=info msg="Migration successfully executed" id="create dashboard acl table" duration=969.683µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.24407215Z level=info msg="Executing migration" id="add index dashboard_acl_dashboard_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.245518579Z level=info msg="Migration successfully executed" id="add index dashboard_acl_dashboard_id" duration=1.445979ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.248766411Z level=info msg="Executing migration" id="add unique index dashboard_acl_dashboard_id_user_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.250070608Z level=info msg="Migration successfully executed" id="add unique index dashboard_acl_dashboard_id_user_id" duration=1.303446ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.255952384Z level=info msg="Executing migration" id="add unique index dashboard_acl_dashboard_id_team_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.257061538Z level=info msg="Migration successfully executed" id="add unique index dashboard_acl_dashboard_id_team_id" duration=1.108364ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.260780716Z level=info msg="Executing migration" id="add index dashboard_acl_user_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.261955171Z level=info msg="Migration successfully executed" id="add index dashboard_acl_user_id" duration=1.177485ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.268204921Z level=info msg="Executing migration" id="add index dashboard_acl_team_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.270253428Z level=info msg="Migration successfully executed" id="add index dashboard_acl_team_id" duration=2.059707ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.273212046Z level=info msg="Executing migration" id="add index dashboard_acl_org_id_role" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.273965096Z level=info msg="Migration successfully executed" id="add index dashboard_acl_org_id_role" duration=752.72µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.349261168Z level=info msg="Executing migration" id="add index dashboard_permission" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.350936149Z level=info msg="Migration successfully executed" id="add index dashboard_permission" duration=1.676351ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.357810968Z level=info msg="Executing migration" id="save default acl rules in dashboard_acl table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.358389325Z level=info msg="Migration successfully executed" id="save default acl rules in dashboard_acl table" duration=578.987µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.363905967Z level=info msg="Executing migration" id="delete acl rules for deleted dashboards and folders" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.364284082Z level=info msg="Migration successfully executed" id="delete acl rules for deleted dashboards and folders" duration=380.655µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.368210882Z level=info msg="Executing migration" id="create tag table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.369072903Z level=info msg="Migration successfully executed" id="create tag table" duration=853.131µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.37269416Z level=info msg="Executing migration" id="add index tag.key_value" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.373789384Z level=info msg="Migration successfully executed" id="add index tag.key_value" duration=1.094984ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.379462718Z level=info msg="Executing migration" id="create login attempt table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.380309189Z level=info msg="Migration successfully executed" id="create login attempt table" duration=848.791µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.386452418Z level=info msg="Executing migration" id="add index login_attempt.username" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.387214527Z level=info msg="Migration successfully executed" id="add index login_attempt.username" duration=762.749µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.392673078Z level=info msg="Executing migration" id="drop index IDX_login_attempt_username - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.393479198Z level=info msg="Migration successfully executed" id="drop index IDX_login_attempt_username - v1" duration=805.66µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.400936095Z level=info msg="Executing migration" id="Rename table login_attempt to login_attempt_tmp_qwerty - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.418338539Z level=info msg="Migration successfully executed" id="Rename table login_attempt to login_attempt_tmp_qwerty - v1" duration=17.404294ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.421983466Z level=info msg="Executing migration" id="create login_attempt v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.422817727Z level=info msg="Migration successfully executed" id="create login_attempt v2" duration=834.361µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.427255634Z level=info msg="Executing migration" id="create index IDX_login_attempt_username - v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.428824975Z level=info msg="Migration successfully executed" id="create index IDX_login_attempt_username - v2" duration=1.578911ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.436929129Z level=info msg="Executing migration" id="copy login_attempt v1 to v2" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.437250093Z level=info msg="Migration successfully executed" id="copy login_attempt v1 to v2" duration=321.354µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.44091702Z level=info msg="Executing migration" id="drop login_attempt_tmp_qwerty" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.441881723Z level=info msg="Migration successfully executed" id="drop login_attempt_tmp_qwerty" duration=966.193µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.445682352Z level=info msg="Executing migration" id="create user auth table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.447254062Z level=info msg="Migration successfully executed" id="create user auth table" duration=1.5741ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.450854309Z level=info msg="Executing migration" id="create index IDX_user_auth_auth_module_auth_id - v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.452447879Z level=info msg="Migration successfully executed" id="create index IDX_user_auth_auth_module_auth_id - v1" duration=1.5975ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.457110849Z level=info msg="Executing migration" id="alter user_auth.auth_id to length 190" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.457383213Z level=info msg="Migration successfully executed" id="alter user_auth.auth_id to length 190" duration=267.964µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.46099601Z level=info msg="Executing migration" id="Add OAuth access token to user_auth" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.469651851Z level=info msg="Migration successfully executed" id="Add OAuth access token to user_auth" duration=8.656651ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.92878983Z level=info msg="Executing migration" id="Add OAuth refresh token to user_auth" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.937179539Z level=info msg="Migration successfully executed" id="Add OAuth refresh token to user_auth" duration=8.394719ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.942073892Z level=info msg="Executing migration" id="Add OAuth token type to user_auth" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.947471532Z level=info msg="Migration successfully executed" id="Add OAuth token type to user_auth" duration=5.39753ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.952009231Z level=info msg="Executing migration" id="Add OAuth expiry to user_auth" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.957537333Z level=info msg="Migration successfully executed" id="Add OAuth expiry to user_auth" duration=5.527672ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.961189061Z level=info msg="Executing migration" id="Add index to user_id column in user_auth" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.962067782Z level=info msg="Migration successfully executed" id="Add index to user_id column in user_auth" duration=878.651µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.970929867Z level=info msg="Executing migration" id="Add OAuth ID token to user_auth" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.978143961Z level=info msg="Migration successfully executed" id="Add OAuth ID token to user_auth" duration=7.215524ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.983614472Z level=info msg="Executing migration" id="create server_lock table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.984441473Z level=info msg="Migration successfully executed" id="create server_lock table" duration=819.681µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.987695825Z level=info msg="Executing migration" id="add index server_lock.operation_uid" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.988756409Z level=info msg="Migration successfully executed" id="add index server_lock.operation_uid" duration=1.059824ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.99577612Z level=info msg="Executing migration" id="create user auth token table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:02.996597171Z level=info msg="Migration successfully executed" id="create user auth token table" duration=820.951µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.001197481Z level=info msg="Executing migration" id="add unique index user_auth_token.auth_token" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.003146286Z level=info msg="Migration successfully executed" id="add unique index user_auth_token.auth_token" duration=1.953226ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.006771993Z level=info msg="Executing migration" id="add unique index user_auth_token.prev_auth_token" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.007755945Z level=info msg="Migration successfully executed" id="add unique index user_auth_token.prev_auth_token" duration=983.772µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.037830394Z level=info msg="Executing migration" id="add index user_auth_token.user_id" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.039468475Z level=info msg="Migration successfully executed" id="add index user_auth_token.user_id" duration=1.638302ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.046979511Z level=info msg="Executing migration" id="Add revoked_at to the user auth token" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.052746966Z level=info msg="Migration successfully executed" id="Add revoked_at to the user auth token" duration=5.767325ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.056806248Z level=info msg="Executing migration" id="add index user_auth_token.revoked_at" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.057767911Z level=info msg="Migration successfully executed" id="add index user_auth_token.revoked_at" duration=961.473µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.062584243Z level=info msg="Executing migration" id="create cache_data table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.063793748Z level=info msg="Migration successfully executed" id="create cache_data table" duration=1.181495ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.067480576Z level=info msg="Executing migration" id="add unique index cache_data.cache_key" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.069047546Z level=info msg="Migration successfully executed" id="add unique index cache_data.cache_key" duration=1.56491ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.072793235Z level=info msg="Executing migration" id="create short_url table v1" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.073648126Z level=info msg="Migration successfully executed" id="create short_url table v1" duration=847.15µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.080835988Z level=info msg="Executing migration" id="add index short_url.org_id-uid" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.082581741Z level=info msg="Migration successfully executed" id="add index short_url.org_id-uid" duration=1.745513ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.090104308Z level=info msg="Executing migration" id="alter table short_url alter column created_by type to bigint" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.090234219Z level=info msg="Migration successfully executed" id="alter table short_url alter column created_by type to bigint" duration=130.201µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.093170047Z level=info msg="Executing migration" id="delete alert_definition table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.093309189Z level=info msg="Migration successfully executed" id="delete alert_definition table" duration=150.192µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.097635785Z level=info msg="Executing migration" id="recreate alert_definition table" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.098564377Z level=info msg="Migration successfully executed" id="recreate alert_definition table" duration=926.832µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.102001531Z level=info msg="Executing migration" id="add index in alert_definition on org_id and title columns" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.102972614Z level=info msg="Migration successfully executed" id="add index in alert_definition on org_id and title columns" duration=971.443µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.106244526Z level=info msg="Executing migration" id="add index in alert_definition on org_id and uid columns" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.106942415Z level=info msg="Migration successfully executed" id="add index in alert_definition on org_id and uid columns" duration=697.399µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.112347245Z level=info msg="Executing migration" id="alter alert_definition table data column to mediumtext in mysql" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.112412426Z level=info msg="Migration successfully executed" id="alter alert_definition table data column to mediumtext in mysql" duration=65.691µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.115939921Z level=info msg="Executing migration" id="drop index in alert_definition on org_id and title columns" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.117366049Z level=info msg="Migration successfully executed" id="drop index in alert_definition on org_id and title columns" duration=1.434328ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.125849039Z level=info msg="Executing migration" id="drop index in alert_definition on org_id and uid columns" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.12674572Z level=info msg="Migration successfully executed" id="drop index in alert_definition on org_id and uid columns" duration=896.731µs 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.136453656Z level=info msg="Executing migration" id="add unique index in alert_definition on org_id and title columns" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.138177248Z level=info msg="Migration successfully executed" id="add unique index in alert_definition on org_id and title columns" duration=1.722632ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.141806265Z level=info msg="Executing migration" id="add unique index in alert_definition on org_id and uid columns" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.143356604Z level=info msg="Migration successfully executed" id="add unique index in alert_definition on org_id and uid columns" duration=1.54979ms 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.146935691Z level=info msg="Executing migration" id="Add column paused in alert_definition" 09:44:47 grafana | logger=migrator t=2024-07-04T09:42:03.153283593Z level=info msg="Migration successfully executed" id="Add column paused in alert_definition" duration=6.348732ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.15769602Z level=info msg="Executing migration" id="drop alert_definition table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.158665792Z level=info msg="Migration successfully executed" id="drop alert_definition table" duration=969.662µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.162006395Z level=info msg="Executing migration" id="delete alert_definition_version table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.162115046Z level=info msg="Migration successfully executed" id="delete alert_definition_version table" duration=108.981µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.167677908Z level=info msg="Executing migration" id="recreate alert_definition_version table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.168557059Z level=info msg="Migration successfully executed" id="recreate alert_definition_version table" duration=878.801µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.176852717Z level=info msg="Executing migration" id="add index in alert_definition_version table on alert_definition_id and version columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.178787562Z level=info msg="Migration successfully executed" id="add index in alert_definition_version table on alert_definition_id and version columns" duration=1.933755ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.182331897Z level=info msg="Executing migration" id="add index in alert_definition_version table on alert_definition_uid and version columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.183704465Z level=info msg="Migration successfully executed" id="add index in alert_definition_version table on alert_definition_uid and version columns" duration=1.379468ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.187760837Z level=info msg="Executing migration" id="alter alert_definition_version table data column to mediumtext in mysql" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.187834848Z level=info msg="Migration successfully executed" id="alter alert_definition_version table data column to mediumtext in mysql" duration=66.311µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.190995619Z level=info msg="Executing migration" id="drop alert_definition_version table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.191930711Z level=info msg="Migration successfully executed" id="drop alert_definition_version table" duration=934.632µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.195232504Z level=info msg="Executing migration" id="create alert_instance table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.19647489Z level=info msg="Migration successfully executed" id="create alert_instance table" duration=1.241646ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.200854646Z level=info msg="Executing migration" id="add index in alert_instance table on def_org_id, def_uid and current_state columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.20190597Z level=info msg="Migration successfully executed" id="add index in alert_instance table on def_org_id, def_uid and current_state columns" duration=1.051354ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.207002256Z level=info msg="Executing migration" id="add index in alert_instance table on def_org_id, current_state columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.207994898Z level=info msg="Migration successfully executed" id="add index in alert_instance table on def_org_id, current_state columns" duration=992.792µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.214114207Z level=info msg="Executing migration" id="add column current_state_end to alert_instance" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.220414498Z level=info msg="Migration successfully executed" id="add column current_state_end to alert_instance" duration=6.299701ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.22519951Z level=info msg="Executing migration" id="remove index def_org_id, def_uid, current_state on alert_instance" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.226480687Z level=info msg="Migration successfully executed" id="remove index def_org_id, def_uid, current_state on alert_instance" duration=1.281307ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.230056063Z level=info msg="Executing migration" id="remove index def_org_id, current_state on alert_instance" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.231188718Z level=info msg="Migration successfully executed" id="remove index def_org_id, current_state on alert_instance" duration=1.132955ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.235604755Z level=info msg="Executing migration" id="rename def_org_id to rule_org_id in alert_instance" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.261493388Z level=info msg="Migration successfully executed" id="rename def_org_id to rule_org_id in alert_instance" duration=25.886134ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.300072286Z level=info msg="Executing migration" id="rename def_uid to rule_uid in alert_instance" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.329628417Z level=info msg="Migration successfully executed" id="rename def_uid to rule_uid in alert_instance" duration=29.561811ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.333276644Z level=info msg="Executing migration" id="add index rule_org_id, rule_uid, current_state on alert_instance" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.334093315Z level=info msg="Migration successfully executed" id="add index rule_org_id, rule_uid, current_state on alert_instance" duration=808.591µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.337783092Z level=info msg="Executing migration" id="add index rule_org_id, current_state on alert_instance" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.338832646Z level=info msg="Migration successfully executed" id="add index rule_org_id, current_state on alert_instance" duration=1.049254ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.343282343Z level=info msg="Executing migration" id="add current_reason column related to current_state" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.352564073Z level=info msg="Migration successfully executed" id="add current_reason column related to current_state" duration=9.28107ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.356350232Z level=info msg="Executing migration" id="add result_fingerprint column to alert_instance" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.360525646Z level=info msg="Migration successfully executed" id="add result_fingerprint column to alert_instance" duration=4.176374ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.512702638Z level=info msg="Executing migration" id="create alert_rule table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.514010725Z level=info msg="Migration successfully executed" id="create alert_rule table" duration=1.310637ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.520804233Z level=info msg="Executing migration" id="add index in alert_rule on org_id and title columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.521951238Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id and title columns" duration=1.147705ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.525398352Z level=info msg="Executing migration" id="add index in alert_rule on org_id and uid columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.526523907Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id and uid columns" duration=1.125695ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.530205854Z level=info msg="Executing migration" id="add index in alert_rule on org_id, namespace_uid, group_uid columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.531966017Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, namespace_uid, group_uid columns" duration=1.758843ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.537351627Z level=info msg="Executing migration" id="alter alert_rule table data column to mediumtext in mysql" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.537486748Z level=info msg="Migration successfully executed" id="alter alert_rule table data column to mediumtext in mysql" duration=135.521µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.541243337Z level=info msg="Executing migration" id="add column for to alert_rule" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.548844035Z level=info msg="Migration successfully executed" id="add column for to alert_rule" duration=7.599678ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.551739552Z level=info msg="Executing migration" id="add column annotations to alert_rule" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.555986657Z level=info msg="Migration successfully executed" id="add column annotations to alert_rule" duration=4.247035ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.565446119Z level=info msg="Executing migration" id="add column labels to alert_rule" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.574902191Z level=info msg="Migration successfully executed" id="add column labels to alert_rule" duration=9.460962ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.578134173Z level=info msg="Executing migration" id="remove unique index from alert_rule on org_id, title columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.578835032Z level=info msg="Migration successfully executed" id="remove unique index from alert_rule on org_id, title columns" duration=700.939µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.582062383Z level=info msg="Executing migration" id="add index in alert_rule on org_id, namespase_uid and title columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.582815363Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, namespase_uid and title columns" duration=752.62µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.593861815Z level=info msg="Executing migration" id="add dashboard_uid column to alert_rule" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.603123505Z level=info msg="Migration successfully executed" id="add dashboard_uid column to alert_rule" duration=9.2747ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.606735581Z level=info msg="Executing migration" id="add panel_id column to alert_rule" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.610929325Z level=info msg="Migration successfully executed" id="add panel_id column to alert_rule" duration=4.193644ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.682460448Z level=info msg="Executing migration" id="add index in alert_rule on org_id, dashboard_uid and panel_id columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.684001718Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, dashboard_uid and panel_id columns" duration=1.54047ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.68955965Z level=info msg="Executing migration" id="add rule_group_idx column to alert_rule" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.697119197Z level=info msg="Migration successfully executed" id="add rule_group_idx column to alert_rule" duration=7.560647ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.70273698Z level=info msg="Executing migration" id="add is_paused column to alert_rule table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.708613155Z level=info msg="Migration successfully executed" id="add is_paused column to alert_rule table" duration=5.875935ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.714169237Z level=info msg="Executing migration" id="fix is_paused column for alert_rule table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.714233048Z level=info msg="Migration successfully executed" id="fix is_paused column for alert_rule table" duration=64.671µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.718361211Z level=info msg="Executing migration" id="create alert_rule_version table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.719429345Z level=info msg="Migration successfully executed" id="create alert_rule_version table" duration=1.067704ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.723133743Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_org_id, rule_uid and version columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.724709623Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_org_id, rule_uid and version columns" duration=1.57554ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.729845849Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_org_id, rule_namespace_uid and rule_group columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.730845552Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_org_id, rule_namespace_uid and rule_group columns" duration=999.653µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.735418822Z level=info msg="Executing migration" id="alter alert_rule_version table data column to mediumtext in mysql" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.735482492Z level=info msg="Migration successfully executed" id="alter alert_rule_version table data column to mediumtext in mysql" duration=64.381µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.739399813Z level=info msg="Executing migration" id="add column for to alert_rule_version" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.748135235Z level=info msg="Migration successfully executed" id="add column for to alert_rule_version" duration=8.740172ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.75390252Z level=info msg="Executing migration" id="add column annotations to alert_rule_version" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.760010149Z level=info msg="Migration successfully executed" id="add column annotations to alert_rule_version" duration=6.106629ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.765167885Z level=info msg="Executing migration" id="add column labels to alert_rule_version" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.771640219Z level=info msg="Migration successfully executed" id="add column labels to alert_rule_version" duration=6.466114ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.775156494Z level=info msg="Executing migration" id="add rule_group_idx column to alert_rule_version" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.781343264Z level=info msg="Migration successfully executed" id="add rule_group_idx column to alert_rule_version" duration=6.19066ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.787767246Z level=info msg="Executing migration" id="add is_paused column to alert_rule_versions table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.794096288Z level=info msg="Migration successfully executed" id="add is_paused column to alert_rule_versions table" duration=6.337492ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.798424454Z level=info msg="Executing migration" id="fix is_paused column for alert_rule_version table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.798492755Z level=info msg="Migration successfully executed" id="fix is_paused column for alert_rule_version table" duration=69.001µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.80433315Z level=info msg="Executing migration" id=create_alert_configuration_table 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.806076883Z level=info msg="Migration successfully executed" id=create_alert_configuration_table duration=1.733062ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.809410776Z level=info msg="Executing migration" id="Add column default in alert_configuration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.815754698Z level=info msg="Migration successfully executed" id="Add column default in alert_configuration" duration=6.343502ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.819168422Z level=info msg="Executing migration" id="alert alert_configuration alertmanager_configuration column from TEXT to MEDIUMTEXT if mysql" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.819288923Z level=info msg="Migration successfully executed" id="alert alert_configuration alertmanager_configuration column from TEXT to MEDIUMTEXT if mysql" duration=120.391µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.822455804Z level=info msg="Executing migration" id="add column org_id in alert_configuration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.828757455Z level=info msg="Migration successfully executed" id="add column org_id in alert_configuration" duration=6.301291ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.833096871Z level=info msg="Executing migration" id="add index in alert_configuration table on org_id column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.834227256Z level=info msg="Migration successfully executed" id="add index in alert_configuration table on org_id column" duration=1.130375ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.83996457Z level=info msg="Executing migration" id="add configuration_hash column to alert_configuration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.846631486Z level=info msg="Migration successfully executed" id="add configuration_hash column to alert_configuration" duration=6.669806ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.892146572Z level=info msg="Executing migration" id=create_ngalert_configuration_table 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.893816804Z level=info msg="Migration successfully executed" id=create_ngalert_configuration_table duration=1.670032ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.897532631Z level=info msg="Executing migration" id="add index in ngalert_configuration on org_id column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.900337148Z level=info msg="Migration successfully executed" id="add index in ngalert_configuration on org_id column" duration=2.804407ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.910460128Z level=info msg="Executing migration" id="add column send_alerts_to in ngalert_configuration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.917092324Z level=info msg="Migration successfully executed" id="add column send_alerts_to in ngalert_configuration" duration=6.631436ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.920121203Z level=info msg="Executing migration" id="create provenance_type table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.920962974Z level=info msg="Migration successfully executed" id="create provenance_type table" duration=836.19µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.92530319Z level=info msg="Executing migration" id="add index to uniquify (record_key, record_type, org_id) columns" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.926473255Z level=info msg="Migration successfully executed" id="add index to uniquify (record_key, record_type, org_id) columns" duration=1.169805ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.931034914Z level=info msg="Executing migration" id="create alert_image table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.93233272Z level=info msg="Migration successfully executed" id="create alert_image table" duration=1.298436ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.935352999Z level=info msg="Executing migration" id="add unique index on token to alert_image table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.936425243Z level=info msg="Migration successfully executed" id="add unique index on token to alert_image table" duration=1.072304ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.939581684Z level=info msg="Executing migration" id="support longer URLs in alert_image table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.939751386Z level=info msg="Migration successfully executed" id="support longer URLs in alert_image table" duration=169.352µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.94472523Z level=info msg="Executing migration" id=create_alert_configuration_history_table 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.945820544Z level=info msg="Migration successfully executed" id=create_alert_configuration_history_table duration=1.094824ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.951742291Z level=info msg="Executing migration" id="drop non-unique orgID index on alert_configuration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.953733766Z level=info msg="Migration successfully executed" id="drop non-unique orgID index on alert_configuration" duration=1.991165ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.956905077Z level=info msg="Executing migration" id="drop unique orgID index on alert_configuration if exists" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.957378973Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop unique orgID index on alert_configuration if exists" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.960788727Z level=info msg="Executing migration" id="extract alertmanager configuration history to separate table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.961207503Z level=info msg="Migration successfully executed" id="extract alertmanager configuration history to separate table" duration=418.556µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.964225722Z level=info msg="Executing migration" id="add unique index on orgID to alert_configuration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.965390407Z level=info msg="Migration successfully executed" id="add unique index on orgID to alert_configuration" duration=1.156385ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.968399906Z level=info msg="Executing migration" id="add last_applied column to alert_configuration_history" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.975467897Z level=info msg="Migration successfully executed" id="add last_applied column to alert_configuration_history" duration=7.066171ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.980306319Z level=info msg="Executing migration" id="create library_element table v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.981384583Z level=info msg="Migration successfully executed" id="create library_element table v1" duration=1.078154ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.989348126Z level=info msg="Executing migration" id="add index library_element org_id-folder_id-name-kind" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.990579692Z level=info msg="Migration successfully executed" id="add index library_element org_id-folder_id-name-kind" duration=1.230626ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.993868294Z level=info msg="Executing migration" id="create library_element_connection table v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:03.995481665Z level=info msg="Migration successfully executed" id="create library_element_connection table v1" duration=1.612661ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.034277985Z level=info msg="Executing migration" id="add index library_element_connection element_id-kind-connection_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.036724657Z level=info msg="Migration successfully executed" id="add index library_element_connection element_id-kind-connection_id" duration=2.447222ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.040374644Z level=info msg="Executing migration" id="add unique index library_element org_id_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.041490528Z level=info msg="Migration successfully executed" id="add unique index library_element org_id_uid" duration=1.115734ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.044785001Z level=info msg="Executing migration" id="increase max description length to 2048" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.044815971Z level=info msg="Migration successfully executed" id="increase max description length to 2048" duration=31.95µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.050660326Z level=info msg="Executing migration" id="alter library_element model to mediumtext" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.050727587Z level=info msg="Migration successfully executed" id="alter library_element model to mediumtext" duration=68.071µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.053158548Z level=info msg="Executing migration" id="add library_element folder uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.063961828Z level=info msg="Migration successfully executed" id="add library_element folder uid" duration=10.81017ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.067153449Z level=info msg="Executing migration" id="populate library_element folder_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.067550584Z level=info msg="Migration successfully executed" id="populate library_element folder_uid" duration=396.965µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.070839196Z level=info msg="Executing migration" id="add index library_element org_id-folder_uid-name-kind" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.071829709Z level=info msg="Migration successfully executed" id="add index library_element org_id-folder_uid-name-kind" duration=990.403µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.074732477Z level=info msg="Executing migration" id="clone move dashboard alerts to unified alerting" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.075187492Z level=info msg="Migration successfully executed" id="clone move dashboard alerts to unified alerting" duration=454.125µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.078309363Z level=info msg="Executing migration" id="create data_keys table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.079849442Z level=info msg="Migration successfully executed" id="create data_keys table" duration=1.539339ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.083170525Z level=info msg="Executing migration" id="create secrets table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.084651104Z level=info msg="Migration successfully executed" id="create secrets table" duration=1.479999ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.091010457Z level=info msg="Executing migration" id="rename data_keys name column to id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.12537816Z level=info msg="Migration successfully executed" id="rename data_keys name column to id" duration=34.367903ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.135291217Z level=info msg="Executing migration" id="add name column into data_keys" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.144202542Z level=info msg="Migration successfully executed" id="add name column into data_keys" duration=8.912115ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.147363213Z level=info msg="Executing migration" id="copy data_keys id column values into name" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.147638107Z level=info msg="Migration successfully executed" id="copy data_keys id column values into name" duration=284.134µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.150931709Z level=info msg="Executing migration" id="rename data_keys name column to label" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.181398942Z level=info msg="Migration successfully executed" id="rename data_keys name column to label" duration=30.467573ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.18438955Z level=info msg="Executing migration" id="rename data_keys id column back to name" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.215008925Z level=info msg="Migration successfully executed" id="rename data_keys id column back to name" duration=30.618905ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.675712845Z level=info msg="Executing migration" id="create kv_store table v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.677430877Z level=info msg="Migration successfully executed" id="create kv_store table v1" duration=1.722262ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.681959935Z level=info msg="Executing migration" id="add index kv_store.org_id-namespace-key" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.684016482Z level=info msg="Migration successfully executed" id="add index kv_store.org_id-namespace-key" duration=2.059687ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.687824281Z level=info msg="Executing migration" id="update dashboard_uid and panel_id from existing annotations" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.688205346Z level=info msg="Migration successfully executed" id="update dashboard_uid and panel_id from existing annotations" duration=372.515µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.691484898Z level=info msg="Executing migration" id="create permission table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.692454491Z level=info msg="Migration successfully executed" id="create permission table" duration=968.993µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.695806704Z level=info msg="Executing migration" id="add unique index permission.role_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.696906348Z level=info msg="Migration successfully executed" id="add unique index permission.role_id" duration=1.099684ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.700055139Z level=info msg="Executing migration" id="add unique index role_id_action_scope" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.701280184Z level=info msg="Migration successfully executed" id="add unique index role_id_action_scope" duration=1.216945ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.706293229Z level=info msg="Executing migration" id="create role table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.707375843Z level=info msg="Migration successfully executed" id="create role table" duration=1.082544ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.714642567Z level=info msg="Executing migration" id="add column display_name" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.725923912Z level=info msg="Migration successfully executed" id="add column display_name" duration=11.283015ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.72888513Z level=info msg="Executing migration" id="add column group_name" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.734641204Z level=info msg="Migration successfully executed" id="add column group_name" duration=5.751384ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.738743357Z level=info msg="Executing migration" id="add index role.org_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.739807441Z level=info msg="Migration successfully executed" id="add index role.org_id" duration=1.064054ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.743019113Z level=info msg="Executing migration" id="add unique index role_org_id_name" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.744099417Z level=info msg="Migration successfully executed" id="add unique index role_org_id_name" duration=1.086584ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.747848875Z level=info msg="Executing migration" id="add index role_org_id_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.748958789Z level=info msg="Migration successfully executed" id="add index role_org_id_uid" duration=1.109814ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.752454104Z level=info msg="Executing migration" id="create team role table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.75370009Z level=info msg="Migration successfully executed" id="create team role table" duration=1.244096ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.758922307Z level=info msg="Executing migration" id="add index team_role.org_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.76069344Z level=info msg="Migration successfully executed" id="add index team_role.org_id" duration=1.770823ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.765626534Z level=info msg="Executing migration" id="add unique index team_role_org_id_team_id_role_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.766732728Z level=info msg="Migration successfully executed" id="add unique index team_role_org_id_team_id_role_id" duration=1.105964ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.772014616Z level=info msg="Executing migration" id="add index team_role.team_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.773929581Z level=info msg="Migration successfully executed" id="add index team_role.team_id" duration=1.914655ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.777917072Z level=info msg="Executing migration" id="create user role table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.778811654Z level=info msg="Migration successfully executed" id="create user role table" duration=895.282µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.783627106Z level=info msg="Executing migration" id="add index user_role.org_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.784939273Z level=info msg="Migration successfully executed" id="add index user_role.org_id" duration=1.311257ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.789191788Z level=info msg="Executing migration" id="add unique index user_role_org_id_user_id_role_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.790603026Z level=info msg="Migration successfully executed" id="add unique index user_role_org_id_user_id_role_id" duration=1.410468ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.796987338Z level=info msg="Executing migration" id="add index user_role.user_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.798171484Z level=info msg="Migration successfully executed" id="add index user_role.user_id" duration=1.184396ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.804290152Z level=info msg="Executing migration" id="create builtin role table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.805255385Z level=info msg="Migration successfully executed" id="create builtin role table" duration=964.763µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.809996716Z level=info msg="Executing migration" id="add index builtin_role.role_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.812260895Z level=info msg="Migration successfully executed" id="add index builtin_role.role_id" duration=2.262129ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.816246996Z level=info msg="Executing migration" id="add index builtin_role.name" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.817671205Z level=info msg="Migration successfully executed" id="add index builtin_role.name" duration=1.425249ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.821601146Z level=info msg="Executing migration" id="Add column org_id to builtin_role table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.829824162Z level=info msg="Migration successfully executed" id="Add column org_id to builtin_role table" duration=8.222596ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.834822476Z level=info msg="Executing migration" id="add index builtin_role.org_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.83592111Z level=info msg="Migration successfully executed" id="add index builtin_role.org_id" duration=1.098474ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.841397301Z level=info msg="Executing migration" id="add unique index builtin_role_org_id_role_id_role" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.842625386Z level=info msg="Migration successfully executed" id="add unique index builtin_role_org_id_role_id_role" duration=1.229855ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.848739496Z level=info msg="Executing migration" id="Remove unique index role_org_id_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.849756379Z level=info msg="Migration successfully executed" id="Remove unique index role_org_id_uid" duration=1.016763ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.855551093Z level=info msg="Executing migration" id="add unique index role.uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.856978462Z level=info msg="Migration successfully executed" id="add unique index role.uid" duration=1.445339ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.862809977Z level=info msg="Executing migration" id="create seed assignment table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.864223595Z level=info msg="Migration successfully executed" id="create seed assignment table" duration=1.413098ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.868740673Z level=info msg="Executing migration" id="add unique index builtin_role_role_name" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.870358664Z level=info msg="Migration successfully executed" id="add unique index builtin_role_role_name" duration=1.617911ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.877332344Z level=info msg="Executing migration" id="add column hidden to role table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.888468388Z level=info msg="Migration successfully executed" id="add column hidden to role table" duration=11.114373ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.893886987Z level=info msg="Executing migration" id="permission kind migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.90260004Z level=info msg="Migration successfully executed" id="permission kind migration" duration=8.712823ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.907048707Z level=info msg="Executing migration" id="permission attribute migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.913146176Z level=info msg="Migration successfully executed" id="permission attribute migration" duration=6.096809ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.921944809Z level=info msg="Executing migration" id="permission identifier migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:04.93051141Z level=info msg="Migration successfully executed" id="permission identifier migration" duration=8.566331ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.782356732Z level=info msg="Executing migration" id="add permission identifier index" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.784004323Z level=info msg="Migration successfully executed" id="add permission identifier index" duration=1.653861ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.79228053Z level=info msg="Executing migration" id="add permission action scope role_id index" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.794059793Z level=info msg="Migration successfully executed" id="add permission action scope role_id index" duration=1.783873ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.800282023Z level=info msg="Executing migration" id="remove permission role_id action scope index" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.80160141Z level=info msg="Migration successfully executed" id="remove permission role_id action scope index" duration=1.323167ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.81171835Z level=info msg="Executing migration" id="create query_history table v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.813079268Z level=info msg="Migration successfully executed" id="create query_history table v1" duration=1.363287ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.820651205Z level=info msg="Executing migration" id="add index query_history.org_id-created_by-datasource_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.822908564Z level=info msg="Migration successfully executed" id="add index query_history.org_id-created_by-datasource_uid" duration=2.257169ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.826897465Z level=info msg="Executing migration" id="alter table query_history alter column created_by type to bigint" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.827126678Z level=info msg="Migration successfully executed" id="alter table query_history alter column created_by type to bigint" duration=231.903µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.832551338Z level=info msg="Executing migration" id="rbac disabled migrator" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.832589799Z level=info msg="Migration successfully executed" id="rbac disabled migrator" duration=40.321µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.837504882Z level=info msg="Executing migration" id="teams permissions migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.83810539Z level=info msg="Migration successfully executed" id="teams permissions migration" duration=600.788µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.843258886Z level=info msg="Executing migration" id="dashboard permissions" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.844500582Z level=info msg="Migration successfully executed" id="dashboard permissions" duration=1.244186ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.848393442Z level=info msg="Executing migration" id="dashboard permissions uid scopes" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.849317284Z level=info msg="Migration successfully executed" id="dashboard permissions uid scopes" duration=923.292µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.855736937Z level=info msg="Executing migration" id="drop managed folder create actions" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.856315424Z level=info msg="Migration successfully executed" id="drop managed folder create actions" duration=576.767µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.861949477Z level=info msg="Executing migration" id="alerting notification permissions" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.862590535Z level=info msg="Migration successfully executed" id="alerting notification permissions" duration=641.528µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.866926751Z level=info msg="Executing migration" id="create query_history_star table v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.867960424Z level=info msg="Migration successfully executed" id="create query_history_star table v1" duration=1.033303ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.87233813Z level=info msg="Executing migration" id="add index query_history.user_id-query_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.873621617Z level=info msg="Migration successfully executed" id="add index query_history.user_id-query_uid" duration=1.283327ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.87931102Z level=info msg="Executing migration" id="add column org_id in query_history_star" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.891141223Z level=info msg="Migration successfully executed" id="add column org_id in query_history_star" duration=11.829772ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.897824389Z level=info msg="Executing migration" id="alter table query_history_star_mig column user_id type to bigint" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.898059762Z level=info msg="Migration successfully executed" id="alter table query_history_star_mig column user_id type to bigint" duration=235.153µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.903848416Z level=info msg="Executing migration" id="create correlation table v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.906400819Z level=info msg="Migration successfully executed" id="create correlation table v1" duration=2.550773ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.911416894Z level=info msg="Executing migration" id="add index correlations.uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.912946803Z level=info msg="Migration successfully executed" id="add index correlations.uid" duration=1.528789ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.916668781Z level=info msg="Executing migration" id="add index correlations.source_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.918484965Z level=info msg="Migration successfully executed" id="add index correlations.source_uid" duration=1.817484ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.924789846Z level=info msg="Executing migration" id="add correlation config column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.934093486Z level=info msg="Migration successfully executed" id="add correlation config column" duration=9.303049ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.938991909Z level=info msg="Executing migration" id="drop index IDX_correlation_uid - v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.939789469Z level=info msg="Migration successfully executed" id="drop index IDX_correlation_uid - v1" duration=797.01µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.945060907Z level=info msg="Executing migration" id="drop index IDX_correlation_source_uid - v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.946807209Z level=info msg="Migration successfully executed" id="drop index IDX_correlation_source_uid - v1" duration=1.755892ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.954039312Z level=info msg="Executing migration" id="Rename table correlation to correlation_tmp_qwerty - v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.979847775Z level=info msg="Migration successfully executed" id="Rename table correlation to correlation_tmp_qwerty - v1" duration=25.806023ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.987008297Z level=info msg="Executing migration" id="create correlation v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.988434875Z level=info msg="Migration successfully executed" id="create correlation v2" duration=1.425898ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.996409668Z level=info msg="Executing migration" id="create index IDX_correlation_uid - v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:07.998340783Z level=info msg="Migration successfully executed" id="create index IDX_correlation_uid - v2" duration=1.930565ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.003636331Z level=info msg="Executing migration" id="create index IDX_correlation_source_uid - v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.005618126Z level=info msg="Migration successfully executed" id="create index IDX_correlation_source_uid - v2" duration=1.982215ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.010480399Z level=info msg="Executing migration" id="create index IDX_correlation_org_id - v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.011643084Z level=info msg="Migration successfully executed" id="create index IDX_correlation_org_id - v2" duration=1.162585ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.015263761Z level=info msg="Executing migration" id="copy correlation v1 to v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.015776167Z level=info msg="Migration successfully executed" id="copy correlation v1 to v2" duration=522.737µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.020398807Z level=info msg="Executing migration" id="drop correlation_tmp_qwerty" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.021200717Z level=info msg="Migration successfully executed" id="drop correlation_tmp_qwerty" duration=805.61µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.024140605Z level=info msg="Executing migration" id="add provisioning column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.030254984Z level=info msg="Migration successfully executed" id="add provisioning column" duration=6.113959ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.035915536Z level=info msg="Executing migration" id="create entity_events table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.036524854Z level=info msg="Migration successfully executed" id="create entity_events table" duration=609.448µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.04083569Z level=info msg="Executing migration" id="create dashboard public config v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.041542119Z level=info msg="Migration successfully executed" id="create dashboard public config v1" duration=706.359µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.046694495Z level=info msg="Executing migration" id="drop index UQE_dashboard_public_config_uid - v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.047390324Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop index UQE_dashboard_public_config_uid - v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.050763468Z level=info msg="Executing migration" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.051284424Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.056126127Z level=info msg="Executing migration" id="Drop old dashboard public config table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.056946097Z level=info msg="Migration successfully executed" id="Drop old dashboard public config table" duration=826.861µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.060726876Z level=info msg="Executing migration" id="recreate dashboard public config v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.062532119Z level=info msg="Migration successfully executed" id="recreate dashboard public config v1" duration=1.804413ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.069475308Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_uid - v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.070735265Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_uid - v1" duration=1.252457ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.074486493Z level=info msg="Executing migration" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.076214445Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v1" duration=1.727792ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.080248207Z level=info msg="Executing migration" id="drop index UQE_dashboard_public_config_uid - v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.082145742Z level=info msg="Migration successfully executed" id="drop index UQE_dashboard_public_config_uid - v2" duration=1.897935ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.087836275Z level=info msg="Executing migration" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.089108121Z level=info msg="Migration successfully executed" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v2" duration=1.271586ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.092354143Z level=info msg="Executing migration" id="Drop public config table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.093437597Z level=info msg="Migration successfully executed" id="Drop public config table" duration=1.081154ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.09757482Z level=info msg="Executing migration" id="Recreate dashboard public config v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.099710628Z level=info msg="Migration successfully executed" id="Recreate dashboard public config v2" duration=2.134608ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.170160234Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_uid - v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.172039608Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_uid - v2" duration=1.880714ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.177705921Z level=info msg="Executing migration" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.179063519Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v2" duration=1.358268ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.184386817Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_access_token - v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.186290312Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_access_token - v2" duration=1.903425ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.192063846Z level=info msg="Executing migration" id="Rename table dashboard_public_config to dashboard_public - v2" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.213550062Z level=info msg="Migration successfully executed" id="Rename table dashboard_public_config to dashboard_public - v2" duration=21.486666ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.217732106Z level=info msg="Executing migration" id="add annotations_enabled column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.224727186Z level=info msg="Migration successfully executed" id="add annotations_enabled column" duration=6.98713ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.229906143Z level=info msg="Executing migration" id="add time_selection_enabled column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.239383775Z level=info msg="Migration successfully executed" id="add time_selection_enabled column" duration=9.471162ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.243563139Z level=info msg="Executing migration" id="delete orphaned public dashboards" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.243796112Z level=info msg="Migration successfully executed" id="delete orphaned public dashboards" duration=233.073µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.248487182Z level=info msg="Executing migration" id="add share column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.257925074Z level=info msg="Migration successfully executed" id="add share column" duration=9.437272ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.262530003Z level=info msg="Executing migration" id="backfill empty share column fields with default of public" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.262867837Z level=info msg="Migration successfully executed" id="backfill empty share column fields with default of public" duration=340.194µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.268677122Z level=info msg="Executing migration" id="create file table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.270554456Z level=info msg="Migration successfully executed" id="create file table" duration=1.876594ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.274646679Z level=info msg="Executing migration" id="file table idx: path natural pk" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.275747793Z level=info msg="Migration successfully executed" id="file table idx: path natural pk" duration=1.100585ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.285476888Z level=info msg="Executing migration" id="file table idx: parent_folder_path_hash fast folder retrieval" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.287240861Z level=info msg="Migration successfully executed" id="file table idx: parent_folder_path_hash fast folder retrieval" duration=1.786553ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.29259643Z level=info msg="Executing migration" id="create file_meta table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.294073589Z level=info msg="Migration successfully executed" id="create file_meta table" duration=1.476949ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.299861523Z level=info msg="Executing migration" id="file table idx: path key" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.300987768Z level=info msg="Migration successfully executed" id="file table idx: path key" duration=1.125585ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.305274733Z level=info msg="Executing migration" id="set path collation in file table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.305493456Z level=info msg="Migration successfully executed" id="set path collation in file table" duration=218.923µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.312781329Z level=info msg="Executing migration" id="migrate contents column to mediumblob for MySQL" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.312920311Z level=info msg="Migration successfully executed" id="migrate contents column to mediumblob for MySQL" duration=138.912µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.319488596Z level=info msg="Executing migration" id="managed permissions migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.320497599Z level=info msg="Migration successfully executed" id="managed permissions migration" duration=1.008183ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.324174366Z level=info msg="Executing migration" id="managed folder permissions alert actions migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.32445304Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions migration" duration=278.184µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.330937443Z level=info msg="Executing migration" id="RBAC action name migrator" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.33222371Z level=info msg="Migration successfully executed" id="RBAC action name migrator" duration=1.286107ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.339211689Z level=info msg="Executing migration" id="Add UID column to playlist" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.35476812Z level=info msg="Migration successfully executed" id="Add UID column to playlist" duration=15.557771ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.359887865Z level=info msg="Executing migration" id="Update uid column values in playlist" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.36023752Z level=info msg="Migration successfully executed" id="Update uid column values in playlist" duration=349.965µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.364042029Z level=info msg="Executing migration" id="Add index for uid in playlist" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.365733611Z level=info msg="Migration successfully executed" id="Add index for uid in playlist" duration=1.692862ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.372724351Z level=info msg="Executing migration" id="update group index for alert rules" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.373262228Z level=info msg="Migration successfully executed" id="update group index for alert rules" duration=538.687µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.380151286Z level=info msg="Executing migration" id="managed folder permissions alert actions repeated migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.380485871Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions repeated migration" duration=333.975µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.387143756Z level=info msg="Executing migration" id="admin only folder/dashboard permission" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.387981327Z level=info msg="Migration successfully executed" id="admin only folder/dashboard permission" duration=837.291µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.393827332Z level=info msg="Executing migration" id="add action column to seed_assignment" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.403055011Z level=info msg="Migration successfully executed" id="add action column to seed_assignment" duration=9.235779ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.407342036Z level=info msg="Executing migration" id="add scope column to seed_assignment" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.415182117Z level=info msg="Migration successfully executed" id="add scope column to seed_assignment" duration=7.839221ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.42088582Z level=info msg="Executing migration" id="remove unique index builtin_role_role_name before nullable update" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.422106626Z level=info msg="Migration successfully executed" id="remove unique index builtin_role_role_name before nullable update" duration=1.213996ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.429221278Z level=info msg="Executing migration" id="update seed_assignment role_name column to nullable" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.509552771Z level=info msg="Migration successfully executed" id="update seed_assignment role_name column to nullable" duration=80.327853ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.549682658Z level=info msg="Executing migration" id="add unique index builtin_role_name back" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.552373473Z level=info msg="Migration successfully executed" id="add unique index builtin_role_name back" duration=2.692795ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.557451448Z level=info msg="Executing migration" id="add unique index builtin_role_action_scope" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.558714634Z level=info msg="Migration successfully executed" id="add unique index builtin_role_action_scope" duration=1.255446ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.562451442Z level=info msg="Executing migration" id="add primary key to seed_assigment" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.589735193Z level=info msg="Migration successfully executed" id="add primary key to seed_assigment" duration=27.281771ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.596008554Z level=info msg="Executing migration" id="add origin column to seed_assignment" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.607116117Z level=info msg="Migration successfully executed" id="add origin column to seed_assignment" duration=11.099923ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.611402712Z level=info msg="Executing migration" id="add origin to plugin seed_assignment" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.611884198Z level=info msg="Migration successfully executed" id="add origin to plugin seed_assignment" duration=481.736µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.619228663Z level=info msg="Executing migration" id="prevent seeding OnCall access" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.619555627Z level=info msg="Migration successfully executed" id="prevent seeding OnCall access" duration=328.234µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.624881795Z level=info msg="Executing migration" id="managed folder permissions alert actions repeated fixed migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.625368071Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions repeated fixed migration" duration=485.766µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.630394966Z level=info msg="Executing migration" id="managed folder permissions library panel actions migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.630616529Z level=info msg="Migration successfully executed" id="managed folder permissions library panel actions migration" duration=221.553µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.634476049Z level=info msg="Executing migration" id="migrate external alertmanagers to datsourcse" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.634709482Z level=info msg="Migration successfully executed" id="migrate external alertmanagers to datsourcse" duration=234.093µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.640106131Z level=info msg="Executing migration" id="create folder table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.641884234Z level=info msg="Migration successfully executed" id="create folder table" duration=1.777583ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.648239146Z level=info msg="Executing migration" id="Add index for parent_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.649638234Z level=info msg="Migration successfully executed" id="Add index for parent_uid" duration=1.399228ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.654614988Z level=info msg="Executing migration" id="Add unique index for folder.uid and folder.org_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.656516542Z level=info msg="Migration successfully executed" id="Add unique index for folder.uid and folder.org_id" duration=1.901174ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.666066045Z level=info msg="Executing migration" id="Update folder title length" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.666102016Z level=info msg="Migration successfully executed" id="Update folder title length" duration=37.331µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.673934377Z level=info msg="Executing migration" id="Add unique index for folder.title and folder.parent_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.677452132Z level=info msg="Migration successfully executed" id="Add unique index for folder.title and folder.parent_uid" duration=3.517385ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.681612386Z level=info msg="Executing migration" id="Remove unique index for folder.title and folder.parent_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.683052484Z level=info msg="Migration successfully executed" id="Remove unique index for folder.title and folder.parent_uid" duration=1.439578ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.687636983Z level=info msg="Executing migration" id="Add unique index for title, parent_uid, and org_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.689281134Z level=info msg="Migration successfully executed" id="Add unique index for title, parent_uid, and org_id" duration=1.637981ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.693001212Z level=info msg="Executing migration" id="Sync dashboard and folder table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.693526979Z level=info msg="Migration successfully executed" id="Sync dashboard and folder table" duration=525.927µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.697904005Z level=info msg="Executing migration" id="Remove ghost folders from the folder table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.698174258Z level=info msg="Migration successfully executed" id="Remove ghost folders from the folder table" duration=270.513µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.703366255Z level=info msg="Executing migration" id="Remove unique index UQE_folder_uid_org_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.705592384Z level=info msg="Migration successfully executed" id="Remove unique index UQE_folder_uid_org_id" duration=2.227849ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.720203882Z level=info msg="Executing migration" id="Add unique index UQE_folder_org_id_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.722475181Z level=info msg="Migration successfully executed" id="Add unique index UQE_folder_org_id_uid" duration=2.270859ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.727760739Z level=info msg="Executing migration" id="Remove unique index UQE_folder_title_parent_uid_org_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.729707864Z level=info msg="Migration successfully executed" id="Remove unique index UQE_folder_title_parent_uid_org_id" duration=1.946985ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.735816123Z level=info msg="Executing migration" id="Add unique index UQE_folder_org_id_parent_uid_title" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.737061599Z level=info msg="Migration successfully executed" id="Add unique index UQE_folder_org_id_parent_uid_title" duration=1.245236ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.740678496Z level=info msg="Executing migration" id="Remove index IDX_folder_parent_uid_org_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.742485949Z level=info msg="Migration successfully executed" id="Remove index IDX_folder_parent_uid_org_id" duration=1.807543ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.746504011Z level=info msg="Executing migration" id="create anon_device table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.748735069Z level=info msg="Migration successfully executed" id="create anon_device table" duration=2.230438ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.756721792Z level=info msg="Executing migration" id="add unique index anon_device.device_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.75809605Z level=info msg="Migration successfully executed" id="add unique index anon_device.device_id" duration=1.373848ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.764058957Z level=info msg="Executing migration" id="add index anon_device.updated_at" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.76592302Z level=info msg="Migration successfully executed" id="add index anon_device.updated_at" duration=1.871314ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.806656134Z level=info msg="Executing migration" id="create signing_key table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.807960531Z level=info msg="Migration successfully executed" id="create signing_key table" duration=1.308267ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.812209846Z level=info msg="Executing migration" id="add unique index signing_key.key_id" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.814762639Z level=info msg="Migration successfully executed" id="add unique index signing_key.key_id" duration=2.565033ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.820512793Z level=info msg="Executing migration" id="set legacy alert migration status in kvstore" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.8226554Z level=info msg="Migration successfully executed" id="set legacy alert migration status in kvstore" duration=2.155127ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.833346418Z level=info msg="Executing migration" id="migrate record of created folders during legacy migration to kvstore" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.833860325Z level=info msg="Migration successfully executed" id="migrate record of created folders during legacy migration to kvstore" duration=519.417µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.840352868Z level=info msg="Executing migration" id="Add folder_uid for dashboard" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.853022501Z level=info msg="Migration successfully executed" id="Add folder_uid for dashboard" duration=12.667183ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.859622646Z level=info msg="Executing migration" id="Populate dashboard folder_uid column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.860466737Z level=info msg="Migration successfully executed" id="Populate dashboard folder_uid column" duration=846.601µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.864227035Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_folder_uid_title" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.864248975Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_folder_uid_title" duration=23.66µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.869677745Z level=info msg="Executing migration" id="Delete unique index for dashboard_org_id_folder_id_title" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.871000852Z level=info msg="Migration successfully executed" id="Delete unique index for dashboard_org_id_folder_id_title" duration=1.324017ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.877482636Z level=info msg="Executing migration" id="Delete unique index for dashboard_org_id_folder_uid_title" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.877532977Z level=info msg="Migration successfully executed" id="Delete unique index for dashboard_org_id_folder_uid_title" duration=46.96µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.882406219Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_folder_uid_title_is_folder" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.885052923Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_folder_uid_title_is_folder" duration=2.647144ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.891448756Z level=info msg="Executing migration" id="Restore index for dashboard_org_id_folder_id_title" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.893115637Z level=info msg="Migration successfully executed" id="Restore index for dashboard_org_id_folder_id_title" duration=1.668171ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.929231982Z level=info msg="Executing migration" id="create sso_setting table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.931501891Z level=info msg="Migration successfully executed" id="create sso_setting table" duration=2.266699ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.937550309Z level=info msg="Executing migration" id="copy kvstore migration status to each org" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.938880216Z level=info msg="Migration successfully executed" id="copy kvstore migration status to each org" duration=1.332487ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.943418754Z level=info msg="Executing migration" id="add back entry for orgid=0 migrated status" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.94387698Z level=info msg="Migration successfully executed" id="add back entry for orgid=0 migrated status" duration=460.006µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.950166441Z level=info msg="Executing migration" id="managed dashboard permissions annotation actions migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.951214385Z level=info msg="Migration successfully executed" id="managed dashboard permissions annotation actions migration" duration=1.047834ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.958717521Z level=info msg="Executing migration" id="create cloud_migration table v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.960282301Z level=info msg="Migration successfully executed" id="create cloud_migration table v1" duration=1.56536ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.963945419Z level=info msg="Executing migration" id="create cloud_migration_run table v1" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.965476298Z level=info msg="Migration successfully executed" id="create cloud_migration_run table v1" duration=1.530389ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.970535693Z level=info msg="Executing migration" id="add stack_id column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.981395633Z level=info msg="Migration successfully executed" id="add stack_id column" duration=10.84441ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.986777892Z level=info msg="Executing migration" id="add region_slug column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:08.997965466Z level=info msg="Migration successfully executed" id="add region_slug column" duration=11.193064ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.001959427Z level=info msg="Executing migration" id="add cluster_slug column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.011038924Z level=info msg="Migration successfully executed" id="add cluster_slug column" duration=9.087287ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.014968635Z level=info msg="Executing migration" id="add migration uid column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.02157612Z level=info msg="Migration successfully executed" id="add migration uid column" duration=6.607335ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.027249853Z level=info msg="Executing migration" id="Update uid column values for migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.027429225Z level=info msg="Migration successfully executed" id="Update uid column values for migration" duration=179.042µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.031275515Z level=info msg="Executing migration" id="Add unique index migration_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.03326151Z level=info msg="Migration successfully executed" id="Add unique index migration_uid" duration=1.985795ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.040546184Z level=info msg="Executing migration" id="add migration run uid column" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.050515722Z level=info msg="Migration successfully executed" id="add migration run uid column" duration=9.968048ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.053960446Z level=info msg="Executing migration" id="Update uid column values for migration run" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.054128679Z level=info msg="Migration successfully executed" id="Update uid column values for migration run" duration=168.153µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.059225724Z level=info msg="Executing migration" id="Add unique index migration_run_uid" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.06041546Z level=info msg="Migration successfully executed" id="Add unique index migration_run_uid" duration=1.185745ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.066030512Z level=info msg="Executing migration" id="alter kv_store.value to longtext" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.066100333Z level=info msg="Migration successfully executed" id="alter kv_store.value to longtext" duration=70.131µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.069901981Z level=info msg="Executing migration" id="add notification_settings column to alert_rule table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.082869788Z level=info msg="Migration successfully executed" id="add notification_settings column to alert_rule table" duration=12.939277ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.088643543Z level=info msg="Executing migration" id="add notification_settings column to alert_rule_version table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.098647421Z level=info msg="Migration successfully executed" id="add notification_settings column to alert_rule_version table" duration=10.003748ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.102867966Z level=info msg="Executing migration" id="removing scope from alert.instances:read action migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.103572185Z level=info msg="Migration successfully executed" id="removing scope from alert.instances:read action migration" duration=709.629µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.10792009Z level=info msg="Executing migration" id="managed folder permissions alerting silences actions migration" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.108304355Z level=info msg="Migration successfully executed" id="managed folder permissions alerting silences actions migration" duration=383.725µs 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.112397248Z level=info msg="Executing migration" id="add record column to alert_rule table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.124427193Z level=info msg="Migration successfully executed" id="add record column to alert_rule table" duration=12.030485ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.153459136Z level=info msg="Executing migration" id="add record column to alert_rule_version table" 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.167508727Z level=info msg="Migration successfully executed" id="add record column to alert_rule_version table" duration=14.052181ms 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.173099439Z level=info msg="migrations completed" performed=572 skipped=0 duration=9.400858662s 09:44:48 grafana | logger=migrator t=2024-07-04T09:42:09.173800848Z level=info msg="Unlocking database" 09:44:48 grafana | logger=sqlstore t=2024-07-04T09:42:09.19338936Z level=info msg="Created default admin" user=admin 09:44:48 grafana | logger=sqlstore t=2024-07-04T09:42:09.193632173Z level=info msg="Created default organization" 09:44:48 grafana | logger=secrets t=2024-07-04T09:42:09.199040142Z level=info msg="Envelope encryption state" enabled=true currentprovider=secretKey.v1 09:44:48 grafana | logger=plugin.angulardetectorsprovider.dynamic t=2024-07-04T09:42:09.273156886Z level=info msg="Restored cache from database" duration=894.922µs 09:44:48 grafana | logger=plugin.store t=2024-07-04T09:42:09.275108911Z level=info msg="Loading plugins..." 09:44:48 grafana | logger=plugins.registration t=2024-07-04T09:42:09.315627172Z level=error msg="Could not register plugin" pluginId=xychart error="plugin xychart is already registered" 09:44:48 grafana | logger=plugins.initialization t=2024-07-04T09:42:09.315664892Z level=error msg="Could not initialize plugin" pluginId=xychart error="plugin xychart is already registered" 09:44:48 grafana | logger=local.finder t=2024-07-04T09:42:09.315745753Z level=warn msg="Skipping finding plugins as directory does not exist" path=/usr/share/grafana/plugins-bundled 09:44:48 grafana | logger=plugin.store t=2024-07-04T09:42:09.315763084Z level=info msg="Plugins loaded" count=54 duration=40.655673ms 09:44:48 grafana | logger=query_data t=2024-07-04T09:42:09.320388333Z level=info msg="Query Service initialization" 09:44:48 grafana | logger=live.push_http t=2024-07-04T09:42:09.324440475Z level=info msg="Live Push Gateway initialization" 09:44:48 grafana | logger=ngalert.notifier.alertmanager org=1 t=2024-07-04T09:42:09.333320099Z level=info msg="Applying new configuration to Alertmanager" configHash=d2c56faca6af2a5772ff4253222f7386 09:44:48 grafana | logger=ngalert.state.manager t=2024-07-04T09:42:09.34424931Z level=info msg="Running in alternative execution of Error/NoData mode" 09:44:48 grafana | logger=infra.usagestats.collector t=2024-07-04T09:42:09.34969627Z level=info msg="registering usage stat providers" usageStatsProvidersLen=2 09:44:48 grafana | logger=provisioning.datasources t=2024-07-04T09:42:09.352446206Z level=info msg="inserting datasource from configuration" name=PolicyPrometheus uid=dkSf71fnz 09:44:48 grafana | logger=provisioning.alerting t=2024-07-04T09:42:09.378241487Z level=info msg="starting to provision alerting" 09:44:48 grafana | logger=provisioning.alerting t=2024-07-04T09:42:09.378302128Z level=info msg="finished to provision alerting" 09:44:48 grafana | logger=grafanaStorageLogger t=2024-07-04T09:42:09.379318891Z level=info msg="Storage starting" 09:44:48 grafana | logger=ngalert.state.manager t=2024-07-04T09:42:09.379190349Z level=info msg="Warming state cache for startup" 09:44:48 grafana | logger=ngalert.multiorg.alertmanager t=2024-07-04T09:42:09.381751152Z level=info msg="Starting MultiOrg Alertmanager" 09:44:48 grafana | logger=ngalert.state.manager t=2024-07-04T09:42:09.382265729Z level=info msg="State cache has been initialized" states=0 duration=3.07158ms 09:44:48 grafana | logger=ngalert.scheduler t=2024-07-04T09:42:09.38232567Z level=info msg="Starting scheduler" tickInterval=10s maxAttempts=1 09:44:48 grafana | logger=ticker t=2024-07-04T09:42:09.382456201Z level=info msg=starting first_tick=2024-07-04T09:42:10Z 09:44:48 grafana | logger=provisioning.dashboard t=2024-07-04T09:42:09.384283925Z level=info msg="starting to provision dashboards" 09:44:48 grafana | logger=http.server t=2024-07-04T09:42:09.384726801Z level=info msg="HTTP Server Listen" address=[::]:3000 protocol=http subUrl= socket= 09:44:48 grafana | logger=plugins.update.checker t=2024-07-04T09:42:09.450956602Z level=info msg="Update check succeeded" duration=69.138889ms 09:44:48 grafana | logger=grafana.update.checker t=2024-07-04T09:42:09.462851075Z level=info msg="Update check succeeded" duration=82.43078ms 09:44:48 grafana | logger=sqlstore.transactions t=2024-07-04T09:42:09.482918173Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=0 code="database is locked" 09:44:48 grafana | logger=sqlstore.transactions t=2024-07-04T09:42:09.490000285Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=0 code="database is locked" 09:44:48 grafana | logger=sqlstore.transactions t=2024-07-04T09:42:09.495965411Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=1 code="database is locked" 09:44:48 grafana | logger=sqlstore.transactions t=2024-07-04T09:42:09.502245312Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=1 code="database is locked" 09:44:48 grafana | logger=sqlstore.transactions t=2024-07-04T09:42:09.509511686Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=2 code="database is locked" 09:44:48 grafana | logger=sqlstore.transactions t=2024-07-04T09:42:09.514396858Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=2 code="database is locked" 09:44:48 grafana | logger=sqlstore.transactions t=2024-07-04T09:42:09.528971206Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=3 code="database is locked" 09:44:48 grafana | logger=plugin.angulardetectorsprovider.dynamic t=2024-07-04T09:42:09.531843133Z level=info msg="Patterns update finished" duration=149.928129ms 09:44:48 grafana | logger=sqlstore.transactions t=2024-07-04T09:42:09.552877513Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=0 code="database is locked" 09:44:48 grafana | logger=grafana-apiserver t=2024-07-04T09:42:09.707429121Z level=info msg="Adding GroupVersion playlist.grafana.app v0alpha1 to ResourceManager" 09:44:48 grafana | logger=grafana-apiserver t=2024-07-04T09:42:09.708337993Z level=info msg="Adding GroupVersion featuretoggle.grafana.app v0alpha1 to ResourceManager" 09:44:48 grafana | logger=provisioning.dashboard t=2024-07-04T09:42:09.909390459Z level=info msg="finished to provision dashboards" 09:44:48 grafana | logger=infra.usagestats t=2024-07-04T09:43:57.390096588Z level=info msg="Usage stats are ready to report" 09:44:48 =================================== 09:44:48 ======== Logs from kafka ======== 09:44:48 kafka | ===> User 09:44:48 kafka | uid=1000(appuser) gid=1000(appuser) groups=1000(appuser) 09:44:48 kafka | ===> Configuring ... 09:44:48 kafka | Running in Zookeeper mode... 09:44:48 kafka | ===> Running preflight checks ... 09:44:48 kafka | ===> Check if /var/lib/kafka/data is writable ... 09:44:48 kafka | ===> Check if Zookeeper is healthy ... 09:44:48 kafka | [2024-07-04 09:41:59,466] INFO Client environment:zookeeper.version=3.8.3-6ad6d364c7c0bcf0de452d54ebefa3058098ab56, built on 2023-10-05 10:34 UTC (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:host.name=kafka (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:java.version=11.0.22 (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:java.vendor=Azul Systems, Inc. (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:java.home=/usr/lib/jvm/java-11-zulu-openjdk-ca (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:java.class.path=/usr/share/java/cp-base-new/scala-logging_2.13-3.9.4.jar:/usr/share/java/cp-base-new/jackson-datatype-jdk8-2.14.2.jar:/usr/share/java/cp-base-new/kafka-storage-api-7.6.1-ccs.jar:/usr/share/java/cp-base-new/commons-validator-1.7.jar:/usr/share/java/cp-base-new/logredactor-1.0.12.jar:/usr/share/java/cp-base-new/jolokia-core-1.7.1.jar:/usr/share/java/cp-base-new/re2j-1.6.jar:/usr/share/java/cp-base-new/scala-library-2.13.10.jar:/usr/share/java/cp-base-new/commons-io-2.11.0.jar:/usr/share/java/cp-base-new/kafka-clients-7.6.1-ccs.jar:/usr/share/java/cp-base-new/utility-belt-7.6.1.jar:/usr/share/java/cp-base-new/commons-cli-1.4.jar:/usr/share/java/cp-base-new/slf4j-reload4j-1.7.36.jar:/usr/share/java/cp-base-new/kafka-server-common-7.6.1-ccs.jar:/usr/share/java/cp-base-new/kafka-group-coordinator-7.6.1-ccs.jar:/usr/share/java/cp-base-new/common-utils-7.6.1.jar:/usr/share/java/cp-base-new/jackson-annotations-2.14.2.jar:/usr/share/java/cp-base-new/json-simple-1.1.1.jar:/usr/share/java/cp-base-new/jackson-module-scala_2.13-2.14.2.jar:/usr/share/java/cp-base-new/scala-java8-compat_2.13-1.0.2.jar:/usr/share/java/cp-base-new/minimal-json-0.9.5.jar:/usr/share/java/cp-base-new/kafka-storage-7.6.1-ccs.jar:/usr/share/java/cp-base-new/zstd-jni-1.5.5-1.jar:/usr/share/java/cp-base-new/jackson-dataformat-yaml-2.14.2.jar:/usr/share/java/cp-base-new/slf4j-api-1.7.36.jar:/usr/share/java/cp-base-new/paranamer-2.8.jar:/usr/share/java/cp-base-new/commons-beanutils-1.9.4.jar:/usr/share/java/cp-base-new/jmx_prometheus_javaagent-0.18.0.jar:/usr/share/java/cp-base-new/reload4j-1.2.25.jar:/usr/share/java/cp-base-new/jackson-core-2.14.2.jar:/usr/share/java/cp-base-new/disk-usage-agent-7.6.1.jar:/usr/share/java/cp-base-new/commons-digester-2.1.jar:/usr/share/java/cp-base-new/argparse4j-0.7.0.jar:/usr/share/java/cp-base-new/audience-annotations-0.12.0.jar:/usr/share/java/cp-base-new/kafka-raft-7.6.1-ccs.jar:/usr/share/java/cp-base-new/gson-2.9.0.jar:/usr/share/java/cp-base-new/snakeyaml-2.0.jar:/usr/share/java/cp-base-new/jopt-simple-5.0.4.jar:/usr/share/java/cp-base-new/error_prone_annotations-2.10.0.jar:/usr/share/java/cp-base-new/lz4-java-1.8.0.jar:/usr/share/java/cp-base-new/logredactor-metrics-1.0.12.jar:/usr/share/java/cp-base-new/checker-qual-3.19.0.jar:/usr/share/java/cp-base-new/kafka-metadata-7.6.1-ccs.jar:/usr/share/java/cp-base-new/pcollections-4.0.1.jar:/usr/share/java/cp-base-new/commons-logging-1.2.jar:/usr/share/java/cp-base-new/scala-reflect-2.13.10.jar:/usr/share/java/cp-base-new/scala-collection-compat_2.13-2.10.0.jar:/usr/share/java/cp-base-new/metrics-core-2.2.0.jar:/usr/share/java/cp-base-new/zookeeper-jute-3.8.3.jar:/usr/share/java/cp-base-new/jackson-dataformat-csv-2.14.2.jar:/usr/share/java/cp-base-new/commons-collections-3.2.2.jar:/usr/share/java/cp-base-new/caffeine-2.9.3.jar:/usr/share/java/cp-base-new/jolokia-jvm-1.7.1.jar:/usr/share/java/cp-base-new/kafka-tools-api-7.6.1-ccs.jar:/usr/share/java/cp-base-new/zookeeper-3.8.3.jar:/usr/share/java/cp-base-new/metrics-core-4.1.12.1.jar:/usr/share/java/cp-base-new/kafka_2.13-7.6.1-ccs.jar:/usr/share/java/cp-base-new/jackson-databind-2.14.2.jar:/usr/share/java/cp-base-new/snappy-java-1.1.10.5.jar:/usr/share/java/cp-base-new/jose4j-0.9.5.jar (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:os.version=4.15.0-192-generic (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:user.name=appuser (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:user.home=/home/appuser (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:user.dir=/home/appuser (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:os.memory.free=493MB (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:os.memory.max=8042MB (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,467] INFO Client environment:os.memory.total=504MB (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,470] INFO Initiating client connection, connectString=zookeeper:2181 sessionTimeout=40000 watcher=io.confluent.admin.utils.ZookeeperConnectionWatcher@b7f23d9 (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,473] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) 09:44:48 kafka | [2024-07-04 09:41:59,477] INFO jute.maxbuffer value is 1048575 Bytes (org.apache.zookeeper.ClientCnxnSocket) 09:44:48 kafka | [2024-07-04 09:41:59,483] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | [2024-07-04 09:41:59,502] INFO Opening socket connection to server zookeeper/172.17.0.2:2181. (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | [2024-07-04 09:41:59,502] INFO SASL config status: Will not attempt to authenticate using SASL (unknown error) (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | [2024-07-04 09:41:59,509] INFO Socket connection established, initiating session, client: /172.17.0.6:44036, server: zookeeper/172.17.0.2:2181 (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | [2024-07-04 09:41:59,610] INFO Session establishment complete on server zookeeper/172.17.0.2:2181, session id = 0x10000032f9c0000, negotiated timeout = 40000 (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | [2024-07-04 09:41:59,731] INFO Session: 0x10000032f9c0000 closed (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:41:59,732] INFO EventThread shut down for session: 0x10000032f9c0000 (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | Using log4j config /etc/kafka/log4j.properties 09:44:48 kafka | ===> Launching ... 09:44:48 kafka | ===> Launching kafka ... 09:44:48 kafka | [2024-07-04 09:42:00,505] INFO Registered kafka:type=kafka.Log4jController MBean (kafka.utils.Log4jControllerRegistration$) 09:44:48 kafka | [2024-07-04 09:42:00,869] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) 09:44:48 kafka | [2024-07-04 09:42:00,947] INFO Registered signal handlers for TERM, INT, HUP (org.apache.kafka.common.utils.LoggingSignalHandler) 09:44:48 kafka | [2024-07-04 09:42:00,948] INFO starting (kafka.server.KafkaServer) 09:44:48 kafka | [2024-07-04 09:42:00,949] INFO Connecting to zookeeper on zookeeper:2181 (kafka.server.KafkaServer) 09:44:48 kafka | [2024-07-04 09:42:00,962] INFO [ZooKeeperClient Kafka server] Initializing a new session to zookeeper:2181. (kafka.zookeeper.ZooKeeperClient) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:zookeeper.version=3.8.4-9316c2a7a97e1666d8f4593f34dd6fc36ecc436c, built on 2024-02-12 22:16 UTC (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:host.name=kafka (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:java.version=11.0.22 (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:java.vendor=Azul Systems, Inc. (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:java.home=/usr/lib/jvm/java-11-zulu-openjdk-ca (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:java.class.path=/usr/bin/../share/java/kafka/scala-logging_2.13-3.9.4.jar:/usr/bin/../share/java/kafka/jersey-common-2.39.1.jar:/usr/bin/../share/java/kafka/swagger-annotations-2.2.8.jar:/usr/bin/../share/java/kafka/connect-json-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-storage-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/commons-validator-1.7.jar:/usr/bin/../share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/bin/../share/java/kafka/aopalliance-repackaged-2.6.1.jar:/usr/bin/../share/java/kafka/netty-transport-4.1.100.Final.jar:/usr/bin/../share/java/kafka/rocksdbjni-7.9.2.jar:/usr/bin/../share/java/kafka/jetty-http-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka-log4j-appender-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/commons-io-2.11.0.jar:/usr/bin/../share/java/kafka/kafka-clients-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/javax.activation-api-1.2.0.jar:/usr/bin/../share/java/kafka/jetty-security-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/commons-cli-1.4.jar:/usr/bin/../share/java/kafka/jetty-server-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/slf4j-reload4j-1.7.36.jar:/usr/bin/../share/java/kafka/scala-reflect-2.13.11.jar:/usr/bin/../share/java/kafka/kafka-server-common-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-group-coordinator-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-buffer-4.1.100.Final.jar:/usr/bin/../share/java/kafka/connect-runtime-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/trogdor-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.ws.rs-api-2.1.6.jar:/usr/bin/../share/java/kafka/jakarta.annotation-api-1.3.5.jar:/usr/bin/../share/java/kafka/kafka-streams-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/scala-java8-compat_2.13-1.0.2.jar:/usr/bin/../share/java/kafka/javax.ws.rs-api-2.1.1.jar:/usr/bin/../share/java/kafka/zookeeper-jute-3.8.4.jar:/usr/bin/../share/java/kafka/netty-resolver-4.1.100.Final.jar:/usr/bin/../share/java/kafka/hk2-api-2.6.1.jar:/usr/bin/../share/java/kafka/kafka-storage-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/zstd-jni-1.5.5-1.jar:/usr/bin/../share/java/kafka/jackson-dataformat-csv-2.13.5.jar:/usr/bin/../share/java/kafka/kafka.jar:/usr/bin/../share/java/kafka/scala-library-2.13.11.jar:/usr/bin/../share/java/kafka/jakarta.inject-2.6.1.jar:/usr/bin/../share/java/kafka/jakarta.xml.bind-api-2.3.3.jar:/usr/bin/../share/java/kafka/jetty-continuation-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/connect-transforms-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jose4j-0.9.4.jar:/usr/bin/../share/java/kafka/hk2-locator-2.6.1.jar:/usr/bin/../share/java/kafka/reflections-0.10.2.jar:/usr/bin/../share/java/kafka/slf4j-api-1.7.36.jar:/usr/bin/../share/java/kafka/paranamer-2.8.jar:/usr/bin/../share/java/kafka/commons-beanutils-1.9.4.jar:/usr/bin/../share/java/kafka/jaxb-api-2.3.1.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-2.39.1.jar:/usr/bin/../share/java/kafka/hk2-utils-2.6.1.jar:/usr/bin/../share/java/kafka/jackson-module-scala_2.13-2.13.5.jar:/usr/bin/../share/java/kafka/reload4j-1.2.25.jar:/usr/bin/../share/java/kafka/connect-mirror-client-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-servlet-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jackson-core-2.13.5.jar:/usr/bin/../share/java/kafka/jersey-hk2-2.39.1.jar:/usr/bin/../share/java/kafka/jackson-databind-2.13.5.jar:/usr/bin/../share/java/kafka/jetty-servlets-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jersey-client-2.39.1.jar:/usr/bin/../share/java/kafka/osgi-resource-locator-1.0.3.jar:/usr/bin/../share/java/kafka/commons-digester-2.1.jar:/usr/bin/../share/java/kafka/netty-transport-native-epoll-4.1.100.Final.jar:/usr/bin/../share/java/kafka/argparse4j-0.7.0.jar:/usr/bin/../share/java/kafka/connect-mirror-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-datatype-jdk8-2.13.5.jar:/usr/bin/../share/java/kafka/audience-annotations-0.12.0.jar:/usr/bin/../share/java/kafka/jackson-module-jaxb-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/kafka-raft-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/javax.annotation-api-1.3.2.jar:/usr/bin/../share/java/kafka/maven-artifact-3.8.8.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-json-provider-2.13.5.jar:/usr/bin/../share/java/kafka/jakarta.validation-api-2.0.2.jar:/usr/bin/../share/java/kafka/zookeeper-3.8.4.jar:/usr/bin/../share/java/kafka/jersey-server-2.39.1.jar:/usr/bin/../share/java/kafka/commons-lang3-3.8.1.jar:/usr/bin/../share/java/kafka/jopt-simple-5.0.4.jar:/usr/bin/../share/java/kafka/error_prone_annotations-2.10.0.jar:/usr/bin/../share/java/kafka/lz4-java-1.8.0.jar:/usr/bin/../share/java/kafka/jetty-util-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka-tools-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.activation-api-1.2.2.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-core-2.39.1.jar:/usr/bin/../share/java/kafka/checker-qual-3.19.0.jar:/usr/bin/../share/java/kafka/kafka-metadata-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/pcollections-4.0.1.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-base-2.13.5.jar:/usr/bin/../share/java/kafka/commons-logging-1.2.jar:/usr/bin/../share/java/kafka/jsr305-3.0.2.jar:/usr/bin/../share/java/kafka/jetty-io-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/netty-codec-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-transport-native-unix-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/scala-collection-compat_2.13-2.10.0.jar:/usr/bin/../share/java/kafka/metrics-core-2.2.0.jar:/usr/bin/../share/java/kafka/kafka-streams-test-utils-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-streams-examples-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-handler-4.1.100.Final.jar:/usr/bin/../share/java/kafka/commons-collections-3.2.2.jar:/usr/bin/../share/java/kafka/javassist-3.29.2-GA.jar:/usr/bin/../share/java/kafka/caffeine-2.9.3.jar:/usr/bin/../share/java/kafka/plexus-utils-3.3.1.jar:/usr/bin/../share/java/kafka/kafka-tools-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/activation-1.1.1.jar:/usr/bin/../share/java/kafka/netty-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/kafka-streams-scala_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/metrics-core-4.1.12.1.jar:/usr/bin/../share/java/kafka/jline-3.25.1.jar:/usr/bin/../share/java/kafka/jetty-client-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-basic-auth-extension-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-util-ajax-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/snappy-java-1.1.10.5.jar:/usr/bin/../share/java/kafka/kafka-shell-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-transport-classes-epoll-4.1.100.Final.jar:/usr/bin/../share/java/confluent-telemetry/* (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:os.version=4.15.0-192-generic (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:user.name=appuser (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:user.home=/home/appuser (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:user.dir=/home/appuser (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:os.memory.free=1008MB (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,966] INFO Client environment:os.memory.max=1024MB (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,967] INFO Client environment:os.memory.total=1024MB (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,968] INFO Initiating client connection, connectString=zookeeper:2181 sessionTimeout=18000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@447a020 (org.apache.zookeeper.ZooKeeper) 09:44:48 kafka | [2024-07-04 09:42:00,972] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket) 09:44:48 kafka | [2024-07-04 09:42:00,979] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | [2024-07-04 09:42:00,980] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient) 09:44:48 kafka | [2024-07-04 09:42:00,985] INFO Opening socket connection to server zookeeper/172.17.0.2:2181. (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | [2024-07-04 09:42:00,991] INFO Socket connection established, initiating session, client: /172.17.0.6:44038, server: zookeeper/172.17.0.2:2181 (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | [2024-07-04 09:42:01,002] INFO Session establishment complete on server zookeeper/172.17.0.2:2181, session id = 0x10000032f9c0001, negotiated timeout = 18000 (org.apache.zookeeper.ClientCnxn) 09:44:48 kafka | [2024-07-04 09:42:01,013] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient) 09:44:48 kafka | [2024-07-04 09:42:01,390] INFO Cluster ID = 5fWDTH-mSmOqNb5HKZw_Tw (kafka.server.KafkaServer) 09:44:48 kafka | [2024-07-04 09:42:01,394] WARN No meta.properties file under dir /var/lib/kafka/data/meta.properties (kafka.server.BrokerMetadataCheckpoint) 09:44:48 kafka | [2024-07-04 09:42:01,451] INFO KafkaConfig values: 09:44:48 kafka | advertised.listeners = PLAINTEXT://kafka:9092,PLAINTEXT_HOST://localhost:29092 09:44:48 kafka | alter.config.policy.class.name = null 09:44:48 kafka | alter.log.dirs.replication.quota.window.num = 11 09:44:48 kafka | alter.log.dirs.replication.quota.window.size.seconds = 1 09:44:48 kafka | authorizer.class.name = 09:44:48 kafka | auto.create.topics.enable = true 09:44:48 kafka | auto.include.jmx.reporter = true 09:44:48 kafka | auto.leader.rebalance.enable = true 09:44:48 kafka | background.threads = 10 09:44:48 kafka | broker.heartbeat.interval.ms = 2000 09:44:48 kafka | broker.id = 1 09:44:48 kafka | broker.id.generation.enable = true 09:44:48 kafka | broker.rack = null 09:44:48 kafka | broker.session.timeout.ms = 9000 09:44:48 kafka | client.quota.callback.class = null 09:44:48 kafka | compression.type = producer 09:44:48 kafka | connection.failed.authentication.delay.ms = 100 09:44:48 kafka | connections.max.idle.ms = 600000 09:44:48 kafka | connections.max.reauth.ms = 0 09:44:48 kafka | control.plane.listener.name = null 09:44:48 kafka | controlled.shutdown.enable = true 09:44:48 kafka | controlled.shutdown.max.retries = 3 09:44:48 kafka | controlled.shutdown.retry.backoff.ms = 5000 09:44:48 kafka | controller.listener.names = null 09:44:48 kafka | controller.quorum.append.linger.ms = 25 09:44:48 kafka | controller.quorum.election.backoff.max.ms = 1000 09:44:48 kafka | controller.quorum.election.timeout.ms = 1000 09:44:48 kafka | controller.quorum.fetch.timeout.ms = 2000 09:44:48 kafka | controller.quorum.request.timeout.ms = 2000 09:44:48 kafka | controller.quorum.retry.backoff.ms = 20 09:44:48 kafka | controller.quorum.voters = [] 09:44:48 kafka | controller.quota.window.num = 11 09:44:48 kafka | controller.quota.window.size.seconds = 1 09:44:48 kafka | controller.socket.timeout.ms = 30000 09:44:48 kafka | create.topic.policy.class.name = null 09:44:48 kafka | default.replication.factor = 1 09:44:48 kafka | delegation.token.expiry.check.interval.ms = 3600000 09:44:48 kafka | delegation.token.expiry.time.ms = 86400000 09:44:48 kafka | delegation.token.master.key = null 09:44:48 kafka | delegation.token.max.lifetime.ms = 604800000 09:44:48 kafka | delegation.token.secret.key = null 09:44:48 kafka | delete.records.purgatory.purge.interval.requests = 1 09:44:48 kafka | delete.topic.enable = true 09:44:48 kafka | early.start.listeners = null 09:44:48 kafka | fetch.max.bytes = 57671680 09:44:48 kafka | fetch.purgatory.purge.interval.requests = 1000 09:44:48 kafka | group.consumer.assignors = [org.apache.kafka.coordinator.group.assignor.RangeAssignor] 09:44:48 kafka | group.consumer.heartbeat.interval.ms = 5000 09:44:48 kafka | group.consumer.max.heartbeat.interval.ms = 15000 09:44:48 kafka | group.consumer.max.session.timeout.ms = 60000 09:44:48 kafka | group.consumer.max.size = 2147483647 09:44:48 kafka | group.consumer.min.heartbeat.interval.ms = 5000 09:44:48 kafka | group.consumer.min.session.timeout.ms = 45000 09:44:48 kafka | group.consumer.session.timeout.ms = 45000 09:44:48 kafka | group.coordinator.new.enable = false 09:44:48 kafka | group.coordinator.threads = 1 09:44:48 kafka | group.initial.rebalance.delay.ms = 3000 09:44:48 kafka | group.max.session.timeout.ms = 1800000 09:44:48 kafka | group.max.size = 2147483647 09:44:48 kafka | group.min.session.timeout.ms = 6000 09:44:48 kafka | initial.broker.registration.timeout.ms = 60000 09:44:48 kafka | inter.broker.listener.name = PLAINTEXT 09:44:48 kafka | inter.broker.protocol.version = 3.6-IV2 09:44:48 kafka | kafka.metrics.polling.interval.secs = 10 09:44:48 kafka | kafka.metrics.reporters = [] 09:44:48 kafka | leader.imbalance.check.interval.seconds = 300 09:44:48 kafka | leader.imbalance.per.broker.percentage = 10 09:44:48 kafka | listener.security.protocol.map = PLAINTEXT:PLAINTEXT,PLAINTEXT_HOST:PLAINTEXT 09:44:48 kafka | listeners = PLAINTEXT://0.0.0.0:9092,PLAINTEXT_HOST://0.0.0.0:29092 09:44:48 kafka | log.cleaner.backoff.ms = 15000 09:44:48 kafka | log.cleaner.dedupe.buffer.size = 134217728 09:44:48 kafka | log.cleaner.delete.retention.ms = 86400000 09:44:48 kafka | log.cleaner.enable = true 09:44:48 kafka | log.cleaner.io.buffer.load.factor = 0.9 09:44:48 kafka | log.cleaner.io.buffer.size = 524288 09:44:48 kafka | log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308 09:44:48 kafka | log.cleaner.max.compaction.lag.ms = 9223372036854775807 09:44:48 kafka | log.cleaner.min.cleanable.ratio = 0.5 09:44:48 kafka | log.cleaner.min.compaction.lag.ms = 0 09:44:48 kafka | log.cleaner.threads = 1 09:44:48 kafka | log.cleanup.policy = [delete] 09:44:48 kafka | log.dir = /tmp/kafka-logs 09:44:48 kafka | log.dirs = /var/lib/kafka/data 09:44:48 kafka | log.flush.interval.messages = 9223372036854775807 09:44:48 kafka | log.flush.interval.ms = null 09:44:48 kafka | log.flush.offset.checkpoint.interval.ms = 60000 09:44:48 kafka | log.flush.scheduler.interval.ms = 9223372036854775807 09:44:48 kafka | log.flush.start.offset.checkpoint.interval.ms = 60000 09:44:48 kafka | log.index.interval.bytes = 4096 09:44:48 kafka | log.index.size.max.bytes = 10485760 09:44:48 kafka | log.local.retention.bytes = -2 09:44:48 kafka | log.local.retention.ms = -2 09:44:48 kafka | log.message.downconversion.enable = true 09:44:48 kafka | log.message.format.version = 3.0-IV1 09:44:48 kafka | log.message.timestamp.after.max.ms = 9223372036854775807 09:44:48 kafka | log.message.timestamp.before.max.ms = 9223372036854775807 09:44:48 kafka | log.message.timestamp.difference.max.ms = 9223372036854775807 09:44:48 kafka | log.message.timestamp.type = CreateTime 09:44:48 kafka | log.preallocate = false 09:44:48 kafka | log.retention.bytes = -1 09:44:48 kafka | log.retention.check.interval.ms = 300000 09:44:48 kafka | log.retention.hours = 168 09:44:48 kafka | log.retention.minutes = null 09:44:48 kafka | log.retention.ms = null 09:44:48 kafka | log.roll.hours = 168 09:44:48 kafka | log.roll.jitter.hours = 0 09:44:48 kafka | log.roll.jitter.ms = null 09:44:48 kafka | log.roll.ms = null 09:44:48 kafka | log.segment.bytes = 1073741824 09:44:48 kafka | log.segment.delete.delay.ms = 60000 09:44:48 kafka | max.connection.creation.rate = 2147483647 09:44:48 kafka | max.connections = 2147483647 09:44:48 kafka | max.connections.per.ip = 2147483647 09:44:48 kafka | max.connections.per.ip.overrides = 09:44:48 kafka | max.incremental.fetch.session.cache.slots = 1000 09:44:48 kafka | message.max.bytes = 1048588 09:44:48 kafka | metadata.log.dir = null 09:44:48 kafka | metadata.log.max.record.bytes.between.snapshots = 20971520 09:44:48 kafka | metadata.log.max.snapshot.interval.ms = 3600000 09:44:48 kafka | metadata.log.segment.bytes = 1073741824 09:44:48 kafka | metadata.log.segment.min.bytes = 8388608 09:44:48 kafka | metadata.log.segment.ms = 604800000 09:44:48 kafka | metadata.max.idle.interval.ms = 500 09:44:48 kafka | metadata.max.retention.bytes = 104857600 09:44:48 kafka | metadata.max.retention.ms = 604800000 09:44:48 kafka | metric.reporters = [] 09:44:48 kafka | metrics.num.samples = 2 09:44:48 kafka | metrics.recording.level = INFO 09:44:48 kafka | metrics.sample.window.ms = 30000 09:44:48 kafka | min.insync.replicas = 1 09:44:48 kafka | node.id = 1 09:44:48 kafka | num.io.threads = 8 09:44:48 kafka | num.network.threads = 3 09:44:48 kafka | num.partitions = 1 09:44:48 kafka | num.recovery.threads.per.data.dir = 1 09:44:48 kafka | num.replica.alter.log.dirs.threads = null 09:44:48 kafka | num.replica.fetchers = 1 09:44:48 kafka | offset.metadata.max.bytes = 4096 09:44:48 kafka | offsets.commit.required.acks = -1 09:44:48 kafka | offsets.commit.timeout.ms = 5000 09:44:48 kafka | offsets.load.buffer.size = 5242880 09:44:48 kafka | offsets.retention.check.interval.ms = 600000 09:44:48 kafka | offsets.retention.minutes = 10080 09:44:48 kafka | offsets.topic.compression.codec = 0 09:44:48 kafka | offsets.topic.num.partitions = 50 09:44:48 kafka | offsets.topic.replication.factor = 1 09:44:48 kafka | offsets.topic.segment.bytes = 104857600 09:44:48 kafka | password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding 09:44:48 kafka | password.encoder.iterations = 4096 09:44:48 kafka | password.encoder.key.length = 128 09:44:48 kafka | password.encoder.keyfactory.algorithm = null 09:44:48 kafka | password.encoder.old.secret = null 09:44:48 kafka | password.encoder.secret = null 09:44:48 kafka | principal.builder.class = class org.apache.kafka.common.security.authenticator.DefaultKafkaPrincipalBuilder 09:44:48 kafka | process.roles = [] 09:44:48 kafka | producer.id.expiration.check.interval.ms = 600000 09:44:48 kafka | producer.id.expiration.ms = 86400000 09:44:48 kafka | producer.purgatory.purge.interval.requests = 1000 09:44:48 kafka | queued.max.request.bytes = -1 09:44:48 kafka | queued.max.requests = 500 09:44:48 kafka | quota.window.num = 11 09:44:48 kafka | quota.window.size.seconds = 1 09:44:48 kafka | remote.log.index.file.cache.total.size.bytes = 1073741824 09:44:48 kafka | remote.log.manager.task.interval.ms = 30000 09:44:48 kafka | remote.log.manager.task.retry.backoff.max.ms = 30000 09:44:48 kafka | remote.log.manager.task.retry.backoff.ms = 500 09:44:48 kafka | remote.log.manager.task.retry.jitter = 0.2 09:44:48 kafka | remote.log.manager.thread.pool.size = 10 09:44:48 kafka | remote.log.metadata.custom.metadata.max.bytes = 128 09:44:48 kafka | remote.log.metadata.manager.class.name = org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemoteLogMetadataManager 09:44:48 kafka | remote.log.metadata.manager.class.path = null 09:44:48 kafka | remote.log.metadata.manager.impl.prefix = rlmm.config. 09:44:48 kafka | remote.log.metadata.manager.listener.name = null 09:44:48 kafka | remote.log.reader.max.pending.tasks = 100 09:44:48 kafka | remote.log.reader.threads = 10 09:44:48 kafka | remote.log.storage.manager.class.name = null 09:44:48 kafka | remote.log.storage.manager.class.path = null 09:44:48 kafka | remote.log.storage.manager.impl.prefix = rsm.config. 09:44:48 kafka | remote.log.storage.system.enable = false 09:44:48 kafka | replica.fetch.backoff.ms = 1000 09:44:48 kafka | replica.fetch.max.bytes = 1048576 09:44:48 kafka | replica.fetch.min.bytes = 1 09:44:48 kafka | replica.fetch.response.max.bytes = 10485760 09:44:48 kafka | replica.fetch.wait.max.ms = 500 09:44:48 kafka | replica.high.watermark.checkpoint.interval.ms = 5000 09:44:48 kafka | replica.lag.time.max.ms = 30000 09:44:48 kafka | replica.selector.class = null 09:44:48 kafka | replica.socket.receive.buffer.bytes = 65536 09:44:48 kafka | replica.socket.timeout.ms = 30000 09:44:48 kafka | replication.quota.window.num = 11 09:44:48 kafka | replication.quota.window.size.seconds = 1 09:44:48 kafka | request.timeout.ms = 30000 09:44:48 kafka | reserved.broker.max.id = 1000 09:44:48 kafka | sasl.client.callback.handler.class = null 09:44:48 kafka | sasl.enabled.mechanisms = [GSSAPI] 09:44:48 kafka | sasl.jaas.config = null 09:44:48 kafka | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 kafka | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 kafka | sasl.kerberos.principal.to.local.rules = [DEFAULT] 09:44:48 kafka | sasl.kerberos.service.name = null 09:44:48 kafka | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 kafka | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 kafka | sasl.login.callback.handler.class = null 09:44:48 kafka | sasl.login.class = null 09:44:48 kafka | sasl.login.connect.timeout.ms = null 09:44:48 kafka | sasl.login.read.timeout.ms = null 09:44:48 kafka | sasl.login.refresh.buffer.seconds = 300 09:44:48 kafka | sasl.login.refresh.min.period.seconds = 60 09:44:48 kafka | sasl.login.refresh.window.factor = 0.8 09:44:48 kafka | sasl.login.refresh.window.jitter = 0.05 09:44:48 kafka | sasl.login.retry.backoff.max.ms = 10000 09:44:48 kafka | sasl.login.retry.backoff.ms = 100 09:44:48 kafka | sasl.mechanism.controller.protocol = GSSAPI 09:44:48 kafka | sasl.mechanism.inter.broker.protocol = GSSAPI 09:44:48 kafka | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 kafka | sasl.oauthbearer.expected.audience = null 09:44:48 kafka | sasl.oauthbearer.expected.issuer = null 09:44:48 kafka | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 kafka | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 kafka | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 kafka | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 kafka | sasl.oauthbearer.scope.claim.name = scope 09:44:48 kafka | sasl.oauthbearer.sub.claim.name = sub 09:44:48 kafka | sasl.oauthbearer.token.endpoint.url = null 09:44:48 kafka | sasl.server.callback.handler.class = null 09:44:48 kafka | sasl.server.max.receive.size = 524288 09:44:48 kafka | security.inter.broker.protocol = PLAINTEXT 09:44:48 kafka | security.providers = null 09:44:48 kafka | server.max.startup.time.ms = 9223372036854775807 09:44:48 kafka | socket.connection.setup.timeout.max.ms = 30000 09:44:48 kafka | socket.connection.setup.timeout.ms = 10000 09:44:48 kafka | socket.listen.backlog.size = 50 09:44:48 kafka | socket.receive.buffer.bytes = 102400 09:44:48 kafka | socket.request.max.bytes = 104857600 09:44:48 kafka | socket.send.buffer.bytes = 102400 09:44:48 kafka | ssl.cipher.suites = [] 09:44:48 kafka | ssl.client.auth = none 09:44:48 kafka | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 kafka | ssl.endpoint.identification.algorithm = https 09:44:48 kafka | ssl.engine.factory.class = null 09:44:48 kafka | ssl.key.password = null 09:44:48 kafka | ssl.keymanager.algorithm = SunX509 09:44:48 kafka | ssl.keystore.certificate.chain = null 09:44:48 kafka | ssl.keystore.key = null 09:44:48 kafka | ssl.keystore.location = null 09:44:48 kafka | ssl.keystore.password = null 09:44:48 kafka | ssl.keystore.type = JKS 09:44:48 kafka | ssl.principal.mapping.rules = DEFAULT 09:44:48 kafka | ssl.protocol = TLSv1.3 09:44:48 kafka | ssl.provider = null 09:44:48 kafka | ssl.secure.random.implementation = null 09:44:48 kafka | ssl.trustmanager.algorithm = PKIX 09:44:48 kafka | ssl.truststore.certificates = null 09:44:48 kafka | ssl.truststore.location = null 09:44:48 kafka | ssl.truststore.password = null 09:44:48 kafka | ssl.truststore.type = JKS 09:44:48 kafka | transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000 09:44:48 kafka | transaction.max.timeout.ms = 900000 09:44:48 kafka | transaction.partition.verification.enable = true 09:44:48 kafka | transaction.remove.expired.transaction.cleanup.interval.ms = 3600000 09:44:48 kafka | transaction.state.log.load.buffer.size = 5242880 09:44:48 kafka | transaction.state.log.min.isr = 2 09:44:48 kafka | transaction.state.log.num.partitions = 50 09:44:48 kafka | transaction.state.log.replication.factor = 3 09:44:48 kafka | transaction.state.log.segment.bytes = 104857600 09:44:48 kafka | transactional.id.expiration.ms = 604800000 09:44:48 kafka | unclean.leader.election.enable = false 09:44:48 kafka | unstable.api.versions.enable = false 09:44:48 kafka | zookeeper.clientCnxnSocket = null 09:44:48 kafka | zookeeper.connect = zookeeper:2181 09:44:48 kafka | zookeeper.connection.timeout.ms = null 09:44:48 kafka | zookeeper.max.in.flight.requests = 10 09:44:48 kafka | zookeeper.metadata.migration.enable = false 09:44:48 kafka | zookeeper.metadata.migration.min.batch.size = 200 09:44:48 kafka | zookeeper.session.timeout.ms = 18000 09:44:48 kafka | zookeeper.set.acl = false 09:44:48 kafka | zookeeper.ssl.cipher.suites = null 09:44:48 kafka | zookeeper.ssl.client.enable = false 09:44:48 kafka | zookeeper.ssl.crl.enable = false 09:44:48 kafka | zookeeper.ssl.enabled.protocols = null 09:44:48 kafka | zookeeper.ssl.endpoint.identification.algorithm = HTTPS 09:44:48 kafka | zookeeper.ssl.keystore.location = null 09:44:48 kafka | zookeeper.ssl.keystore.password = null 09:44:48 kafka | zookeeper.ssl.keystore.type = null 09:44:48 kafka | zookeeper.ssl.ocsp.enable = false 09:44:48 kafka | zookeeper.ssl.protocol = TLSv1.2 09:44:48 kafka | zookeeper.ssl.truststore.location = null 09:44:48 kafka | zookeeper.ssl.truststore.password = null 09:44:48 kafka | zookeeper.ssl.truststore.type = null 09:44:48 kafka | (kafka.server.KafkaConfig) 09:44:48 kafka | [2024-07-04 09:42:01,492] INFO [ThrottledChannelReaper-Fetch]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) 09:44:48 kafka | [2024-07-04 09:42:01,493] INFO [ThrottledChannelReaper-Produce]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) 09:44:48 kafka | [2024-07-04 09:42:01,494] INFO [ThrottledChannelReaper-Request]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) 09:44:48 kafka | [2024-07-04 09:42:01,496] INFO [ThrottledChannelReaper-ControllerMutation]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) 09:44:48 kafka | [2024-07-04 09:42:01,524] INFO Loading logs from log dirs ArraySeq(/var/lib/kafka/data) (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:01,527] INFO No logs found to be loaded in /var/lib/kafka/data (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:01,537] INFO Loaded 0 logs in 13ms (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:01,539] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:01,540] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:01,564] INFO Starting the log cleaner (kafka.log.LogCleaner) 09:44:48 kafka | [2024-07-04 09:42:01,612] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner$CleanerThread) 09:44:48 kafka | [2024-07-04 09:42:01,627] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) 09:44:48 kafka | [2024-07-04 09:42:01,641] INFO Feature ZK node at path: /feature does not exist (kafka.server.FinalizedFeatureChangeListener) 09:44:48 kafka | [2024-07-04 09:42:01,679] INFO [zk-broker-1-to-controller-forwarding-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) 09:44:48 kafka | [2024-07-04 09:42:02,036] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) 09:44:48 kafka | [2024-07-04 09:42:02,062] INFO [SocketServer listenerType=ZK_BROKER, nodeId=1] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer) 09:44:48 kafka | [2024-07-04 09:42:02,062] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) 09:44:48 kafka | [2024-07-04 09:42:02,068] INFO [SocketServer listenerType=ZK_BROKER, nodeId=1] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT_HOST) (kafka.network.SocketServer) 09:44:48 kafka | [2024-07-04 09:42:02,072] INFO [zk-broker-1-to-controller-alter-partition-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) 09:44:48 kafka | [2024-07-04 09:42:02,095] INFO [ExpirationReaper-1-Produce]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) 09:44:48 kafka | [2024-07-04 09:42:02,100] INFO [ExpirationReaper-1-Fetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) 09:44:48 kafka | [2024-07-04 09:42:02,101] INFO [ExpirationReaper-1-DeleteRecords]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) 09:44:48 kafka | [2024-07-04 09:42:02,102] INFO [ExpirationReaper-1-ElectLeader]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) 09:44:48 kafka | [2024-07-04 09:42:02,102] INFO [ExpirationReaper-1-RemoteFetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) 09:44:48 kafka | [2024-07-04 09:42:02,122] INFO [LogDirFailureHandler]: Starting (kafka.server.ReplicaManager$LogDirFailureHandler) 09:44:48 kafka | [2024-07-04 09:42:02,123] INFO [AddPartitionsToTxnSenderThread-1]: Starting (kafka.server.AddPartitionsToTxnManager) 09:44:48 kafka | [2024-07-04 09:42:02,148] INFO Creating /brokers/ids/1 (is it secure? false) (kafka.zk.KafkaZkClient) 09:44:48 kafka | [2024-07-04 09:42:02,175] INFO Stat of the created znode at /brokers/ids/1 is: 27,27,1720086122163,1720086122163,1,0,0,72057607721582593,258,0,27 09:44:48 kafka | (kafka.zk.KafkaZkClient) 09:44:48 kafka | [2024-07-04 09:42:02,175] INFO Registered broker 1 at path /brokers/ids/1 with addresses: PLAINTEXT://kafka:9092,PLAINTEXT_HOST://localhost:29092, czxid (broker epoch): 27 (kafka.zk.KafkaZkClient) 09:44:48 kafka | [2024-07-04 09:42:02,226] INFO [ControllerEventThread controllerId=1] Starting (kafka.controller.ControllerEventManager$ControllerEventThread) 09:44:48 kafka | [2024-07-04 09:42:02,233] INFO [ExpirationReaper-1-topic]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) 09:44:48 kafka | [2024-07-04 09:42:02,239] INFO [ExpirationReaper-1-Heartbeat]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) 09:44:48 kafka | [2024-07-04 09:42:02,239] INFO [ExpirationReaper-1-Rebalance]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) 09:44:48 kafka | [2024-07-04 09:42:02,252] INFO Successfully created /controller_epoch with initial epoch 0 (kafka.zk.KafkaZkClient) 09:44:48 kafka | [2024-07-04 09:42:02,253] INFO [GroupCoordinator 1]: Starting up. (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:02,261] INFO [Controller id=1] 1 successfully elected as the controller. Epoch incremented to 1 and epoch zk version is now 1 (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,263] INFO [GroupCoordinator 1]: Startup complete. (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:02,265] INFO [Controller id=1] Creating FeatureZNode at path: /feature with contents: FeatureZNode(2,Enabled,Map()) (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,269] INFO Feature ZK node created at path: /feature (kafka.server.FinalizedFeatureChangeListener) 09:44:48 kafka | [2024-07-04 09:42:02,283] INFO [TransactionCoordinator id=1] Starting up. (kafka.coordinator.transaction.TransactionCoordinator) 09:44:48 kafka | [2024-07-04 09:42:02,285] INFO [TransactionCoordinator id=1] Startup complete. (kafka.coordinator.transaction.TransactionCoordinator) 09:44:48 kafka | [2024-07-04 09:42:02,286] INFO [TxnMarkerSenderThread-1]: Starting (kafka.coordinator.transaction.TransactionMarkerChannelManager) 09:44:48 kafka | [2024-07-04 09:42:02,298] INFO [MetadataCache brokerId=1] Updated cache from existing None to latest Features(version=3.6-IV2, finalizedFeatures={}, finalizedFeaturesEpoch=0). (kafka.server.metadata.ZkMetadataCache) 09:44:48 kafka | [2024-07-04 09:42:02,299] INFO [Controller id=1] Registering handlers (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,303] INFO [Controller id=1] Deleting log dir event notifications (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,307] INFO [Controller id=1] Deleting isr change notifications (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,309] INFO [Controller id=1] Initializing controller context (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,321] INFO [Controller id=1] Initialized broker epochs cache: HashMap(1 -> 27) (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,327] DEBUG [Controller id=1] Register BrokerModifications handler for Set(1) (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,327] INFO [ExpirationReaper-1-AlterAcls]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) 09:44:48 kafka | [2024-07-04 09:42:02,333] DEBUG [Channel manager on controller 1]: Controller 1 trying to connect to broker 1 (kafka.controller.ControllerChannelManager) 09:44:48 kafka | [2024-07-04 09:42:02,341] INFO [Controller id=1] Currently active brokers in the cluster: Set(1) (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,341] INFO [RequestSendThread controllerId=1] Starting (kafka.controller.RequestSendThread) 09:44:48 kafka | [2024-07-04 09:42:02,341] INFO [Controller id=1] Currently shutting brokers in the cluster: HashSet() (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,342] INFO [Controller id=1] Current list of topics in the cluster: HashSet() (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,342] INFO [Controller id=1] Fetching topic deletions in progress (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,344] INFO [Controller id=1] List of topics to be deleted: (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,344] INFO [Controller id=1] List of topics ineligible for deletion: (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,345] INFO [Controller id=1] Initializing topic deletion manager (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,345] INFO [Topic Deletion Manager 1] Initializing manager with initial deletions: Set(), initial ineligible deletions: HashSet() (kafka.controller.TopicDeletionManager) 09:44:48 kafka | [2024-07-04 09:42:02,346] INFO [Controller id=1] Sending update metadata request (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,348] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet(1) for 0 partitions (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:02,356] INFO [/config/changes-event-process-thread]: Starting (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) 09:44:48 kafka | [2024-07-04 09:42:02,359] INFO [ReplicaStateMachine controllerId=1] Initializing replica state (kafka.controller.ZkReplicaStateMachine) 09:44:48 kafka | [2024-07-04 09:42:02,360] INFO [ReplicaStateMachine controllerId=1] Triggering online replica state changes (kafka.controller.ZkReplicaStateMachine) 09:44:48 kafka | [2024-07-04 09:42:02,366] INFO [ReplicaStateMachine controllerId=1] Triggering offline replica state changes (kafka.controller.ZkReplicaStateMachine) 09:44:48 kafka | [2024-07-04 09:42:02,367] DEBUG [ReplicaStateMachine controllerId=1] Started replica state machine with initial state -> HashMap() (kafka.controller.ZkReplicaStateMachine) 09:44:48 kafka | [2024-07-04 09:42:02,367] INFO [SocketServer listenerType=ZK_BROKER, nodeId=1] Enabling request processing. (kafka.network.SocketServer) 09:44:48 kafka | [2024-07-04 09:42:02,367] INFO [PartitionStateMachine controllerId=1] Initializing partition state (kafka.controller.ZkPartitionStateMachine) 09:44:48 kafka | [2024-07-04 09:42:02,368] INFO [PartitionStateMachine controllerId=1] Triggering online partition state changes (kafka.controller.ZkPartitionStateMachine) 09:44:48 kafka | [2024-07-04 09:42:02,371] INFO Awaiting socket connections on 0.0.0.0:9092. (kafka.network.DataPlaneAcceptor) 09:44:48 kafka | [2024-07-04 09:42:02,371] DEBUG [PartitionStateMachine controllerId=1] Started partition state machine with initial state -> HashMap() (kafka.controller.ZkPartitionStateMachine) 09:44:48 kafka | [2024-07-04 09:42:02,372] INFO [Controller id=1] Ready to serve as the new controller with epoch 1 (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,373] INFO [Controller id=1, targetBrokerId=1] Node 1 disconnected. (org.apache.kafka.clients.NetworkClient) 09:44:48 kafka | [2024-07-04 09:42:02,376] INFO [Controller id=1] Partitions undergoing preferred replica election: (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,376] WARN [Controller id=1, targetBrokerId=1] Connection to node 1 (kafka/172.17.0.6:9092) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient) 09:44:48 kafka | [2024-07-04 09:42:02,376] INFO [Controller id=1] Partitions that completed preferred replica election: (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,377] INFO [Controller id=1] Skipping preferred replica election for partitions due to topic deletion: (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,377] INFO [Controller id=1] Resuming preferred replica election for partitions: (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,378] INFO Awaiting socket connections on 0.0.0.0:29092. (kafka.network.DataPlaneAcceptor) 09:44:48 kafka | [2024-07-04 09:42:02,378] WARN [RequestSendThread controllerId=1] Controller 1's connection to broker kafka:9092 (id: 1 rack: null) was unsuccessful (kafka.controller.RequestSendThread) 09:44:48 kafka | java.io.IOException: Connection to kafka:9092 (id: 1 rack: null) failed. 09:44:48 kafka | at org.apache.kafka.clients.NetworkClientUtils.awaitReady(NetworkClientUtils.java:70) 09:44:48 kafka | at kafka.controller.RequestSendThread.brokerReady(ControllerChannelManager.scala:298) 09:44:48 kafka | at kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:251) 09:44:48 kafka | at org.apache.kafka.server.util.ShutdownableThread.run(ShutdownableThread.java:130) 09:44:48 kafka | [2024-07-04 09:42:02,378] INFO [Controller id=1] Starting replica leader election (PREFERRED) for partitions triggered by ZkTriggered (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,380] INFO [Controller id=1, targetBrokerId=1] Client requested connection close from node 1 (org.apache.kafka.clients.NetworkClient) 09:44:48 kafka | [2024-07-04 09:42:02,386] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) 09:44:48 kafka | [2024-07-04 09:42:02,386] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) 09:44:48 kafka | [2024-07-04 09:42:02,386] INFO Kafka startTimeMs: 1720086122381 (org.apache.kafka.common.utils.AppInfoParser) 09:44:48 kafka | [2024-07-04 09:42:02,388] INFO [KafkaServer id=1] started (kafka.server.KafkaServer) 09:44:48 kafka | [2024-07-04 09:42:02,391] INFO [Controller id=1] Starting the controller scheduler (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:02,484] INFO [RequestSendThread controllerId=1] Controller 1 connected to kafka:9092 (id: 1 rack: null) for sending state change requests (kafka.controller.RequestSendThread) 09:44:48 kafka | [2024-07-04 09:42:02,549] TRACE [Controller id=1 epoch=1] Received response UpdateMetadataResponseData(errorCode=0) for request UPDATE_METADATA with correlation id 0 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:02,578] INFO [zk-broker-1-to-controller-alter-partition-channel-manager]: Recorded new controller, from now on will use node kafka:9092 (id: 1 rack: null) (kafka.server.BrokerToControllerRequestThread) 09:44:48 kafka | [2024-07-04 09:42:02,869] INFO [zk-broker-1-to-controller-forwarding-channel-manager]: Recorded new controller, from now on will use node kafka:9092 (id: 1 rack: null) (kafka.server.BrokerToControllerRequestThread) 09:44:48 kafka | [2024-07-04 09:42:07,393] INFO [Controller id=1] Processing automatic preferred replica leader election (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:07,393] TRACE [Controller id=1] Checking need to trigger auto leader balancing (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:37,964] INFO Creating topic policy-pdp-pap with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(1)) (kafka.zk.AdminZkClient) 09:44:48 kafka | [2024-07-04 09:42:37,965] DEBUG [Controller id=1] There is no producerId block yet (Zk path version 0), creating the first block (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:37,964] INFO Creating topic __consumer_offsets with configuration {compression.type=producer, cleanup.policy=compact, segment.bytes=104857600} and initial partition assignment HashMap(0 -> ArrayBuffer(1), 1 -> ArrayBuffer(1), 2 -> ArrayBuffer(1), 3 -> ArrayBuffer(1), 4 -> ArrayBuffer(1), 5 -> ArrayBuffer(1), 6 -> ArrayBuffer(1), 7 -> ArrayBuffer(1), 8 -> ArrayBuffer(1), 9 -> ArrayBuffer(1), 10 -> ArrayBuffer(1), 11 -> ArrayBuffer(1), 12 -> ArrayBuffer(1), 13 -> ArrayBuffer(1), 14 -> ArrayBuffer(1), 15 -> ArrayBuffer(1), 16 -> ArrayBuffer(1), 17 -> ArrayBuffer(1), 18 -> ArrayBuffer(1), 19 -> ArrayBuffer(1), 20 -> ArrayBuffer(1), 21 -> ArrayBuffer(1), 22 -> ArrayBuffer(1), 23 -> ArrayBuffer(1), 24 -> ArrayBuffer(1), 25 -> ArrayBuffer(1), 26 -> ArrayBuffer(1), 27 -> ArrayBuffer(1), 28 -> ArrayBuffer(1), 29 -> ArrayBuffer(1), 30 -> ArrayBuffer(1), 31 -> ArrayBuffer(1), 32 -> ArrayBuffer(1), 33 -> ArrayBuffer(1), 34 -> ArrayBuffer(1), 35 -> ArrayBuffer(1), 36 -> ArrayBuffer(1), 37 -> ArrayBuffer(1), 38 -> ArrayBuffer(1), 39 -> ArrayBuffer(1), 40 -> ArrayBuffer(1), 41 -> ArrayBuffer(1), 42 -> ArrayBuffer(1), 43 -> ArrayBuffer(1), 44 -> ArrayBuffer(1), 45 -> ArrayBuffer(1), 46 -> ArrayBuffer(1), 47 -> ArrayBuffer(1), 48 -> ArrayBuffer(1), 49 -> ArrayBuffer(1)) (kafka.zk.AdminZkClient) 09:44:48 kafka | [2024-07-04 09:42:37,977] INFO [Controller id=1] Acquired new producerId block ProducerIdsBlock(assignedBrokerId=1, firstProducerId=0, size=1000) by writing to Zk with path version 1 (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:38,028] INFO [Controller id=1] New topics: [Set(policy-pdp-pap, __consumer_offsets)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(policy-pdp-pap,Some(jb_1oiljSJehU0V5qY0tiA),Map(policy-pdp-pap-0 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=))), TopicIdReplicaAssignment(__consumer_offsets,Some(ZjUdxuMGT2q3AHEVEv_YSw),HashMap(__consumer_offsets-22 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-30 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-25 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-35 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-37 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-38 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-13 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-8 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-21 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-4 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-27 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-7 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-9 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-46 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-41 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-33 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-23 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-49 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-47 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-16 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-28 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-31 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-36 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-42 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-3 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-18 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-15 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-24 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-17 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-48 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-19 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-11 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-2 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-43 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-6 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-14 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-20 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-0 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-44 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-39 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-12 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-45 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-1 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-5 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-26 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-29 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-34 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-10 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-32 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-40 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:38,030] INFO [Controller id=1] New partition creation callback for __consumer_offsets-22,__consumer_offsets-30,__consumer_offsets-25,__consumer_offsets-35,__consumer_offsets-38,__consumer_offsets-13,__consumer_offsets-8,__consumer_offsets-21,__consumer_offsets-4,__consumer_offsets-27,__consumer_offsets-7,__consumer_offsets-9,__consumer_offsets-46,__consumer_offsets-41,__consumer_offsets-33,__consumer_offsets-23,__consumer_offsets-49,__consumer_offsets-47,__consumer_offsets-16,__consumer_offsets-28,__consumer_offsets-31,__consumer_offsets-36,__consumer_offsets-42,__consumer_offsets-3,__consumer_offsets-18,__consumer_offsets-37,policy-pdp-pap-0,__consumer_offsets-15,__consumer_offsets-24,__consumer_offsets-17,__consumer_offsets-48,__consumer_offsets-19,__consumer_offsets-11,__consumer_offsets-2,__consumer_offsets-43,__consumer_offsets-6,__consumer_offsets-14,__consumer_offsets-20,__consumer_offsets-0,__consumer_offsets-44,__consumer_offsets-39,__consumer_offsets-12,__consumer_offsets-45,__consumer_offsets-1,__consumer_offsets-5,__consumer_offsets-26,__consumer_offsets-29,__consumer_offsets-34,__consumer_offsets-10,__consumer_offsets-32,__consumer_offsets-40 (kafka.controller.KafkaController) 09:44:48 kafka | [2024-07-04 09:42:38,033] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-22 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,033] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-30 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,034] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-25 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,034] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-35 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,034] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-38 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,034] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-13 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,034] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-8 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,034] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-21 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,035] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-4 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,035] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-27 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,035] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-7 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,035] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-9 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,035] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-46 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,035] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-41 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,035] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-33 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,036] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-23 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,036] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-49 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,036] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-47 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,036] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-16 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,036] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-28 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,036] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-31 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,037] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-36 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,037] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-42 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,037] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-3 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,037] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-18 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,037] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-37 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,037] INFO [Controller id=1 epoch=1] Changed partition policy-pdp-pap-0 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,037] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-15 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,038] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-24 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,038] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-17 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,038] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-48 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,038] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-19 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,038] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-11 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,038] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-2 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,039] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-43 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,039] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-6 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,039] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-14 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,039] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-20 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,039] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-0 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,039] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-44 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,040] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-39 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,040] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-12 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,040] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-45 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,040] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-1 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,040] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-5 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,041] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-26 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,041] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-29 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,041] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-34 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,041] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-10 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,041] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-32 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,042] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-40 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,042] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,047] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-32 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,048] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-5 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,048] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-44 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,048] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-48 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,048] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-46 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,049] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-20 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,049] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition policy-pdp-pap-0 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,049] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-43 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,049] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-24 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,049] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-6 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,049] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-18 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,050] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-21 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,050] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-1 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,050] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-14 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,050] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-34 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,050] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-16 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,050] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-29 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,051] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-11 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,051] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-0 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,051] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-22 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,051] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-47 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,051] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-36 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,051] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-28 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,052] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-42 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,052] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-9 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,052] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-37 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,052] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-13 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,052] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-30 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,052] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-35 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,053] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-39 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,053] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-12 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,053] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-27 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,053] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-45 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,053] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-19 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,054] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-49 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,054] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-40 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,054] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-41 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,054] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-38 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,054] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-8 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,054] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-7 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,055] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-33 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,055] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-25 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,055] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-31 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,055] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-23 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,055] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-10 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,055] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-2 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,056] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-17 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,056] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-4 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,056] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-15 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,056] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-26 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,056] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-3 from NonExistentReplica to NewReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,057] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-22 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-30 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-25 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-35 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-38 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-13 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-8 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-21 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-4 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-27 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-7 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,290] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-9 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-46 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-41 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-33 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-23 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-49 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-47 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-16 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-28 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-31 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-36 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-42 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-3 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-18 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-37 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition policy-pdp-pap-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-15 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-24 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-17 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-48 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-19 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-11 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-2 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-43 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-6 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-14 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-20 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-44 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-39 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,291] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-12 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,292] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-45 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,292] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,292] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-5 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,292] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-26 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,292] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-29 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,292] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-34 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,292] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-10 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,292] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-32 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,292] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-40 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=13, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-13 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=46, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-46 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=9, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-9 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=42, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-42 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=21, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-21 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=17, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-17 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=30, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-30 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=26, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-26 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=5, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-5 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=38, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-38 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=1, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=34, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-34 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=16, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-16 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=45, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-45 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=12, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-12 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=41, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-41 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=24, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-24 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=20, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-20 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=49, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-49 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,294] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-0 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=29, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-29 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=25, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-25 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=8, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-8 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=37, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-37 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=4, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-4 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=33, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-33 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=15, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-15 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=48, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-48 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=11, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-11 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=44, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-44 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=23, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-23 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=19, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-19 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=32, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-32 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=28, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-28 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=7, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-7 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=40, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-40 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=3, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-3 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=36, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-36 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=47, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-47 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=14, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-14 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=43, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-43 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=10, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-10 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=22, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-22 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=18, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-18 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=31, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-31 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=27, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-27 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=39, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-39 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=6, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-6 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=35, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-35 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='policy-pdp-pap', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition policy-pdp-pap-0 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,295] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=2, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,296] INFO [Controller id=1 epoch=1] Sending LeaderAndIsr request to broker 1 with 51 become-leader and 0 become-follower partitions (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,298] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet(1) for 51 partitions (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-32 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-5 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-44 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-48 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-46 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-20 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition policy-pdp-pap-0 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-43 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-24 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-6 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-18 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-21 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-1 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-14 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-34 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-16 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-29 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-11 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-0 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-22 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-47 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-36 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-28 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-42 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-9 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-37 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,300] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-13 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-30 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-35 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-39 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-12 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-27 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-45 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-19 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-49 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-40 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-41 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-38 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-8 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-7 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-33 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-25 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-31 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-23 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-10 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-2 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-17 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-4 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-15 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-26 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-3 from NewReplica to OnlineReplica (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,301] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,306] INFO [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 for 51 partitions (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,307] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='policy-pdp-pap', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,307] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=13, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,307] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=46, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,307] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=9, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,307] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=42, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,307] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=21, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,307] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=17, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=30, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=26, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=5, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=38, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=1, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=34, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=16, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=45, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=12, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=41, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=24, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=20, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=49, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=29, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=25, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=8, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=37, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=4, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=33, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=15, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=48, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=11, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=44, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=23, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=19, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=32, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=28, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=7, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=40, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=3, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=36, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=47, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=14, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=43, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=10, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,308] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=22, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,309] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=18, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,309] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=31, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,309] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=27, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,309] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=39, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,309] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=6, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,309] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=35, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,309] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=2, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-3 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-18 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-41 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-10 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-33 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-48 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-19 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-34 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-4 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,345] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-11 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-26 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-49 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-39 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-9 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-24 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-31 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-46 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-16 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-25 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-40 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,346] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-47 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-17 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-32 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-37 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-7 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-22 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-29 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-44 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-14 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-23 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-38 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-8 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition policy-pdp-pap-0 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-45 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,347] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-15 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-30 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-0 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-35 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-5 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-20 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-27 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-42 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-12 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-21 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-36 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-6 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-43 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,348] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-13 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,349] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-28 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,350] INFO [ReplicaFetcherManager on broker 1] Removed fetcher for partitions HashSet(__consumer_offsets-22, __consumer_offsets-30, __consumer_offsets-25, __consumer_offsets-35, __consumer_offsets-38, __consumer_offsets-13, __consumer_offsets-8, __consumer_offsets-21, __consumer_offsets-4, __consumer_offsets-27, __consumer_offsets-7, __consumer_offsets-9, __consumer_offsets-46, __consumer_offsets-41, __consumer_offsets-33, __consumer_offsets-23, __consumer_offsets-49, __consumer_offsets-47, __consumer_offsets-16, __consumer_offsets-28, __consumer_offsets-31, __consumer_offsets-36, __consumer_offsets-42, __consumer_offsets-3, __consumer_offsets-18, __consumer_offsets-37, policy-pdp-pap-0, __consumer_offsets-15, __consumer_offsets-24, __consumer_offsets-17, __consumer_offsets-48, __consumer_offsets-19, __consumer_offsets-11, __consumer_offsets-2, __consumer_offsets-43, __consumer_offsets-6, __consumer_offsets-14, __consumer_offsets-20, __consumer_offsets-0, __consumer_offsets-44, __consumer_offsets-39, __consumer_offsets-12, __consumer_offsets-45, __consumer_offsets-1, __consumer_offsets-5, __consumer_offsets-26, __consumer_offsets-29, __consumer_offsets-34, __consumer_offsets-10, __consumer_offsets-32, __consumer_offsets-40) (kafka.server.ReplicaFetcherManager) 09:44:48 kafka | [2024-07-04 09:42:38,350] INFO [Broker id=1] Stopped fetchers as part of LeaderAndIsr request correlationId 1 from controller 1 epoch 1 as part of the become-leader transition for 51 partitions (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,394] INFO [LogLoader partition=__consumer_offsets-3, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,404] INFO Created log for partition __consumer_offsets-3 in /var/lib/kafka/data/__consumer_offsets-3 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,406] INFO [Partition __consumer_offsets-3 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-3 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,407] INFO [Partition __consumer_offsets-3 broker=1] Log loaded for partition __consumer_offsets-3 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,409] INFO [Broker id=1] Leader __consumer_offsets-3 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,420] INFO [LogLoader partition=__consumer_offsets-18, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,420] INFO Created log for partition __consumer_offsets-18 in /var/lib/kafka/data/__consumer_offsets-18 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,420] INFO [Partition __consumer_offsets-18 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-18 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,421] INFO [Partition __consumer_offsets-18 broker=1] Log loaded for partition __consumer_offsets-18 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,421] INFO [Broker id=1] Leader __consumer_offsets-18 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,429] INFO [LogLoader partition=__consumer_offsets-41, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,430] INFO Created log for partition __consumer_offsets-41 in /var/lib/kafka/data/__consumer_offsets-41 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,431] INFO [Partition __consumer_offsets-41 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-41 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,431] INFO [Partition __consumer_offsets-41 broker=1] Log loaded for partition __consumer_offsets-41 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,431] INFO [Broker id=1] Leader __consumer_offsets-41 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,443] INFO [LogLoader partition=__consumer_offsets-10, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,444] INFO Created log for partition __consumer_offsets-10 in /var/lib/kafka/data/__consumer_offsets-10 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,445] INFO [Partition __consumer_offsets-10 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-10 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,445] INFO [Partition __consumer_offsets-10 broker=1] Log loaded for partition __consumer_offsets-10 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,445] INFO [Broker id=1] Leader __consumer_offsets-10 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,456] INFO [LogLoader partition=__consumer_offsets-33, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,459] INFO Created log for partition __consumer_offsets-33 in /var/lib/kafka/data/__consumer_offsets-33 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,461] INFO [Partition __consumer_offsets-33 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-33 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,461] INFO [Partition __consumer_offsets-33 broker=1] Log loaded for partition __consumer_offsets-33 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,464] INFO [Broker id=1] Leader __consumer_offsets-33 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,472] INFO [LogLoader partition=__consumer_offsets-48, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,473] INFO Created log for partition __consumer_offsets-48 in /var/lib/kafka/data/__consumer_offsets-48 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,473] INFO [Partition __consumer_offsets-48 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-48 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,473] INFO [Partition __consumer_offsets-48 broker=1] Log loaded for partition __consumer_offsets-48 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,473] INFO [Broker id=1] Leader __consumer_offsets-48 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,481] INFO [LogLoader partition=__consumer_offsets-19, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,482] INFO Created log for partition __consumer_offsets-19 in /var/lib/kafka/data/__consumer_offsets-19 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,483] INFO [Partition __consumer_offsets-19 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-19 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,483] INFO [Partition __consumer_offsets-19 broker=1] Log loaded for partition __consumer_offsets-19 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,483] INFO [Broker id=1] Leader __consumer_offsets-19 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,491] INFO [LogLoader partition=__consumer_offsets-34, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,492] INFO Created log for partition __consumer_offsets-34 in /var/lib/kafka/data/__consumer_offsets-34 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,492] INFO [Partition __consumer_offsets-34 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-34 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,492] INFO [Partition __consumer_offsets-34 broker=1] Log loaded for partition __consumer_offsets-34 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,492] INFO [Broker id=1] Leader __consumer_offsets-34 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,498] INFO [LogLoader partition=__consumer_offsets-4, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,499] INFO Created log for partition __consumer_offsets-4 in /var/lib/kafka/data/__consumer_offsets-4 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,499] INFO [Partition __consumer_offsets-4 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-4 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,499] INFO [Partition __consumer_offsets-4 broker=1] Log loaded for partition __consumer_offsets-4 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,499] INFO [Broker id=1] Leader __consumer_offsets-4 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,506] INFO [LogLoader partition=__consumer_offsets-11, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,507] INFO Created log for partition __consumer_offsets-11 in /var/lib/kafka/data/__consumer_offsets-11 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,507] INFO [Partition __consumer_offsets-11 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-11 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,507] INFO [Partition __consumer_offsets-11 broker=1] Log loaded for partition __consumer_offsets-11 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,507] INFO [Broker id=1] Leader __consumer_offsets-11 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,514] INFO [LogLoader partition=__consumer_offsets-26, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,515] INFO Created log for partition __consumer_offsets-26 in /var/lib/kafka/data/__consumer_offsets-26 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,515] INFO [Partition __consumer_offsets-26 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-26 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,515] INFO [Partition __consumer_offsets-26 broker=1] Log loaded for partition __consumer_offsets-26 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,515] INFO [Broker id=1] Leader __consumer_offsets-26 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,526] INFO [LogLoader partition=__consumer_offsets-49, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,527] INFO Created log for partition __consumer_offsets-49 in /var/lib/kafka/data/__consumer_offsets-49 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,527] INFO [Partition __consumer_offsets-49 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-49 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,527] INFO [Partition __consumer_offsets-49 broker=1] Log loaded for partition __consumer_offsets-49 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,527] INFO [Broker id=1] Leader __consumer_offsets-49 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,536] INFO [LogLoader partition=__consumer_offsets-39, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,537] INFO Created log for partition __consumer_offsets-39 in /var/lib/kafka/data/__consumer_offsets-39 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,537] INFO [Partition __consumer_offsets-39 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-39 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,537] INFO [Partition __consumer_offsets-39 broker=1] Log loaded for partition __consumer_offsets-39 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,537] INFO [Broker id=1] Leader __consumer_offsets-39 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,544] INFO [LogLoader partition=__consumer_offsets-9, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,545] INFO Created log for partition __consumer_offsets-9 in /var/lib/kafka/data/__consumer_offsets-9 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,545] INFO [Partition __consumer_offsets-9 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-9 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,545] INFO [Partition __consumer_offsets-9 broker=1] Log loaded for partition __consumer_offsets-9 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,545] INFO [Broker id=1] Leader __consumer_offsets-9 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,551] INFO [LogLoader partition=__consumer_offsets-24, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,551] INFO Created log for partition __consumer_offsets-24 in /var/lib/kafka/data/__consumer_offsets-24 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,552] INFO [Partition __consumer_offsets-24 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-24 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,552] INFO [Partition __consumer_offsets-24 broker=1] Log loaded for partition __consumer_offsets-24 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,552] INFO [Broker id=1] Leader __consumer_offsets-24 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,560] INFO [LogLoader partition=__consumer_offsets-31, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,561] INFO Created log for partition __consumer_offsets-31 in /var/lib/kafka/data/__consumer_offsets-31 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,561] INFO [Partition __consumer_offsets-31 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-31 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,561] INFO [Partition __consumer_offsets-31 broker=1] Log loaded for partition __consumer_offsets-31 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,561] INFO [Broker id=1] Leader __consumer_offsets-31 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,635] INFO [LogLoader partition=__consumer_offsets-46, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,636] INFO Created log for partition __consumer_offsets-46 in /var/lib/kafka/data/__consumer_offsets-46 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,636] INFO [Partition __consumer_offsets-46 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-46 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,636] INFO [Partition __consumer_offsets-46 broker=1] Log loaded for partition __consumer_offsets-46 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,637] INFO [Broker id=1] Leader __consumer_offsets-46 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,646] INFO [LogLoader partition=__consumer_offsets-1, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,647] INFO Created log for partition __consumer_offsets-1 in /var/lib/kafka/data/__consumer_offsets-1 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,647] INFO [Partition __consumer_offsets-1 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-1 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,647] INFO [Partition __consumer_offsets-1 broker=1] Log loaded for partition __consumer_offsets-1 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,648] INFO [Broker id=1] Leader __consumer_offsets-1 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,657] INFO [LogLoader partition=__consumer_offsets-16, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,657] INFO Created log for partition __consumer_offsets-16 in /var/lib/kafka/data/__consumer_offsets-16 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,657] INFO [Partition __consumer_offsets-16 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-16 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,657] INFO [Partition __consumer_offsets-16 broker=1] Log loaded for partition __consumer_offsets-16 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,657] INFO [Broker id=1] Leader __consumer_offsets-16 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,673] INFO [LogLoader partition=__consumer_offsets-2, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,674] INFO Created log for partition __consumer_offsets-2 in /var/lib/kafka/data/__consumer_offsets-2 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,674] INFO [Partition __consumer_offsets-2 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-2 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,674] INFO [Partition __consumer_offsets-2 broker=1] Log loaded for partition __consumer_offsets-2 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,674] INFO [Broker id=1] Leader __consumer_offsets-2 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,682] INFO [LogLoader partition=__consumer_offsets-25, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,682] INFO Created log for partition __consumer_offsets-25 in /var/lib/kafka/data/__consumer_offsets-25 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,682] INFO [Partition __consumer_offsets-25 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-25 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,682] INFO [Partition __consumer_offsets-25 broker=1] Log loaded for partition __consumer_offsets-25 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,682] INFO [Broker id=1] Leader __consumer_offsets-25 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,690] INFO [LogLoader partition=__consumer_offsets-40, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,691] INFO Created log for partition __consumer_offsets-40 in /var/lib/kafka/data/__consumer_offsets-40 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,691] INFO [Partition __consumer_offsets-40 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-40 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,691] INFO [Partition __consumer_offsets-40 broker=1] Log loaded for partition __consumer_offsets-40 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,691] INFO [Broker id=1] Leader __consumer_offsets-40 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,697] INFO [LogLoader partition=__consumer_offsets-47, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,697] INFO Created log for partition __consumer_offsets-47 in /var/lib/kafka/data/__consumer_offsets-47 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,697] INFO [Partition __consumer_offsets-47 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-47 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,697] INFO [Partition __consumer_offsets-47 broker=1] Log loaded for partition __consumer_offsets-47 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,697] INFO [Broker id=1] Leader __consumer_offsets-47 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,704] INFO [LogLoader partition=__consumer_offsets-17, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,705] INFO Created log for partition __consumer_offsets-17 in /var/lib/kafka/data/__consumer_offsets-17 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,705] INFO [Partition __consumer_offsets-17 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-17 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,705] INFO [Partition __consumer_offsets-17 broker=1] Log loaded for partition __consumer_offsets-17 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,705] INFO [Broker id=1] Leader __consumer_offsets-17 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,712] INFO [LogLoader partition=__consumer_offsets-32, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,712] INFO Created log for partition __consumer_offsets-32 in /var/lib/kafka/data/__consumer_offsets-32 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,712] INFO [Partition __consumer_offsets-32 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-32 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,713] INFO [Partition __consumer_offsets-32 broker=1] Log loaded for partition __consumer_offsets-32 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,713] INFO [Broker id=1] Leader __consumer_offsets-32 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,718] INFO [LogLoader partition=__consumer_offsets-37, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,719] INFO Created log for partition __consumer_offsets-37 in /var/lib/kafka/data/__consumer_offsets-37 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,719] INFO [Partition __consumer_offsets-37 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-37 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,719] INFO [Partition __consumer_offsets-37 broker=1] Log loaded for partition __consumer_offsets-37 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,719] INFO [Broker id=1] Leader __consumer_offsets-37 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,724] INFO [LogLoader partition=__consumer_offsets-7, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,725] INFO Created log for partition __consumer_offsets-7 in /var/lib/kafka/data/__consumer_offsets-7 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,725] INFO [Partition __consumer_offsets-7 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-7 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,725] INFO [Partition __consumer_offsets-7 broker=1] Log loaded for partition __consumer_offsets-7 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,725] INFO [Broker id=1] Leader __consumer_offsets-7 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,730] INFO [LogLoader partition=__consumer_offsets-22, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,731] INFO Created log for partition __consumer_offsets-22 in /var/lib/kafka/data/__consumer_offsets-22 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,731] INFO [Partition __consumer_offsets-22 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-22 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,731] INFO [Partition __consumer_offsets-22 broker=1] Log loaded for partition __consumer_offsets-22 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,731] INFO [Broker id=1] Leader __consumer_offsets-22 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,736] INFO [LogLoader partition=__consumer_offsets-29, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,736] INFO Created log for partition __consumer_offsets-29 in /var/lib/kafka/data/__consumer_offsets-29 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,736] INFO [Partition __consumer_offsets-29 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-29 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,736] INFO [Partition __consumer_offsets-29 broker=1] Log loaded for partition __consumer_offsets-29 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,736] INFO [Broker id=1] Leader __consumer_offsets-29 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,743] INFO [LogLoader partition=__consumer_offsets-44, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,743] INFO Created log for partition __consumer_offsets-44 in /var/lib/kafka/data/__consumer_offsets-44 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,743] INFO [Partition __consumer_offsets-44 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-44 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,744] INFO [Partition __consumer_offsets-44 broker=1] Log loaded for partition __consumer_offsets-44 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,744] INFO [Broker id=1] Leader __consumer_offsets-44 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,749] INFO [LogLoader partition=__consumer_offsets-14, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,750] INFO Created log for partition __consumer_offsets-14 in /var/lib/kafka/data/__consumer_offsets-14 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,750] INFO [Partition __consumer_offsets-14 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-14 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,750] INFO [Partition __consumer_offsets-14 broker=1] Log loaded for partition __consumer_offsets-14 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,750] INFO [Broker id=1] Leader __consumer_offsets-14 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,756] INFO [LogLoader partition=__consumer_offsets-23, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,756] INFO Created log for partition __consumer_offsets-23 in /var/lib/kafka/data/__consumer_offsets-23 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,756] INFO [Partition __consumer_offsets-23 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-23 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,756] INFO [Partition __consumer_offsets-23 broker=1] Log loaded for partition __consumer_offsets-23 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,756] INFO [Broker id=1] Leader __consumer_offsets-23 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,765] INFO [LogLoader partition=__consumer_offsets-38, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,766] INFO Created log for partition __consumer_offsets-38 in /var/lib/kafka/data/__consumer_offsets-38 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,766] INFO [Partition __consumer_offsets-38 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-38 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,766] INFO [Partition __consumer_offsets-38 broker=1] Log loaded for partition __consumer_offsets-38 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,766] INFO [Broker id=1] Leader __consumer_offsets-38 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,773] INFO [LogLoader partition=__consumer_offsets-8, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,774] INFO Created log for partition __consumer_offsets-8 in /var/lib/kafka/data/__consumer_offsets-8 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,774] INFO [Partition __consumer_offsets-8 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-8 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,774] INFO [Partition __consumer_offsets-8 broker=1] Log loaded for partition __consumer_offsets-8 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,774] INFO [Broker id=1] Leader __consumer_offsets-8 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,779] INFO [LogLoader partition=policy-pdp-pap-0, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,779] INFO Created log for partition policy-pdp-pap-0 in /var/lib/kafka/data/policy-pdp-pap-0 with properties {} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,779] INFO [Partition policy-pdp-pap-0 broker=1] No checkpointed highwatermark is found for partition policy-pdp-pap-0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,779] INFO [Partition policy-pdp-pap-0 broker=1] Log loaded for partition policy-pdp-pap-0 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,780] INFO [Broker id=1] Leader policy-pdp-pap-0 with topic id Some(jb_1oiljSJehU0V5qY0tiA) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,789] INFO [LogLoader partition=__consumer_offsets-45, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,789] INFO Created log for partition __consumer_offsets-45 in /var/lib/kafka/data/__consumer_offsets-45 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,789] INFO [Partition __consumer_offsets-45 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-45 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,789] INFO [Partition __consumer_offsets-45 broker=1] Log loaded for partition __consumer_offsets-45 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,789] INFO [Broker id=1] Leader __consumer_offsets-45 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,796] INFO [LogLoader partition=__consumer_offsets-15, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,796] INFO Created log for partition __consumer_offsets-15 in /var/lib/kafka/data/__consumer_offsets-15 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,796] INFO [Partition __consumer_offsets-15 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-15 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,796] INFO [Partition __consumer_offsets-15 broker=1] Log loaded for partition __consumer_offsets-15 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,796] INFO [Broker id=1] Leader __consumer_offsets-15 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,804] INFO [LogLoader partition=__consumer_offsets-30, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,804] INFO Created log for partition __consumer_offsets-30 in /var/lib/kafka/data/__consumer_offsets-30 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,805] INFO [Partition __consumer_offsets-30 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-30 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,805] INFO [Partition __consumer_offsets-30 broker=1] Log loaded for partition __consumer_offsets-30 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,805] INFO [Broker id=1] Leader __consumer_offsets-30 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,812] INFO [LogLoader partition=__consumer_offsets-0, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,813] INFO Created log for partition __consumer_offsets-0 in /var/lib/kafka/data/__consumer_offsets-0 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,813] INFO [Partition __consumer_offsets-0 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,813] INFO [Partition __consumer_offsets-0 broker=1] Log loaded for partition __consumer_offsets-0 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,813] INFO [Broker id=1] Leader __consumer_offsets-0 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,822] INFO [LogLoader partition=__consumer_offsets-35, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,822] INFO Created log for partition __consumer_offsets-35 in /var/lib/kafka/data/__consumer_offsets-35 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,822] INFO [Partition __consumer_offsets-35 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-35 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,822] INFO [Partition __consumer_offsets-35 broker=1] Log loaded for partition __consumer_offsets-35 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,822] INFO [Broker id=1] Leader __consumer_offsets-35 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,828] INFO [LogLoader partition=__consumer_offsets-5, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,829] INFO Created log for partition __consumer_offsets-5 in /var/lib/kafka/data/__consumer_offsets-5 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,829] INFO [Partition __consumer_offsets-5 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-5 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,829] INFO [Partition __consumer_offsets-5 broker=1] Log loaded for partition __consumer_offsets-5 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,829] INFO [Broker id=1] Leader __consumer_offsets-5 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,837] INFO [LogLoader partition=__consumer_offsets-20, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,838] INFO Created log for partition __consumer_offsets-20 in /var/lib/kafka/data/__consumer_offsets-20 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,838] INFO [Partition __consumer_offsets-20 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-20 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,838] INFO [Partition __consumer_offsets-20 broker=1] Log loaded for partition __consumer_offsets-20 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,839] INFO [Broker id=1] Leader __consumer_offsets-20 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,846] INFO [LogLoader partition=__consumer_offsets-27, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,847] INFO Created log for partition __consumer_offsets-27 in /var/lib/kafka/data/__consumer_offsets-27 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,847] INFO [Partition __consumer_offsets-27 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-27 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,847] INFO [Partition __consumer_offsets-27 broker=1] Log loaded for partition __consumer_offsets-27 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,847] INFO [Broker id=1] Leader __consumer_offsets-27 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,854] INFO [LogLoader partition=__consumer_offsets-42, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,854] INFO Created log for partition __consumer_offsets-42 in /var/lib/kafka/data/__consumer_offsets-42 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,854] INFO [Partition __consumer_offsets-42 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-42 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,854] INFO [Partition __consumer_offsets-42 broker=1] Log loaded for partition __consumer_offsets-42 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,854] INFO [Broker id=1] Leader __consumer_offsets-42 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,861] INFO [LogLoader partition=__consumer_offsets-12, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,862] INFO Created log for partition __consumer_offsets-12 in /var/lib/kafka/data/__consumer_offsets-12 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,862] INFO [Partition __consumer_offsets-12 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-12 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,862] INFO [Partition __consumer_offsets-12 broker=1] Log loaded for partition __consumer_offsets-12 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,862] INFO [Broker id=1] Leader __consumer_offsets-12 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,870] INFO [LogLoader partition=__consumer_offsets-21, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,870] INFO Created log for partition __consumer_offsets-21 in /var/lib/kafka/data/__consumer_offsets-21 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,870] INFO [Partition __consumer_offsets-21 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-21 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,870] INFO [Partition __consumer_offsets-21 broker=1] Log loaded for partition __consumer_offsets-21 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,871] INFO [Broker id=1] Leader __consumer_offsets-21 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,882] INFO [LogLoader partition=__consumer_offsets-36, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,883] INFO Created log for partition __consumer_offsets-36 in /var/lib/kafka/data/__consumer_offsets-36 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,883] INFO [Partition __consumer_offsets-36 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-36 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,883] INFO [Partition __consumer_offsets-36 broker=1] Log loaded for partition __consumer_offsets-36 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,883] INFO [Broker id=1] Leader __consumer_offsets-36 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,890] INFO [LogLoader partition=__consumer_offsets-6, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,891] INFO Created log for partition __consumer_offsets-6 in /var/lib/kafka/data/__consumer_offsets-6 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,891] INFO [Partition __consumer_offsets-6 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-6 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,891] INFO [Partition __consumer_offsets-6 broker=1] Log loaded for partition __consumer_offsets-6 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,891] INFO [Broker id=1] Leader __consumer_offsets-6 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,899] INFO [LogLoader partition=__consumer_offsets-43, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,901] INFO Created log for partition __consumer_offsets-43 in /var/lib/kafka/data/__consumer_offsets-43 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,901] INFO [Partition __consumer_offsets-43 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-43 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,901] INFO [Partition __consumer_offsets-43 broker=1] Log loaded for partition __consumer_offsets-43 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,901] INFO [Broker id=1] Leader __consumer_offsets-43 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,911] INFO [LogLoader partition=__consumer_offsets-13, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,911] INFO Created log for partition __consumer_offsets-13 in /var/lib/kafka/data/__consumer_offsets-13 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,912] INFO [Partition __consumer_offsets-13 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-13 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,912] INFO [Partition __consumer_offsets-13 broker=1] Log loaded for partition __consumer_offsets-13 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,913] INFO [Broker id=1] Leader __consumer_offsets-13 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,919] INFO [LogLoader partition=__consumer_offsets-28, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) 09:44:48 kafka | [2024-07-04 09:42:38,920] INFO Created log for partition __consumer_offsets-28 in /var/lib/kafka/data/__consumer_offsets-28 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) 09:44:48 kafka | [2024-07-04 09:42:38,920] INFO [Partition __consumer_offsets-28 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-28 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,920] INFO [Partition __consumer_offsets-28 broker=1] Log loaded for partition __consumer_offsets-28 with initial high watermark 0 (kafka.cluster.Partition) 09:44:48 kafka | [2024-07-04 09:42:38,920] INFO [Broker id=1] Leader __consumer_offsets-28 with topic id Some(ZjUdxuMGT2q3AHEVEv_YSw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,926] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-3 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,926] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-18 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,926] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-41 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,927] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-10 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,927] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-33 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,927] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-48 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,927] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-19 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,927] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-34 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,927] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-4 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,927] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-11 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,928] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-26 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,928] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-49 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,928] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-39 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,928] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-9 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,928] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-24 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,928] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-31 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,928] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-46 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,929] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-1 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,929] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-16 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,929] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,929] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-25 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,929] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-40 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,929] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-47 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,929] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-17 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,930] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-32 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,930] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-37 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,930] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-7 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,930] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-22 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,930] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-29 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,930] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-44 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,930] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-14 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,931] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-23 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,931] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-38 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,931] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-8 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,931] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition policy-pdp-pap-0 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,931] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-45 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,931] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-15 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,931] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-30 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,932] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-0 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,932] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-35 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,932] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-5 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,932] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-20 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,932] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-27 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,932] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-42 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,932] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-12 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,933] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-21 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,933] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-36 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,933] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-6 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,933] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-43 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,933] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-13 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,933] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-28 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,939] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 3 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,940] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-3 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,941] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 18 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,942] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-18 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,942] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 41 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,942] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-41 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,942] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 10 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,942] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-10 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,942] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 33 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-33 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 48 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-48 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 19 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-19 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 34 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-34 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 4 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-4 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,943] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 11 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-11 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 26 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-26 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 49 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-49 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 39 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-39 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 9 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-9 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,944] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 24 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-24 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 31 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-31 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 46 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-46 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 1 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-1 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 16 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-16 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,945] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 2 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,946] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-2 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,946] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 25 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,946] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-25 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,946] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-3 in 5 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 40 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-40 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-18 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 47 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-47 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-41 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-10 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 17 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-33 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-17 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-48 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,948] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 32 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-19 in 6 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-32 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-34 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-4 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 37 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-11 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-37 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-26 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 7 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-49 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-7 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 22 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-22 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,949] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-39 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 29 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-29 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-9 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 44 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-44 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 14 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-14 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-24 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 23 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,950] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-23 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,951] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-31 in 6 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,951] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 38 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,951] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-46 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,951] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-38 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,951] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-1 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,951] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 8 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,951] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-16 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,952] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-8 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,952] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-2 in 6 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,952] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-25 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,952] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 45 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,952] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-40 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,952] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-45 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,952] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 15 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,952] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-47 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,953] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-15 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,953] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-17 in 5 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,953] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 30 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,953] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-32 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,953] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-30 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,953] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-37 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,953] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 0 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,953] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-7 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,954] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-0 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,954] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-22 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,954] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-29 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,954] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 35 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,954] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-44 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,954] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-35 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,954] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-14 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,955] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 5 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,955] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-23 in 4 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,955] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-5 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,955] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-38 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,955] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 20 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,955] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-8 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,955] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-20 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,955] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-45 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,956] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 27 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,956] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-15 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,956] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-27 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,956] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-30 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,956] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 42 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,956] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-0 in 2 milliseconds for epoch 0, of which 2 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,956] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-42 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,956] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-35 in 2 milliseconds for epoch 0, of which 2 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,957] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 12 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,957] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-5 in 2 milliseconds for epoch 0, of which 2 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,957] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-12 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,957] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-20 in 2 milliseconds for epoch 0, of which 2 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,957] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 21 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,957] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-27 in 1 milliseconds for epoch 0, of which 1 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,957] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-21 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,957] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-42 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,958] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 36 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,958] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-12 in 1 milliseconds for epoch 0, of which 1 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,958] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-36 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,958] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-21 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,958] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 6 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,958] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-36 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-6 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 43 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-43 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-6 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 13 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-13 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-43 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 28 in epoch 0 (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-28 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,959] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-13 in 0 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,960] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-28 in 1 milliseconds for epoch 0, of which 0 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) 09:44:48 kafka | [2024-07-04 09:42:38,962] INFO [Broker id=1] Finished LeaderAndIsr request in 657ms correlationId 1 from controller 1 for 51 partitions (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,966] TRACE [Controller id=1 epoch=1] Received response LeaderAndIsrResponseData(errorCode=0, partitionErrors=[], topics=[LeaderAndIsrTopicError(topicId=ZjUdxuMGT2q3AHEVEv_YSw, partitionErrors=[LeaderAndIsrPartitionError(topicName='', partitionIndex=13, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=46, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=9, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=42, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=21, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=17, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=30, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=26, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=5, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=38, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=1, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=34, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=16, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=45, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=12, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=41, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=24, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=20, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=49, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=0, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=29, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=25, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=8, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=37, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=4, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=33, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=15, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=48, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=11, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=44, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=23, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=19, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=32, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=28, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=7, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=40, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=3, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=36, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=47, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=14, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=43, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=10, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=22, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=18, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=31, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=27, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=39, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=6, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=35, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=2, errorCode=0)]), LeaderAndIsrTopicError(topicId=jb_1oiljSJehU0V5qY0tiA, partitionErrors=[LeaderAndIsrPartitionError(topicName='', partitionIndex=0, errorCode=0)])]) for request LEADER_AND_ISR with correlation id 1 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='policy-pdp-pap', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition policy-pdp-pap-0 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=13, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-13 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=46, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-46 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=9, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-9 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=42, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-42 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=21, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-21 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=17, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-17 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=30, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-30 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=26, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-26 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=5, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-5 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=38, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-38 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=1, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-1 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=34, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-34 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=16, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-16 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=45, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-45 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=12, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-12 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=41, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-41 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=24, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-24 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=20, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-20 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=49, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-49 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-0 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=29, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-29 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=25, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-25 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=8, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-8 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=37, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-37 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=4, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-4 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,972] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=33, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-33 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=15, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-15 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=48, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-48 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=11, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-11 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=44, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-44 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=23, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-23 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=19, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-19 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=32, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-32 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=28, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-28 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=7, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-7 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=40, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-40 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=3, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-3 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=36, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-36 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=47, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-47 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=14, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-14 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=43, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-43 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=10, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-10 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=22, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-22 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=18, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-18 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=31, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-31 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=27, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-27 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=39, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-39 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=6, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-6 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=35, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-35 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,973] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=2, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-2 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,974] INFO [Broker id=1] Add 51 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:38,975] TRACE [Controller id=1 epoch=1] Received response UpdateMetadataResponseData(errorCode=0) for request UPDATE_METADATA with correlation id 2 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) 09:44:48 kafka | [2024-07-04 09:42:39,012] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group c181ccfa-a5ba-49d2-a57d-7770c93062fb in Empty state. Created a new member id consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3-6b61b2ad-4dad-4d27-a110-dd1a2dddc971 and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:39,024] INFO [GroupCoordinator 1]: Preparing to rebalance group c181ccfa-a5ba-49d2-a57d-7770c93062fb in state PreparingRebalance with old generation 0 (__consumer_offsets-27) (reason: Adding new member consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3-6b61b2ad-4dad-4d27-a110-dd1a2dddc971 with group instance id None; client reason: rebalance failed due to MemberIdRequiredException) (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:39,027] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group policy-pap in Empty state. Created a new member id consumer-policy-pap-4-bc3a0567-01a8-42b4-92fb-cd90febcdb27 and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:39,030] INFO [GroupCoordinator 1]: Preparing to rebalance group policy-pap in state PreparingRebalance with old generation 0 (__consumer_offsets-24) (reason: Adding new member consumer-policy-pap-4-bc3a0567-01a8-42b4-92fb-cd90febcdb27 with group instance id None; client reason: rebalance failed due to MemberIdRequiredException) (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:39,702] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group 43992874-9c1b-4e05-b493-a9684c6c5296 in Empty state. Created a new member id consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2-198c82d0-4714-4755-bc78-bfca51e47337 and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:39,711] INFO [GroupCoordinator 1]: Preparing to rebalance group 43992874-9c1b-4e05-b493-a9684c6c5296 in state PreparingRebalance with old generation 0 (__consumer_offsets-35) (reason: Adding new member consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2-198c82d0-4714-4755-bc78-bfca51e47337 with group instance id None; client reason: rebalance failed due to MemberIdRequiredException) (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:42,036] INFO [GroupCoordinator 1]: Stabilized group c181ccfa-a5ba-49d2-a57d-7770c93062fb generation 1 (__consumer_offsets-27) with 1 members (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:42,042] INFO [GroupCoordinator 1]: Stabilized group policy-pap generation 1 (__consumer_offsets-24) with 1 members (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:42,060] INFO [GroupCoordinator 1]: Assignment received from leader consumer-policy-pap-4-bc3a0567-01a8-42b4-92fb-cd90febcdb27 for group policy-pap for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:42,060] INFO [GroupCoordinator 1]: Assignment received from leader consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3-6b61b2ad-4dad-4d27-a110-dd1a2dddc971 for group c181ccfa-a5ba-49d2-a57d-7770c93062fb for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:42,713] INFO [GroupCoordinator 1]: Stabilized group 43992874-9c1b-4e05-b493-a9684c6c5296 generation 1 (__consumer_offsets-35) with 1 members (kafka.coordinator.group.GroupCoordinator) 09:44:48 kafka | [2024-07-04 09:42:42,724] INFO [GroupCoordinator 1]: Assignment received from leader consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2-198c82d0-4714-4755-bc78-bfca51e47337 for group 43992874-9c1b-4e05-b493-a9684c6c5296 for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) 09:44:48 =================================== 09:44:48 ======== Logs from mariadb ======== 09:44:48 mariadb | 2024-07-04 09:41:53+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:10.10.2+maria~ubu2204 started. 09:44:48 mariadb | 2024-07-04 09:41:53+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql' 09:44:48 mariadb | 2024-07-04 09:41:53+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:10.10.2+maria~ubu2204 started. 09:44:48 mariadb | 2024-07-04 09:41:53+00:00 [Note] [Entrypoint]: Initializing database files 09:44:48 mariadb | 2024-07-04 9:41:53 0 [Warning] mariadbd: io_uring_queue_init() failed with ENOSYS: check seccomp filters, and the kernel version (newer than 5.1 required) 09:44:48 mariadb | 2024-07-04 9:41:53 0 [Warning] InnoDB: liburing disabled: falling back to innodb_use_native_aio=OFF 09:44:48 mariadb | 2024-07-04 9:41:53 0 [Warning] 'default-authentication-plugin' is MySQL 5.6 / 5.7 compatible option. To be implemented in later versions. 09:44:48 mariadb | 09:44:48 mariadb | 09:44:48 mariadb | PLEASE REMEMBER TO SET A PASSWORD FOR THE MariaDB root USER ! 09:44:48 mariadb | To do so, start the server, then issue the following command: 09:44:48 mariadb | 09:44:48 mariadb | '/usr/bin/mysql_secure_installation' 09:44:48 mariadb | 09:44:48 mariadb | which will also give you the option of removing the test 09:44:48 mariadb | databases and anonymous user created by default. This is 09:44:48 mariadb | strongly recommended for production servers. 09:44:48 mariadb | 09:44:48 mariadb | See the MariaDB Knowledgebase at https://mariadb.com/kb 09:44:48 mariadb | 09:44:48 mariadb | Please report any problems at https://mariadb.org/jira 09:44:48 mariadb | 09:44:48 mariadb | The latest information about MariaDB is available at https://mariadb.org/. 09:44:48 mariadb | 09:44:48 mariadb | Consider joining MariaDB's strong and vibrant community: 09:44:48 mariadb | https://mariadb.org/get-involved/ 09:44:48 mariadb | 09:44:48 mariadb | 2024-07-04 09:41:56+00:00 [Note] [Entrypoint]: Database files initialized 09:44:48 mariadb | 2024-07-04 09:41:56+00:00 [Note] [Entrypoint]: Starting temporary server 09:44:48 mariadb | 2024-07-04 09:41:56+00:00 [Note] [Entrypoint]: Waiting for server startup 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] mariadbd (server 10.10.2-MariaDB-1:10.10.2+maria~ubu2204) starting as process 95 ... 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: Compressed tables use zlib 1.2.11 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: Number of transaction pools: 1 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: Using crc32 + pclmulqdq instructions 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] mariadbd: O_TMPFILE is not supported on /tmp (disabling future attempts) 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Warning] mariadbd: io_uring_queue_init() failed with ENOSYS: check seccomp filters, and the kernel version (newer than 5.1 required) 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Warning] InnoDB: liburing disabled: falling back to innodb_use_native_aio=OFF 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: Completed initialization of buffer pool 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes) 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: 128 rollback segments are active. 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ... 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB. 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] InnoDB: log sequence number 46590; transaction id 14 09:44:48 mariadb | 2024-07-04 9:41:56 0 [Note] Plugin 'FEEDBACK' is disabled. 09:44:48 mariadb | 2024-07-04 9:41:57 0 [Warning] 'default-authentication-plugin' is MySQL 5.6 / 5.7 compatible option. To be implemented in later versions. 09:44:48 mariadb | 2024-07-04 9:41:57 0 [Warning] 'user' entry 'root@mariadb' ignored in --skip-name-resolve mode. 09:44:48 mariadb | 2024-07-04 9:41:57 0 [Warning] 'proxies_priv' entry '@% root@mariadb' ignored in --skip-name-resolve mode. 09:44:48 mariadb | 2024-07-04 9:41:57 0 [Note] mariadbd: ready for connections. 09:44:48 mariadb | Version: '10.10.2-MariaDB-1:10.10.2+maria~ubu2204' socket: '/run/mysqld/mysqld.sock' port: 0 mariadb.org binary distribution 09:44:48 mariadb | 2024-07-04 09:41:57+00:00 [Note] [Entrypoint]: Temporary server started. 09:44:48 mariadb | 2024-07-04 09:42:00+00:00 [Note] [Entrypoint]: Creating user policy_user 09:44:48 mariadb | 2024-07-04 09:42:00+00:00 [Note] [Entrypoint]: Securing system users (equivalent to running mysql_secure_installation) 09:44:48 mariadb | 09:44:48 mariadb | 09:44:48 mariadb | 2024-07-04 09:42:00+00:00 [Warn] [Entrypoint]: /usr/local/bin/docker-entrypoint.sh: ignoring /docker-entrypoint-initdb.d/db.conf 09:44:48 mariadb | 2024-07-04 09:42:00+00:00 [Note] [Entrypoint]: /usr/local/bin/docker-entrypoint.sh: running /docker-entrypoint-initdb.d/db.sh 09:44:48 mariadb | #!/bin/bash -xv 09:44:48 mariadb | # Copyright 2019,2021 AT&T Intellectual Property. All rights reserved 09:44:48 mariadb | # Modifications Copyright (c) 2022 Nordix Foundation. 09:44:48 mariadb | # 09:44:48 mariadb | # Licensed under the Apache License, Version 2.0 (the "License"); 09:44:48 mariadb | # you may not use this file except in compliance with the License. 09:44:48 mariadb | # You may obtain a copy of the License at 09:44:48 mariadb | # 09:44:48 mariadb | # http://www.apache.org/licenses/LICENSE-2.0 09:44:48 mariadb | # 09:44:48 mariadb | # Unless required by applicable law or agreed to in writing, software 09:44:48 mariadb | # distributed under the License is distributed on an "AS IS" BASIS, 09:44:48 mariadb | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 09:44:48 mariadb | # See the License for the specific language governing permissions and 09:44:48 mariadb | # limitations under the License. 09:44:48 mariadb | 09:44:48 mariadb | for db in migration pooling policyadmin operationshistory clampacm policyclamp 09:44:48 mariadb | do 09:44:48 mariadb | mysql -uroot -p"${MYSQL_ROOT_PASSWORD}" --execute "CREATE DATABASE IF NOT EXISTS ${db};" 09:44:48 mariadb | mysql -uroot -p"${MYSQL_ROOT_PASSWORD}" --execute "GRANT ALL PRIVILEGES ON \`${db}\`.* TO '${MYSQL_USER}'@'%' ;" 09:44:48 mariadb | done 09:44:48 mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp 09:44:48 mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS migration;' 09:44:48 mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `migration`.* TO '\''policy_user'\''@'\''%'\'' ;' 09:44:48 mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp 09:44:48 mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS pooling;' 09:44:48 mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `pooling`.* TO '\''policy_user'\''@'\''%'\'' ;' 09:44:48 mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp 09:44:48 mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS policyadmin;' 09:44:48 mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `policyadmin`.* TO '\''policy_user'\''@'\''%'\'' ;' 09:44:48 mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp 09:44:48 mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS operationshistory;' 09:44:48 mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `operationshistory`.* TO '\''policy_user'\''@'\''%'\'' ;' 09:44:48 mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp 09:44:48 mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS clampacm;' 09:44:48 mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `clampacm`.* TO '\''policy_user'\''@'\''%'\'' ;' 09:44:48 mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp 09:44:48 mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS policyclamp;' 09:44:48 mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `policyclamp`.* TO '\''policy_user'\''@'\''%'\'' ;' 09:44:48 mariadb | 09:44:48 mariadb | mysql -uroot -p"${MYSQL_ROOT_PASSWORD}" --execute "FLUSH PRIVILEGES;" 09:44:48 mariadb | + mysql -uroot -psecret --execute 'FLUSH PRIVILEGES;' 09:44:48 mariadb | mysql -u"${MYSQL_USER}" -p"${MYSQL_PASSWORD}" -f policyclamp < /tmp/policy-clamp-create-tables.sql 09:44:48 mariadb | + mysql -upolicy_user -ppolicy_user -f policyclamp 09:44:48 mariadb | 09:44:48 mariadb | 2024-07-04 09:42:01+00:00 [Note] [Entrypoint]: Stopping temporary server 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] mariadbd (initiated by: unknown): Normal shutdown 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: FTS optimize thread exiting. 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Starting shutdown... 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Dumping buffer pool(s) to /var/lib/mysql/ib_buffer_pool 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Buffer pool(s) dump completed at 240704 9:42:01 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Removed temporary tablespace data file: "./ibtmp1" 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Shutdown completed; log sequence number 392995; transaction id 298 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] mariadbd: Shutdown complete 09:44:48 mariadb | 09:44:48 mariadb | 2024-07-04 09:42:01+00:00 [Note] [Entrypoint]: Temporary server stopped 09:44:48 mariadb | 09:44:48 mariadb | 2024-07-04 09:42:01+00:00 [Note] [Entrypoint]: MariaDB init process done. Ready for start up. 09:44:48 mariadb | 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] mariadbd (server 10.10.2-MariaDB-1:10.10.2+maria~ubu2204) starting as process 1 ... 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Compressed tables use zlib 1.2.11 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Number of transaction pools: 1 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Using crc32 + pclmulqdq instructions 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] mariadbd: O_TMPFILE is not supported on /tmp (disabling future attempts) 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Warning] mariadbd: io_uring_queue_init() failed with ENOSYS: check seccomp filters, and the kernel version (newer than 5.1 required) 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Warning] InnoDB: liburing disabled: falling back to innodb_use_native_aio=OFF 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Completed initialization of buffer pool 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes) 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: 128 rollback segments are active. 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ... 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB. 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: log sequence number 392995; transaction id 299 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] Plugin 'FEEDBACK' is disabled. 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Warning] 'default-authentication-plugin' is MySQL 5.6 / 5.7 compatible option. To be implemented in later versions. 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Warning] You need to use --log-bin to make --expire-logs-days or --binlog-expire-logs-seconds work. 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] Server socket created on IP: '0.0.0.0'. 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] Server socket created on IP: '::'. 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] mariadbd: ready for connections. 09:44:48 mariadb | Version: '10.10.2-MariaDB-1:10.10.2+maria~ubu2204' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution 09:44:48 mariadb | 2024-07-04 9:42:01 0 [Note] InnoDB: Buffer pool(s) load completed at 240704 9:42:01 09:44:48 mariadb | 2024-07-04 9:42:01 3 [Warning] Aborted connection 3 to db: 'unconnected' user: 'unauthenticated' host: '172.17.0.7' (This connection closed normally without authentication) 09:44:48 mariadb | 2024-07-04 9:42:02 30 [Warning] Aborted connection 30 to db: 'unconnected' user: 'unauthenticated' host: '172.17.0.11' (This connection closed normally without authentication) 09:44:48 mariadb | 2024-07-04 9:42:02 31 [Warning] Aborted connection 31 to db: 'unconnected' user: 'unauthenticated' host: '172.17.0.8' (This connection closed normally without authentication) 09:44:48 mariadb | 2024-07-04 9:42:02 32 [Warning] Aborted connection 32 to db: 'unconnected' user: 'unauthenticated' host: '172.17.0.9' (This connection closed normally without authentication) 09:44:48 =================================== 09:44:48 ======== Logs from apex-pdp ======== 09:44:48 policy-apex-pdp | Waiting for mariadb port 3306... 09:44:48 policy-apex-pdp | Waiting for kafka port 9092... 09:44:48 policy-apex-pdp | Waiting for pap port 6969... 09:44:48 policy-apex-pdp | mariadb (172.17.0.3:3306) open 09:44:48 policy-apex-pdp | kafka (172.17.0.6:9092) open 09:44:48 policy-apex-pdp | pap (172.17.0.9:6969) open 09:44:48 policy-apex-pdp | apexApps.sh: running application 'onappf' with command 'java -Dlogback.configurationFile=/opt/app/policy/apex-pdp/etc/logback.xml -cp /opt/app/policy/apex-pdp/etc:/opt/app/policy/apex-pdp/etc/hazelcast:/opt/app/policy/apex-pdp/etc/infinispan:/opt/app/policy/apex-pdp/lib/* -Djavax.net.ssl.keyStore=/opt/app/policy/apex-pdp/etc/ssl/policy-keystore -Djavax.net.ssl.keyStorePassword=Pol1cy_0nap -Djavax.net.ssl.trustStore=/opt/app/policy/apex-pdp/etc/ssl/policy-truststore -Djavax.net.ssl.trustStorePassword=Pol1cy_0nap -Dlogback.configurationFile=/opt/app/policy/apex-pdp/etc/logback.xml -Dhazelcast.config=/opt/app/policy/apex-pdp/etc/hazelcast.xml -Dhazelcast.mancenter.enabled=false org.onap.policy.apex.services.onappf.ApexStarterMain -c /opt/app/policy/apex-pdp/etc/onappf/config/OnapPfConfig.json' 09:44:48 policy-apex-pdp | [2024-07-04T09:42:38.861+00:00|INFO|ApexStarterMain|main] In ApexStarter with parameters [-c, /opt/app/policy/apex-pdp/etc/onappf/config/OnapPfConfig.json] 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.086+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: 09:44:48 policy-apex-pdp | allow.auto.create.topics = true 09:44:48 policy-apex-pdp | auto.commit.interval.ms = 5000 09:44:48 policy-apex-pdp | auto.include.jmx.reporter = true 09:44:48 policy-apex-pdp | auto.offset.reset = latest 09:44:48 policy-apex-pdp | bootstrap.servers = [kafka:9092] 09:44:48 policy-apex-pdp | check.crcs = true 09:44:48 policy-apex-pdp | client.dns.lookup = use_all_dns_ips 09:44:48 policy-apex-pdp | client.id = consumer-43992874-9c1b-4e05-b493-a9684c6c5296-1 09:44:48 policy-apex-pdp | client.rack = 09:44:48 policy-apex-pdp | connections.max.idle.ms = 540000 09:44:48 policy-apex-pdp | default.api.timeout.ms = 60000 09:44:48 policy-apex-pdp | enable.auto.commit = true 09:44:48 policy-apex-pdp | exclude.internal.topics = true 09:44:48 policy-apex-pdp | fetch.max.bytes = 52428800 09:44:48 policy-apex-pdp | fetch.max.wait.ms = 500 09:44:48 policy-apex-pdp | fetch.min.bytes = 1 09:44:48 policy-apex-pdp | group.id = 43992874-9c1b-4e05-b493-a9684c6c5296 09:44:48 policy-apex-pdp | group.instance.id = null 09:44:48 policy-apex-pdp | heartbeat.interval.ms = 3000 09:44:48 policy-apex-pdp | interceptor.classes = [] 09:44:48 policy-apex-pdp | internal.leave.group.on.close = true 09:44:48 policy-apex-pdp | internal.throw.on.fetch.stable.offset.unsupported = false 09:44:48 policy-apex-pdp | isolation.level = read_uncommitted 09:44:48 policy-apex-pdp | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-apex-pdp | max.partition.fetch.bytes = 1048576 09:44:48 policy-apex-pdp | max.poll.interval.ms = 300000 09:44:48 policy-apex-pdp | max.poll.records = 500 09:44:48 policy-apex-pdp | metadata.max.age.ms = 300000 09:44:48 policy-apex-pdp | metric.reporters = [] 09:44:48 policy-apex-pdp | metrics.num.samples = 2 09:44:48 policy-apex-pdp | metrics.recording.level = INFO 09:44:48 policy-apex-pdp | metrics.sample.window.ms = 30000 09:44:48 policy-apex-pdp | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] 09:44:48 policy-apex-pdp | receive.buffer.bytes = 65536 09:44:48 policy-apex-pdp | reconnect.backoff.max.ms = 1000 09:44:48 policy-apex-pdp | reconnect.backoff.ms = 50 09:44:48 policy-apex-pdp | request.timeout.ms = 30000 09:44:48 policy-apex-pdp | retry.backoff.ms = 100 09:44:48 policy-apex-pdp | sasl.client.callback.handler.class = null 09:44:48 policy-apex-pdp | sasl.jaas.config = null 09:44:48 policy-apex-pdp | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 policy-apex-pdp | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 policy-apex-pdp | sasl.kerberos.service.name = null 09:44:48 policy-apex-pdp | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 policy-apex-pdp | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 policy-apex-pdp | sasl.login.callback.handler.class = null 09:44:48 policy-apex-pdp | sasl.login.class = null 09:44:48 policy-apex-pdp | sasl.login.connect.timeout.ms = null 09:44:48 policy-apex-pdp | sasl.login.read.timeout.ms = null 09:44:48 policy-apex-pdp | sasl.login.refresh.buffer.seconds = 300 09:44:48 policy-apex-pdp | sasl.login.refresh.min.period.seconds = 60 09:44:48 policy-apex-pdp | sasl.login.refresh.window.factor = 0.8 09:44:48 policy-apex-pdp | sasl.login.refresh.window.jitter = 0.05 09:44:48 policy-apex-pdp | sasl.login.retry.backoff.max.ms = 10000 09:44:48 policy-apex-pdp | sasl.login.retry.backoff.ms = 100 09:44:48 policy-apex-pdp | sasl.mechanism = GSSAPI 09:44:48 policy-apex-pdp | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 policy-apex-pdp | sasl.oauthbearer.expected.audience = null 09:44:48 policy-apex-pdp | sasl.oauthbearer.expected.issuer = null 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 policy-apex-pdp | sasl.oauthbearer.scope.claim.name = scope 09:44:48 policy-apex-pdp | sasl.oauthbearer.sub.claim.name = sub 09:44:48 policy-apex-pdp | sasl.oauthbearer.token.endpoint.url = null 09:44:48 policy-apex-pdp | security.protocol = PLAINTEXT 09:44:48 policy-apex-pdp | security.providers = null 09:44:48 policy-apex-pdp | send.buffer.bytes = 131072 09:44:48 policy-apex-pdp | session.timeout.ms = 45000 09:44:48 policy-apex-pdp | socket.connection.setup.timeout.max.ms = 30000 09:44:48 policy-apex-pdp | socket.connection.setup.timeout.ms = 10000 09:44:48 policy-apex-pdp | ssl.cipher.suites = null 09:44:48 policy-apex-pdp | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 policy-apex-pdp | ssl.endpoint.identification.algorithm = https 09:44:48 policy-apex-pdp | ssl.engine.factory.class = null 09:44:48 policy-apex-pdp | ssl.key.password = null 09:44:48 policy-apex-pdp | ssl.keymanager.algorithm = SunX509 09:44:48 policy-apex-pdp | ssl.keystore.certificate.chain = null 09:44:48 policy-apex-pdp | ssl.keystore.key = null 09:44:48 policy-apex-pdp | ssl.keystore.location = null 09:44:48 policy-apex-pdp | ssl.keystore.password = null 09:44:48 policy-apex-pdp | ssl.keystore.type = JKS 09:44:48 policy-apex-pdp | ssl.protocol = TLSv1.3 09:44:48 policy-apex-pdp | ssl.provider = null 09:44:48 policy-apex-pdp | ssl.secure.random.implementation = null 09:44:48 policy-apex-pdp | ssl.trustmanager.algorithm = PKIX 09:44:48 policy-apex-pdp | ssl.truststore.certificates = null 09:44:48 policy-apex-pdp | ssl.truststore.location = null 09:44:48 policy-apex-pdp | ssl.truststore.password = null 09:44:48 policy-apex-pdp | ssl.truststore.type = JKS 09:44:48 policy-apex-pdp | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-apex-pdp | 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.233+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.233+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.233+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1720086159231 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.235+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-1, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Subscribed to topic(s): policy-pdp-pap 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.246+00:00|INFO|ServiceManager|main] service manager starting 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.246+00:00|INFO|ServiceManager|main] service manager starting topics 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.247+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=43992874-9c1b-4e05-b493-a9684c6c5296, consumerInstance=policy-apex-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: starting 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.265+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: 09:44:48 policy-apex-pdp | allow.auto.create.topics = true 09:44:48 policy-apex-pdp | auto.commit.interval.ms = 5000 09:44:48 policy-apex-pdp | auto.include.jmx.reporter = true 09:44:48 policy-apex-pdp | auto.offset.reset = latest 09:44:48 policy-apex-pdp | bootstrap.servers = [kafka:9092] 09:44:48 policy-apex-pdp | check.crcs = true 09:44:48 policy-apex-pdp | client.dns.lookup = use_all_dns_ips 09:44:48 policy-apex-pdp | client.id = consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2 09:44:48 policy-apex-pdp | client.rack = 09:44:48 policy-apex-pdp | connections.max.idle.ms = 540000 09:44:48 policy-apex-pdp | default.api.timeout.ms = 60000 09:44:48 policy-apex-pdp | enable.auto.commit = true 09:44:48 policy-apex-pdp | exclude.internal.topics = true 09:44:48 policy-apex-pdp | fetch.max.bytes = 52428800 09:44:48 policy-apex-pdp | fetch.max.wait.ms = 500 09:44:48 policy-apex-pdp | fetch.min.bytes = 1 09:44:48 policy-apex-pdp | group.id = 43992874-9c1b-4e05-b493-a9684c6c5296 09:44:48 policy-apex-pdp | group.instance.id = null 09:44:48 policy-apex-pdp | heartbeat.interval.ms = 3000 09:44:48 policy-apex-pdp | interceptor.classes = [] 09:44:48 policy-apex-pdp | internal.leave.group.on.close = true 09:44:48 policy-apex-pdp | internal.throw.on.fetch.stable.offset.unsupported = false 09:44:48 policy-apex-pdp | isolation.level = read_uncommitted 09:44:48 policy-apex-pdp | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-apex-pdp | max.partition.fetch.bytes = 1048576 09:44:48 policy-apex-pdp | max.poll.interval.ms = 300000 09:44:48 policy-apex-pdp | max.poll.records = 500 09:44:48 policy-apex-pdp | metadata.max.age.ms = 300000 09:44:48 policy-apex-pdp | metric.reporters = [] 09:44:48 policy-apex-pdp | metrics.num.samples = 2 09:44:48 policy-apex-pdp | metrics.recording.level = INFO 09:44:48 policy-apex-pdp | metrics.sample.window.ms = 30000 09:44:48 policy-apex-pdp | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] 09:44:48 policy-apex-pdp | receive.buffer.bytes = 65536 09:44:48 policy-apex-pdp | reconnect.backoff.max.ms = 1000 09:44:48 policy-apex-pdp | reconnect.backoff.ms = 50 09:44:48 policy-apex-pdp | request.timeout.ms = 30000 09:44:48 policy-apex-pdp | retry.backoff.ms = 100 09:44:48 policy-apex-pdp | sasl.client.callback.handler.class = null 09:44:48 policy-apex-pdp | sasl.jaas.config = null 09:44:48 policy-apex-pdp | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 policy-apex-pdp | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 policy-apex-pdp | sasl.kerberos.service.name = null 09:44:48 policy-apex-pdp | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 policy-apex-pdp | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 policy-apex-pdp | sasl.login.callback.handler.class = null 09:44:48 policy-apex-pdp | sasl.login.class = null 09:44:48 policy-apex-pdp | sasl.login.connect.timeout.ms = null 09:44:48 policy-apex-pdp | sasl.login.read.timeout.ms = null 09:44:48 policy-apex-pdp | sasl.login.refresh.buffer.seconds = 300 09:44:48 policy-apex-pdp | sasl.login.refresh.min.period.seconds = 60 09:44:48 policy-apex-pdp | sasl.login.refresh.window.factor = 0.8 09:44:48 policy-apex-pdp | sasl.login.refresh.window.jitter = 0.05 09:44:48 policy-apex-pdp | sasl.login.retry.backoff.max.ms = 10000 09:44:48 policy-apex-pdp | sasl.login.retry.backoff.ms = 100 09:44:48 policy-apex-pdp | sasl.mechanism = GSSAPI 09:44:48 policy-apex-pdp | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 policy-apex-pdp | sasl.oauthbearer.expected.audience = null 09:44:48 policy-apex-pdp | sasl.oauthbearer.expected.issuer = null 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 policy-apex-pdp | sasl.oauthbearer.scope.claim.name = scope 09:44:48 policy-apex-pdp | sasl.oauthbearer.sub.claim.name = sub 09:44:48 policy-apex-pdp | sasl.oauthbearer.token.endpoint.url = null 09:44:48 policy-apex-pdp | security.protocol = PLAINTEXT 09:44:48 policy-apex-pdp | security.providers = null 09:44:48 policy-apex-pdp | send.buffer.bytes = 131072 09:44:48 policy-apex-pdp | session.timeout.ms = 45000 09:44:48 policy-apex-pdp | socket.connection.setup.timeout.max.ms = 30000 09:44:48 policy-apex-pdp | socket.connection.setup.timeout.ms = 10000 09:44:48 policy-apex-pdp | ssl.cipher.suites = null 09:44:48 policy-apex-pdp | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 policy-apex-pdp | ssl.endpoint.identification.algorithm = https 09:44:48 policy-apex-pdp | ssl.engine.factory.class = null 09:44:48 policy-apex-pdp | ssl.key.password = null 09:44:48 policy-apex-pdp | ssl.keymanager.algorithm = SunX509 09:44:48 policy-apex-pdp | ssl.keystore.certificate.chain = null 09:44:48 policy-apex-pdp | ssl.keystore.key = null 09:44:48 policy-apex-pdp | ssl.keystore.location = null 09:44:48 policy-apex-pdp | ssl.keystore.password = null 09:44:48 policy-apex-pdp | ssl.keystore.type = JKS 09:44:48 policy-apex-pdp | ssl.protocol = TLSv1.3 09:44:48 policy-apex-pdp | ssl.provider = null 09:44:48 policy-apex-pdp | ssl.secure.random.implementation = null 09:44:48 policy-apex-pdp | ssl.trustmanager.algorithm = PKIX 09:44:48 policy-apex-pdp | ssl.truststore.certificates = null 09:44:48 policy-apex-pdp | ssl.truststore.location = null 09:44:48 policy-apex-pdp | ssl.truststore.password = null 09:44:48 policy-apex-pdp | ssl.truststore.type = JKS 09:44:48 policy-apex-pdp | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-apex-pdp | 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.273+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.274+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.274+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1720086159273 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.274+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Subscribed to topic(s): policy-pdp-pap 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.274+00:00|INFO|InlineBusTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=029be40a-a332-4e1c-83a0-82703058c4e3, alive=false, publisher=null]]: starting 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.286+00:00|INFO|ProducerConfig|main] ProducerConfig values: 09:44:48 policy-apex-pdp | acks = -1 09:44:48 policy-apex-pdp | auto.include.jmx.reporter = true 09:44:48 policy-apex-pdp | batch.size = 16384 09:44:48 policy-apex-pdp | bootstrap.servers = [kafka:9092] 09:44:48 policy-apex-pdp | buffer.memory = 33554432 09:44:48 policy-apex-pdp | client.dns.lookup = use_all_dns_ips 09:44:48 policy-apex-pdp | client.id = producer-1 09:44:48 policy-apex-pdp | compression.type = none 09:44:48 policy-apex-pdp | connections.max.idle.ms = 540000 09:44:48 policy-apex-pdp | delivery.timeout.ms = 120000 09:44:48 policy-apex-pdp | enable.idempotence = true 09:44:48 policy-apex-pdp | interceptor.classes = [] 09:44:48 policy-apex-pdp | key.serializer = class org.apache.kafka.common.serialization.StringSerializer 09:44:48 policy-apex-pdp | linger.ms = 0 09:44:48 policy-apex-pdp | max.block.ms = 60000 09:44:48 policy-apex-pdp | max.in.flight.requests.per.connection = 5 09:44:48 policy-apex-pdp | max.request.size = 1048576 09:44:48 policy-apex-pdp | metadata.max.age.ms = 300000 09:44:48 policy-apex-pdp | metadata.max.idle.ms = 300000 09:44:48 policy-apex-pdp | metric.reporters = [] 09:44:48 policy-apex-pdp | metrics.num.samples = 2 09:44:48 policy-apex-pdp | metrics.recording.level = INFO 09:44:48 policy-apex-pdp | metrics.sample.window.ms = 30000 09:44:48 policy-apex-pdp | partitioner.adaptive.partitioning.enable = true 09:44:48 policy-apex-pdp | partitioner.availability.timeout.ms = 0 09:44:48 policy-apex-pdp | partitioner.class = null 09:44:48 policy-apex-pdp | partitioner.ignore.keys = false 09:44:48 policy-apex-pdp | receive.buffer.bytes = 32768 09:44:48 policy-apex-pdp | reconnect.backoff.max.ms = 1000 09:44:48 policy-apex-pdp | reconnect.backoff.ms = 50 09:44:48 policy-apex-pdp | request.timeout.ms = 30000 09:44:48 policy-apex-pdp | retries = 2147483647 09:44:48 policy-apex-pdp | retry.backoff.ms = 100 09:44:48 policy-apex-pdp | sasl.client.callback.handler.class = null 09:44:48 policy-apex-pdp | sasl.jaas.config = null 09:44:48 policy-apex-pdp | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 policy-apex-pdp | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 policy-apex-pdp | sasl.kerberos.service.name = null 09:44:48 policy-apex-pdp | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 policy-apex-pdp | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 policy-apex-pdp | sasl.login.callback.handler.class = null 09:44:48 policy-apex-pdp | sasl.login.class = null 09:44:48 policy-apex-pdp | sasl.login.connect.timeout.ms = null 09:44:48 policy-apex-pdp | sasl.login.read.timeout.ms = null 09:44:48 policy-apex-pdp | sasl.login.refresh.buffer.seconds = 300 09:44:48 policy-apex-pdp | sasl.login.refresh.min.period.seconds = 60 09:44:48 policy-apex-pdp | sasl.login.refresh.window.factor = 0.8 09:44:48 policy-apex-pdp | sasl.login.refresh.window.jitter = 0.05 09:44:48 policy-apex-pdp | sasl.login.retry.backoff.max.ms = 10000 09:44:48 policy-apex-pdp | sasl.login.retry.backoff.ms = 100 09:44:48 policy-apex-pdp | sasl.mechanism = GSSAPI 09:44:48 policy-apex-pdp | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 policy-apex-pdp | sasl.oauthbearer.expected.audience = null 09:44:48 policy-apex-pdp | sasl.oauthbearer.expected.issuer = null 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 policy-apex-pdp | sasl.oauthbearer.scope.claim.name = scope 09:44:48 policy-apex-pdp | sasl.oauthbearer.sub.claim.name = sub 09:44:48 policy-apex-pdp | sasl.oauthbearer.token.endpoint.url = null 09:44:48 policy-apex-pdp | security.protocol = PLAINTEXT 09:44:48 policy-apex-pdp | security.providers = null 09:44:48 policy-apex-pdp | send.buffer.bytes = 131072 09:44:48 policy-apex-pdp | socket.connection.setup.timeout.max.ms = 30000 09:44:48 policy-apex-pdp | socket.connection.setup.timeout.ms = 10000 09:44:48 policy-apex-pdp | ssl.cipher.suites = null 09:44:48 policy-apex-pdp | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 policy-apex-pdp | ssl.endpoint.identification.algorithm = https 09:44:48 policy-apex-pdp | ssl.engine.factory.class = null 09:44:48 policy-apex-pdp | ssl.key.password = null 09:44:48 policy-apex-pdp | ssl.keymanager.algorithm = SunX509 09:44:48 policy-apex-pdp | ssl.keystore.certificate.chain = null 09:44:48 policy-apex-pdp | ssl.keystore.key = null 09:44:48 policy-apex-pdp | ssl.keystore.location = null 09:44:48 policy-apex-pdp | ssl.keystore.password = null 09:44:48 policy-apex-pdp | ssl.keystore.type = JKS 09:44:48 policy-apex-pdp | ssl.protocol = TLSv1.3 09:44:48 policy-apex-pdp | ssl.provider = null 09:44:48 policy-apex-pdp | ssl.secure.random.implementation = null 09:44:48 policy-apex-pdp | ssl.trustmanager.algorithm = PKIX 09:44:48 policy-apex-pdp | ssl.truststore.certificates = null 09:44:48 policy-apex-pdp | ssl.truststore.location = null 09:44:48 policy-apex-pdp | ssl.truststore.password = null 09:44:48 policy-apex-pdp | ssl.truststore.type = JKS 09:44:48 policy-apex-pdp | transaction.timeout.ms = 60000 09:44:48 policy-apex-pdp | transactional.id = null 09:44:48 policy-apex-pdp | value.serializer = class org.apache.kafka.common.serialization.StringSerializer 09:44:48 policy-apex-pdp | 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.295+00:00|INFO|KafkaProducer|main] [Producer clientId=producer-1] Instantiated an idempotent producer. 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.310+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.310+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.310+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1720086159310 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.310+00:00|INFO|InlineKafkaTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=029be40a-a332-4e1c-83a0-82703058c4e3, alive=false, publisher=KafkaPublisherWrapper []]]: KAFKA SINK created 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.310+00:00|INFO|ServiceManager|main] service manager starting set alive 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.310+00:00|INFO|ServiceManager|main] service manager starting register pdp status context object 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.312+00:00|INFO|ServiceManager|main] service manager starting topic sinks 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.312+00:00|INFO|ServiceManager|main] service manager starting Pdp Status publisher 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.313+00:00|INFO|ServiceManager|main] service manager starting Register pdp update listener 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.314+00:00|INFO|ServiceManager|main] service manager starting Register pdp state change request dispatcher 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.314+00:00|INFO|ServiceManager|main] service manager starting Message Dispatcher 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.314+00:00|INFO|TopicBase|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=43992874-9c1b-4e05-b493-a9684c6c5296, consumerInstance=policy-apex-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: registering org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher@60a2630a 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.314+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=43992874-9c1b-4e05-b493-a9684c6c5296, consumerInstance=policy-apex-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: register: start not attempted 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.314+00:00|INFO|ServiceManager|main] service manager starting Create REST server 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.326+00:00|INFO|OrderedServiceImpl|Timer-0] ***** OrderedServiceImpl implementers: 09:44:48 policy-apex-pdp | [] 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.329+00:00|INFO|network|Timer-0] [OUT|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"125c7d2b-e164-499b-a407-66ab199c31f7","timestampMs":1720086159314,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.498+00:00|INFO|ServiceManager|main] service manager starting Rest Server 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.498+00:00|INFO|ServiceManager|main] service manager starting 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.498+00:00|INFO|ServiceManager|main] service manager starting REST RestServerParameters 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.498+00:00|INFO|JettyServletServer|main] JettyJerseyServer [Jerseyservlets={/metrics=io.prometheus.client.servlet.jakarta.exporter.MetricsServlet-72b16078==io.prometheus.client.servlet.jakarta.exporter.MetricsServlet@aa16c20f{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:null,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-d78795==org.glassfish.jersey.servlet.ServletContainer@b1764d3c{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=swagger-6969, toString()=JettyServletServer(name=RestServerParameters, host=0.0.0.0, port=6969, sniHostCheck=false, user=policyadmin, password=zb!XztG34, contextPath=/, jettyServer=Server@72c927f1{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@1ac85b0c{/,null,STOPPED}, connector=RestServerParameters@63c5efee{HTTP/1.1, (http/1.1)}{0.0.0.0:6969}, jettyThread=null, servlets={/metrics=io.prometheus.client.servlet.jakarta.exporter.MetricsServlet-72b16078==io.prometheus.client.servlet.jakarta.exporter.MetricsServlet@aa16c20f{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:null,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-d78795==org.glassfish.jersey.servlet.ServletContainer@b1764d3c{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.508+00:00|INFO|ServiceManager|main] service manager started 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.508+00:00|INFO|ServiceManager|main] service manager started 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.508+00:00|INFO|ApexStarterMain|main] Started policy-apex-pdp service successfully. 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.508+00:00|INFO|JettyServletServer|RestServerParameters-6969] JettyJerseyServer [Jerseyservlets={/metrics=io.prometheus.client.servlet.jakarta.exporter.MetricsServlet-72b16078==io.prometheus.client.servlet.jakarta.exporter.MetricsServlet@aa16c20f{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:null,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-d78795==org.glassfish.jersey.servlet.ServletContainer@b1764d3c{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=swagger-6969, toString()=JettyServletServer(name=RestServerParameters, host=0.0.0.0, port=6969, sniHostCheck=false, user=policyadmin, password=zb!XztG34, contextPath=/, jettyServer=Server@72c927f1{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@1ac85b0c{/,null,STOPPED}, connector=RestServerParameters@63c5efee{HTTP/1.1, (http/1.1)}{0.0.0.0:6969}, jettyThread=Thread[RestServerParameters-6969,5,main], servlets={/metrics=io.prometheus.client.servlet.jakarta.exporter.MetricsServlet-72b16078==io.prometheus.client.servlet.jakarta.exporter.MetricsServlet@aa16c20f{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:null,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-d78795==org.glassfish.jersey.servlet.ServletContainer@b1764d3c{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.666+00:00|INFO|Metadata|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] Cluster ID: 5fWDTH-mSmOqNb5HKZw_Tw 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.667+00:00|INFO|TransactionManager|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] ProducerId set to 2 with epoch 0 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.666+00:00|INFO|Metadata|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Cluster ID: 5fWDTH-mSmOqNb5HKZw_Tw 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.678+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Discovered group coordinator kafka:9092 (id: 2147483646 rack: null) 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.690+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] (Re-)joining group 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.706+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Request joining group due to: need to re-join with the given member-id: consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2-198c82d0-4714-4755-bc78-bfca51e47337 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.707+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) 09:44:48 policy-apex-pdp | [2024-07-04T09:42:39.707+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] (Re-)joining group 09:44:48 policy-apex-pdp | [2024-07-04T09:42:40.066+00:00|INFO|GsonMessageBodyHandler|RestServerParameters-6969] Using GSON for REST calls 09:44:48 policy-apex-pdp | [2024-07-04T09:42:40.067+00:00|INFO|YamlMessageBodyHandler|RestServerParameters-6969] Accepting YAML for REST calls 09:44:48 policy-apex-pdp | [2024-07-04T09:42:41.143+00:00|INFO|RequestLog|qtp739264372-33] 172.17.0.1 - - [04/Jul/2024:09:42:41 +0000] "GET / HTTP/1.1" 401 495 "-" "curl/7.58.0" 09:44:48 policy-apex-pdp | [2024-07-04T09:42:42.714+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Successfully joined group with generation Generation{generationId=1, memberId='consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2-198c82d0-4714-4755-bc78-bfca51e47337', protocol='range'} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:42.720+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Finished assignment for group at generation 1: {consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2-198c82d0-4714-4755-bc78-bfca51e47337=Assignment(partitions=[policy-pdp-pap-0])} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:42.727+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Successfully synced group in generation Generation{generationId=1, memberId='consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2-198c82d0-4714-4755-bc78-bfca51e47337', protocol='range'} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:42.728+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Notifying assignor about the new Assignment(partitions=[policy-pdp-pap-0]) 09:44:48 policy-apex-pdp | [2024-07-04T09:42:42.729+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Adding newly assigned partitions: policy-pdp-pap-0 09:44:48 policy-apex-pdp | [2024-07-04T09:42:42.737+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Found no committed offset for partition policy-pdp-pap-0 09:44:48 policy-apex-pdp | [2024-07-04T09:42:42.745+00:00|INFO|SubscriptionState|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-43992874-9c1b-4e05-b493-a9684c6c5296-2, groupId=43992874-9c1b-4e05-b493-a9684c6c5296] Resetting offset for partition policy-pdp-pap-0 to position FetchPosition{offset=1, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:9092 (id: 1 rack: null)], epoch=0}}. 09:44:48 policy-apex-pdp | [2024-07-04T09:42:56.092+00:00|INFO|RequestLog|qtp739264372-30] 172.17.0.4 - policyadmin [04/Jul/2024:09:42:56 +0000] "GET /metrics HTTP/1.1" 200 10652 "-" "Prometheus/2.53.0" 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.313+00:00|INFO|network|Timer-0] [OUT|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"908a25c1-f405-4bad-b634-83eb08f9182d","timestampMs":1720086179313,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.334+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"908a25c1-f405-4bad-b634-83eb08f9182d","timestampMs":1720086179313,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.336+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.502+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"messageName":"PDP_UPDATE","requestId":"83f05f71-f951-4e81-8af2-344adb821102","timestampMs":1720086179433,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.511+00:00|WARN|Registry|KAFKA-source-policy-pdp-pap] replacing previously registered: object:pdp/status/publisher 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.511+00:00|INFO|network|Timer-1] [OUT|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"03bc4065-c817-4394-9c0d-26791ee3b6ff","timestampMs":1720086179510,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.512+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"83f05f71-f951-4e81-8af2-344adb821102","responseStatus":"SUCCESS","responseMessage":"Pdp update successful."},"messageName":"PDP_STATUS","requestId":"8cf42592-1b3a-46c7-8ffb-56547c206e64","timestampMs":1720086179511,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.521+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"03bc4065-c817-4394-9c0d-26791ee3b6ff","timestampMs":1720086179510,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.522+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.526+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"83f05f71-f951-4e81-8af2-344adb821102","responseStatus":"SUCCESS","responseMessage":"Pdp update successful."},"messageName":"PDP_STATUS","requestId":"8cf42592-1b3a-46c7-8ffb-56547c206e64","timestampMs":1720086179511,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.527+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.572+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"bc1c9506-3f24-4c95-b519-1c7dbca5b7b0","timestampMs":1720086179434,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.574+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpStateChange","policies":[],"response":{"responseTo":"bc1c9506-3f24-4c95-b519-1c7dbca5b7b0","responseStatus":"SUCCESS","responseMessage":"State changed to active. No policies found."},"messageName":"PDP_STATUS","requestId":"c6eccbf2-6525-456b-98c1-d9ba1f8088ea","timestampMs":1720086179574,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.583+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpStateChange","policies":[],"response":{"responseTo":"bc1c9506-3f24-4c95-b519-1c7dbca5b7b0","responseStatus":"SUCCESS","responseMessage":"State changed to active. No policies found."},"messageName":"PDP_STATUS","requestId":"c6eccbf2-6525-456b-98c1-d9ba1f8088ea","timestampMs":1720086179574,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.584+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.615+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"68fa5061-72d2-4f2d-ba41-4a7727b3c604","timestampMs":1720086179596,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.617+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"68fa5061-72d2-4f2d-ba41-4a7727b3c604","responseStatus":"SUCCESS","responseMessage":"Pdp already updated"},"messageName":"PDP_STATUS","requestId":"c6d6d97d-4b4b-476f-966f-02ef731a732f","timestampMs":1720086179616,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.625+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-apex-pdp | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"68fa5061-72d2-4f2d-ba41-4a7727b3c604","responseStatus":"SUCCESS","responseMessage":"Pdp already updated"},"messageName":"PDP_STATUS","requestId":"c6d6d97d-4b4b-476f-966f-02ef731a732f","timestampMs":1720086179616,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-apex-pdp | [2024-07-04T09:42:59.626+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS 09:44:48 policy-apex-pdp | [2024-07-04T09:43:01.187+00:00|INFO|RequestLog|qtp739264372-31] 172.17.0.1 - policyadmin [04/Jul/2024:09:43:01 +0000] "GET /policy/apex-pdp/v1/healthcheck HTTP/1.1" 200 109 "-" "curl/7.58.0" 09:44:48 policy-apex-pdp | [2024-07-04T09:43:38.924+00:00|INFO|RequestLog|qtp739264372-26] 172.17.0.7 - policyadmin [04/Jul/2024:09:43:38 +0000] "GET /policy/apex-pdp/v1/healthcheck?null HTTP/1.1" 200 109 "-" "python-requests/2.32.3" 09:44:48 policy-apex-pdp | [2024-07-04T09:43:40.709+00:00|INFO|RequestLog|qtp739264372-27] 172.17.0.7 - policyadmin [04/Jul/2024:09:43:40 +0000] "GET /metrics?null HTTP/1.1" 200 11011 "-" "python-requests/2.32.3" 09:44:48 policy-apex-pdp | [2024-07-04T09:43:40.784+00:00|INFO|RequestLog|qtp739264372-30] 172.17.0.7 - policyadmin [04/Jul/2024:09:43:40 +0000] "GET /policy/apex-pdp/v1/healthcheck?null HTTP/1.1" 200 109 "-" "python-requests/2.32.3" 09:44:48 policy-apex-pdp | [2024-07-04T09:43:56.081+00:00|INFO|RequestLog|qtp739264372-33] 172.17.0.4 - policyadmin [04/Jul/2024:09:43:56 +0000] "GET /metrics HTTP/1.1" 200 10654 "-" "Prometheus/2.53.0" 09:44:48 =================================== 09:44:48 ======== Logs from api ======== 09:44:48 policy-api | Waiting for mariadb port 3306... 09:44:48 policy-api | mariadb (172.17.0.3:3306) open 09:44:48 policy-api | Waiting for policy-db-migrator port 6824... 09:44:48 policy-api | policy-db-migrator (172.17.0.7:6824) open 09:44:48 policy-api | Policy api config file: /opt/app/policy/api/etc/apiParameters.yaml 09:44:48 policy-api | 09:44:48 policy-api | . ____ _ __ _ _ 09:44:48 policy-api | /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \ 09:44:48 policy-api | ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \ 09:44:48 policy-api | \\/ ___)| |_)| | | | | || (_| | ) ) ) ) 09:44:48 policy-api | ' |____| .__|_| |_|_| |_\__, | / / / / 09:44:48 policy-api | =========|_|==============|___/=/_/_/_/ 09:44:48 policy-api | :: Spring Boot :: (v3.1.10) 09:44:48 policy-api | 09:44:48 policy-api | [2024-07-04T09:42:15.275+00:00|INFO|Version|background-preinit] HV000001: Hibernate Validator 8.0.1.Final 09:44:48 policy-api | [2024-07-04T09:42:15.355+00:00|INFO|PolicyApiApplication|main] Starting PolicyApiApplication using Java 17.0.11 with PID 29 (/app/api.jar started by policy in /opt/app/policy/api/bin) 09:44:48 policy-api | [2024-07-04T09:42:15.356+00:00|INFO|PolicyApiApplication|main] No active profile set, falling back to 1 default profile: "default" 09:44:48 policy-api | [2024-07-04T09:42:17.337+00:00|INFO|RepositoryConfigurationDelegate|main] Bootstrapping Spring Data JPA repositories in DEFAULT mode. 09:44:48 policy-api | [2024-07-04T09:42:17.573+00:00|INFO|RepositoryConfigurationDelegate|main] Finished Spring Data repository scanning in 226 ms. Found 6 JPA repository interfaces. 09:44:48 policy-api | [2024-07-04T09:42:18.416+00:00|INFO|TomcatWebServer|main] Tomcat initialized with port(s): 6969 (http) 09:44:48 policy-api | [2024-07-04T09:42:18.426+00:00|INFO|Http11NioProtocol|main] Initializing ProtocolHandler ["http-nio-6969"] 09:44:48 policy-api | [2024-07-04T09:42:18.428+00:00|INFO|StandardService|main] Starting service [Tomcat] 09:44:48 policy-api | [2024-07-04T09:42:18.428+00:00|INFO|StandardEngine|main] Starting Servlet engine: [Apache Tomcat/10.1.19] 09:44:48 policy-api | [2024-07-04T09:42:18.528+00:00|INFO|[/policy/api/v1]|main] Initializing Spring embedded WebApplicationContext 09:44:48 policy-api | [2024-07-04T09:42:18.528+00:00|INFO|ServletWebServerApplicationContext|main] Root WebApplicationContext: initialization completed in 3105 ms 09:44:48 policy-api | [2024-07-04T09:42:18.886+00:00|INFO|LogHelper|main] HHH000204: Processing PersistenceUnitInfo [name: default] 09:44:48 policy-api | [2024-07-04T09:42:18.967+00:00|INFO|Version|main] HHH000412: Hibernate ORM core version 6.3.2.Final 09:44:48 policy-api | [2024-07-04T09:42:19.079+00:00|INFO|RegionFactoryInitiator|main] HHH000026: Second-level cache disabled 09:44:48 policy-api | [2024-07-04T09:42:19.382+00:00|INFO|SpringPersistenceUnitInfo|main] No LoadTimeWeaver setup: ignoring JPA class transformer 09:44:48 policy-api | [2024-07-04T09:42:19.414+00:00|INFO|HikariDataSource|main] HikariPool-1 - Starting... 09:44:48 policy-api | [2024-07-04T09:42:19.511+00:00|INFO|HikariPool|main] HikariPool-1 - Added connection org.mariadb.jdbc.Connection@320a8ebf 09:44:48 policy-api | [2024-07-04T09:42:19.514+00:00|INFO|HikariDataSource|main] HikariPool-1 - Start completed. 09:44:48 policy-api | [2024-07-04T09:42:21.556+00:00|INFO|JtaPlatformInitiator|main] HHH000489: No JTA platform available (set 'hibernate.transaction.jta.platform' to enable JTA platform integration) 09:44:48 policy-api | [2024-07-04T09:42:21.561+00:00|INFO|LocalContainerEntityManagerFactoryBean|main] Initialized JPA EntityManagerFactory for persistence unit 'default' 09:44:48 policy-api | [2024-07-04T09:42:22.326+00:00|WARN|ApiDatabaseInitializer|main] Detected multi-versioned type: policytypes/onap.policies.monitoring.tcagen2.v2.yaml 09:44:48 policy-api | [2024-07-04T09:42:23.300+00:00|INFO|ApiDatabaseInitializer|main] Multi-versioned Service Template [onap.policies.Monitoring, onap.policies.monitoring.tcagen2] 09:44:48 policy-api | [2024-07-04T09:42:24.542+00:00|WARN|JpaBaseConfiguration$JpaWebConfiguration|main] spring.jpa.open-in-view is enabled by default. Therefore, database queries may be performed during view rendering. Explicitly configure spring.jpa.open-in-view to disable this warning 09:44:48 policy-api | [2024-07-04T09:42:24.731+00:00|INFO|DefaultSecurityFilterChain|main] Will secure any request with [org.springframework.security.web.session.DisableEncodeUrlFilter@19ad99fa, org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter@5331be15, org.springframework.security.web.context.SecurityContextHolderFilter@6f3628c2, org.springframework.security.web.header.HeaderWriterFilter@7f1f6c22, org.springframework.security.web.authentication.logout.LogoutFilter@cf97391, org.springframework.security.web.authentication.www.BasicAuthenticationFilter@46fdb4d, org.springframework.security.web.savedrequest.RequestCacheAwareFilter@321f97d9, org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter@64e06cf0, org.springframework.security.web.authentication.AnonymousAuthenticationFilter@32a53a59, org.springframework.security.web.access.ExceptionTranslationFilter@7111b312, org.springframework.security.web.access.intercept.AuthorizationFilter@73c72238] 09:44:48 policy-api | [2024-07-04T09:42:25.458+00:00|INFO|EndpointLinksResolver|main] Exposing 3 endpoint(s) beneath base path '' 09:44:48 policy-api | [2024-07-04T09:42:25.540+00:00|INFO|Http11NioProtocol|main] Starting ProtocolHandler ["http-nio-6969"] 09:44:48 policy-api | [2024-07-04T09:42:25.572+00:00|INFO|TomcatWebServer|main] Tomcat started on port(s): 6969 (http) with context path '/policy/api/v1' 09:44:48 policy-api | [2024-07-04T09:42:25.594+00:00|INFO|PolicyApiApplication|main] Started PolicyApiApplication in 11.041 seconds (process running for 11.694) 09:44:48 policy-api | [2024-07-04T09:42:39.915+00:00|INFO|[/policy/api/v1]|http-nio-6969-exec-2] Initializing Spring DispatcherServlet 'dispatcherServlet' 09:44:48 policy-api | [2024-07-04T09:42:39.915+00:00|INFO|DispatcherServlet|http-nio-6969-exec-2] Initializing Servlet 'dispatcherServlet' 09:44:48 policy-api | [2024-07-04T09:42:39.917+00:00|INFO|DispatcherServlet|http-nio-6969-exec-2] Completed initialization in 1 ms 09:44:48 policy-api | [2024-07-04T09:43:39.121+00:00|INFO|OrderedServiceImpl|http-nio-6969-exec-3] ***** OrderedServiceImpl implementers: 09:44:48 policy-api | [] 09:44:48 =================================== 09:44:48 ======== Logs from csit-tests ======== 09:44:48 policy-csit | Invoking the robot tests from: apex-pdp-test.robot apex-slas.robot 09:44:48 policy-csit | Run Robot test 09:44:48 policy-csit | ROBOT_VARIABLES=-v DATA:/opt/robotworkspace/models/models-examples/src/main/resources/policies 09:44:48 policy-csit | -v NODETEMPLATES:/opt/robotworkspace/models/models-examples/src/main/resources/nodetemplates 09:44:48 policy-csit | -v POLICY_API_IP:policy-api:6969 09:44:48 policy-csit | -v POLICY_RUNTIME_ACM_IP:policy-clamp-runtime-acm:6969 09:44:48 policy-csit | -v POLICY_PARTICIPANT_SIM_IP:policy-clamp-ac-sim-ppnt:6969 09:44:48 policy-csit | -v POLICY_PAP_IP:policy-pap:6969 09:44:48 policy-csit | -v APEX_IP:policy-apex-pdp:6969 09:44:48 policy-csit | -v APEX_EVENTS_IP:policy-apex-pdp:23324 09:44:48 policy-csit | -v KAFKA_IP:kafka:9092 09:44:48 policy-csit | -v PROMETHEUS_IP:prometheus:9090 09:44:48 policy-csit | -v POLICY_PDPX_IP:policy-xacml-pdp:6969 09:44:48 policy-csit | -v POLICY_DROOLS_IP:policy-drools-pdp:9696 09:44:48 policy-csit | -v DROOLS_IP:policy-drools-apps:6969 09:44:48 policy-csit | -v DROOLS_IP_2:policy-drools-apps:9696 09:44:48 policy-csit | -v TEMP_FOLDER:/tmp/distribution 09:44:48 policy-csit | -v DISTRIBUTION_IP:policy-distribution:6969 09:44:48 policy-csit | -v TEST_ENV: 09:44:48 policy-csit | -v JAEGER_IP:jaeger:16686 09:44:48 policy-csit | Starting Robot test suites ... 09:44:48 policy-csit | ============================================================================== 09:44:48 policy-csit | Apex-Pdp-Test & Apex-Slas 09:44:48 policy-csit | ============================================================================== 09:44:48 policy-csit | Apex-Pdp-Test & Apex-Slas.Apex-Pdp-Test 09:44:48 policy-csit | ============================================================================== 09:44:48 policy-csit | Healthcheck :: Runs Apex PDP Health check | PASS | 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | ExecuteApexSampleDomainPolicy | FAIL | 09:44:48 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | ExecuteApexTestPnfPolicy | FAIL | 09:44:48 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | ExecuteApexTestPnfPolicyWithMetadataSet | FAIL | 09:44:48 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | Metrics :: Verify policy-apex-pdp is exporting prometheus metrics | FAIL | 09:44:48 policy-csit | '# HELP jvm_classes_currently_loaded The number of classes that are currently loaded in the JVM 09:44:48 policy-csit | # TYPE jvm_classes_currently_loaded gauge 09:44:48 policy-csit | jvm_classes_currently_loaded 7533.0 09:44:48 policy-csit | # HELP jvm_classes_loaded_total The total number of classes that have been loaded since the JVM has started execution 09:44:48 policy-csit | # TYPE jvm_classes_loaded_total counter 09:44:48 policy-csit | jvm_classes_loaded_total 7533.0 09:44:48 policy-csit | # HELP jvm_classes_unloaded_total The total number of classes that have been unloaded since the JVM has started execution 09:44:48 policy-csit | # TYPE jvm_classes_unloaded_total counter 09:44:48 policy-csit | jvm_classes_unloaded_total 0.0 09:44:48 policy-csit | # HELP process_cpu_seconds_total Total user and system CPU time spent in seconds. 09:44:48 policy-csit | # TYPE process_cpu_seconds_total counter 09:44:48 policy-csit | process_cpu_seconds_total 7.72 09:44:48 policy-csit | # HELP process_start_time_seconds Start time of the process since unix epoch in seconds. 09:44:48 policy-csit | # TYPE process_start_time_seconds gauge 09:44:48 policy-csit | process_start_time_seconds 1.720086157998E9 09:44:48 policy-csit | [ Message content over the limit has been removed. ] 09:44:48 policy-csit | # TYPE pdpa_policy_deployments_total counter 09:44:48 policy-csit | # HELP jvm_memory_pool_allocated_bytes_created Total bytes allocated in a given JVM memory pool. Only updated after GC, not continuously. 09:44:48 policy-csit | # TYPE jvm_memory_pool_allocated_bytes_created gauge 09:44:48 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="CodeHeap 'profiled nmethods'",} 1.720086159595E9 09:44:48 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="G1 Old Gen",} 1.720086159613E9 09:44:48 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="G1 Eden Space",} 1.720086159613E9 09:44:48 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="CodeHeap 'non-profiled nmethods'",} 1.720086159613E9 09:44:48 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="G1 Survivor Space",} 1.720086159613E9 09:44:48 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="Compressed Class Space",} 1.720086159613E9 09:44:48 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="Metaspace",} 1.720086159613E9 09:44:48 policy-csit | jvm_memory_pool_allocated_bytes_created{pool="CodeHeap 'non-nmethods'",} 1.720086159613E9 09:44:48 policy-csit | ' does not contain 'pdpa_policy_deployments_total{operation="deploy",status="TOTAL",} 3.0' 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | Apex-Pdp-Test & Apex-Slas.Apex-Pdp-Test | FAIL | 09:44:48 policy-csit | 5 tests, 1 passed, 4 failed 09:44:48 policy-csit | ============================================================================== 09:44:48 policy-csit | Apex-Pdp-Test & Apex-Slas.Apex-Slas 09:44:48 policy-csit | ============================================================================== 09:44:48 policy-csit | Healthcheck :: Runs Apex PDP Health check | PASS | 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | ValidatePolicyExecutionAndEventRateLowComplexity :: Validate that ... | FAIL | 09:44:48 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | ValidatePolicyExecutionAndEventRateModerateComplexity :: Validate ... | FAIL | 09:44:48 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | ValidatePolicyExecutionAndEventRateHighComplexity :: Validate that... | FAIL | 09:44:48 policy-csit | Url: http://policy-api:6969/policy/api/v1/policytypes/onap.policies.native.Apex/versions/1.0.0/policies?null Expected status: 201 != 200 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | WaitForPrometheusServer :: Sleep time to wait for Prometheus serve... | PASS | 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | ValidatePolicyExecutionTimes :: Validate policy execution times us... | FAIL | 09:44:48 policy-csit | Resolving variable '${resp['data']['result'][0]['value'][1]}' failed: IndexError: list index out of range 09:44:48 policy-csit | ------------------------------------------------------------------------------ 09:44:48 policy-csit | Apex-Pdp-Test & Apex-Slas.Apex-Slas | FAIL | 09:44:48 policy-csit | 6 tests, 2 passed, 4 failed 09:44:48 policy-csit | ============================================================================== 09:44:48 policy-csit | Apex-Pdp-Test & Apex-Slas | FAIL | 09:44:48 policy-csit | 11 tests, 3 passed, 8 failed 09:44:48 policy-csit | ============================================================================== 09:44:48 policy-csit | Output: /tmp/results/output.xml 09:44:48 policy-csit | Log: /tmp/results/log.html 09:44:48 policy-csit | Report: /tmp/results/report.html 09:44:48 policy-csit | RESULT: 8 09:44:48 =================================== 09:44:48 ======== Logs from policy-db-migrator ======== 09:44:48 policy-db-migrator | Waiting for mariadb port 3306... 09:44:48 policy-db-migrator | nc: connect to mariadb (172.17.0.3) port 3306 (tcp) failed: Connection refused 09:44:48 policy-db-migrator | nc: connect to mariadb (172.17.0.3) port 3306 (tcp) failed: Connection refused 09:44:48 policy-db-migrator | nc: connect to mariadb (172.17.0.3) port 3306 (tcp) failed: Connection refused 09:44:48 policy-db-migrator | nc: connect to mariadb (172.17.0.3) port 3306 (tcp) failed: Connection refused 09:44:48 policy-db-migrator | nc: connect to mariadb (172.17.0.3) port 3306 (tcp) failed: Connection refused 09:44:48 policy-db-migrator | nc: connect to mariadb (172.17.0.3) port 3306 (tcp) failed: Connection refused 09:44:48 policy-db-migrator | nc: connect to mariadb (172.17.0.3) port 3306 (tcp) failed: Connection refused 09:44:48 policy-db-migrator | nc: connect to mariadb (172.17.0.3) port 3306 (tcp) failed: Connection refused 09:44:48 policy-db-migrator | Connection to mariadb (172.17.0.3) 3306 port [tcp/mysql] succeeded! 09:44:48 policy-db-migrator | 321 blocks 09:44:48 policy-db-migrator | Preparing upgrade release version: 0800 09:44:48 policy-db-migrator | Preparing upgrade release version: 0900 09:44:48 policy-db-migrator | Preparing upgrade release version: 1000 09:44:48 policy-db-migrator | Preparing upgrade release version: 1100 09:44:48 policy-db-migrator | Preparing upgrade release version: 1200 09:44:48 policy-db-migrator | Preparing upgrade release version: 1300 09:44:48 policy-db-migrator | Done 09:44:48 policy-db-migrator | name version 09:44:48 policy-db-migrator | policyadmin 0 09:44:48 policy-db-migrator | policyadmin: upgrade available: 0 -> 1300 09:44:48 policy-db-migrator | upgrade: 0 -> 1300 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0100-jpapdpgroup_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpgroup_properties (name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, PROPERTIES VARCHAR(255) DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) DEFAULT NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0110-jpapdpstatistics_enginestats.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpstatistics_enginestats (AVERAGEEXECUTIONTIME DOUBLE DEFAULT NULL, ENGINEID VARCHAR(255) DEFAULT NULL, ENGINETIMESTAMP BIGINT DEFAULT NULL, ENGINEWORKERSTATE INT DEFAULT NULL, EVENTCOUNT BIGINT DEFAULT NULL, LASTENTERTIME BIGINT DEFAULT NULL, LASTEXECUTIONTIME BIGINT DEFAULT NULL, LASTSTART BIGINT DEFAULT NULL, UPTIME BIGINT DEFAULT NULL, timeStamp datetime DEFAULT NULL, name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0120-jpapdpsubgroup_policies.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpsubgroup_policies (name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, parentLocalName VARCHAR(120) DEFAULT NULL, localName VARCHAR(120) DEFAULT NULL, parentKeyVersion VARCHAR(15) DEFAULT NULL, parentKeyName VARCHAR(120) DEFAULT NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0130-jpapdpsubgroup_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpsubgroup_properties (parentLocalName VARCHAR(120) DEFAULT NULL, localName VARCHAR(120) DEFAULT NULL, parentKeyVersion VARCHAR(15) DEFAULT NULL, parentKeyName VARCHAR(120) DEFAULT NULL, PROPERTIES VARCHAR(255) DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) DEFAULT NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0140-jpapdpsubgroup_supportedpolicytypes.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpsubgroup_supportedpolicytypes (name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, parentLocalName VARCHAR(120) DEFAULT NULL, localName VARCHAR(120) DEFAULT NULL, parentKeyVersion VARCHAR(15) DEFAULT NULL, parentKeyName VARCHAR(120) DEFAULT NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0150-jpatoscacapabilityassignment_attributes.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilityassignment_attributes (name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, ATTRIBUTES LONGTEXT DEFAULT NULL, ATTRIBUTES_KEY VARCHAR(255) DEFAULT NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0160-jpatoscacapabilityassignment_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilityassignment_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0170-jpatoscacapabilityassignment_occurrences.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilityassignment_occurrences (name VARCHAR(120) NULL, version VARCHAR(20) NULL, OCCURRENCES INT DEFAULT NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0180-jpatoscacapabilityassignment_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilityassignment_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGTEXT NULL, PROPERTIES_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0190-jpatoscacapabilitytype_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilitytype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0200-jpatoscacapabilitytype_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilitytype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0210-jpatoscadatatype_constraints.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscadatatype_constraints (name VARCHAR(120) NULL, version VARCHAR(20) NULL, CONSTRAINTS VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0220-jpatoscadatatype_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscadatatype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0230-jpatoscadatatype_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscadatatype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0240-jpatoscanodetemplate_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscanodetemplate_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0250-jpatoscanodetemplate_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscanodetemplate_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGTEXT NULL, PROPERTIES_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0260-jpatoscanodetype_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscanodetype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0270-jpatoscanodetype_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscanodetype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0280-jpatoscapolicy_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicy_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0290-jpatoscapolicy_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicy_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGTEXT NULL, PROPERTIES_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0300-jpatoscapolicy_targets.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicy_targets (name VARCHAR(120) NULL, version VARCHAR(20) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0310-jpatoscapolicytype_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicytype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0320-jpatoscapolicytype_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicytype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0330-jpatoscapolicytype_targets.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicytype_targets (name VARCHAR(120) NULL, version VARCHAR(20) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0340-jpatoscapolicytype_triggers.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicytype_triggers (name VARCHAR(120) NULL, version VARCHAR(20) NULL, TRIGGERS VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0350-jpatoscaproperty_constraints.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscaproperty_constraints (parentLocalName VARCHAR(120) NULL, localName VARCHAR(120) NULL, parentKeyVersion VARCHAR(15) NULL, parentKeyName VARCHAR(120) NULL, CONSTRAINTS VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0360-jpatoscaproperty_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscaproperty_metadata (parentLocalName VARCHAR(120) NULL, localName VARCHAR(120) NULL, parentKeyVersion VARCHAR(15) NULL, parentKeyName VARCHAR(120) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0370-jpatoscarelationshiptype_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarelationshiptype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0380-jpatoscarelationshiptype_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarelationshiptype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0390-jpatoscarequirement_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarequirement_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0400-jpatoscarequirement_occurrences.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarequirement_occurrences (name VARCHAR(120) NULL, version VARCHAR(20) NULL, OCCURRENCES INT DEFAULT NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0410-jpatoscarequirement_properties.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarequirement_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGTEXT NULL, PROPERTIES_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0420-jpatoscaservicetemplate_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscaservicetemplate_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0430-jpatoscatopologytemplate_inputs.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscatopologytemplate_inputs (parentLocalName VARCHAR(120) NULL, localName VARCHAR(120) NULL, parentKeyVersion VARCHAR(15) NULL, parentKeyName VARCHAR(120) NULL, INPUTS LONGBLOB DEFAULT NULL, INPUTS_KEY VARCHAR(255) NULL) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0440-pdpgroup_pdpsubgroup.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpgroup_pdpsubgroup (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_PDPGROUP_PDPSUBGROUP (name, version, parentLocalName, localName, parentKeyVersion, parentKeyName)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0450-pdpgroup.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpgroup (`DESCRIPTION` VARCHAR(255) NULL, PDPGROUPSTATE INT DEFAULT NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_PDPGROUP (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0460-pdppolicystatus.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS pdppolicystatus (DEPLOY BOOLEAN DEFAULT 0, PDPGROUP VARCHAR(255) DEFAULT NULL, PDPTYPE VARCHAR(255) DEFAULT NULL, STATE INT DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, PRIMARY KEY PK_PDPPOLICYSTATUS (parentLocalName, localName, parentKeyVersion, parentKeyName)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0470-pdp.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS pdp (HEALTHY INT DEFAULT NULL, MESSAGE VARCHAR(255) DEFAULT NULL, PDPSTATE INT DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_PDP (parentLocalName, localName, parentKeyVersion, parentKeyName)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0480-pdpstatistics.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpstatistics (PDPGROUPNAME VARCHAR(120) NULL, PDPSUBGROUPNAME VARCHAR(120) NULL, POLICYDEPLOYCOUNT BIGINT DEFAULT NULL, POLICYDEPLOYFAILCOUNT BIGINT DEFAULT NULL, POLICYDEPLOYSUCCESSCOUNT BIGINT DEFAULT NULL, POLICYEXECUTEDCOUNT BIGINT DEFAULT NULL, POLICYEXECUTEDFAILCOUNT BIGINT DEFAULT NULL, POLICYEXECUTEDSUCCESSCOUNT BIGINT DEFAULT NULL, timeStamp datetime NOT NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_PDPSTATISTICS (timeStamp, name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0490-pdpsubgroup_pdp.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpsubgroup_pdp (pdpParentKeyName VARCHAR(120) NOT NULL, pdpParentKeyVersion VARCHAR(15) NOT NULL, pdpParentLocalName VARCHAR(120) NOT NULL, pdpLocalName VARCHAR(120) NOT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_PDPSUBGROUP_PDP (pdpParentKeyName, pdpParentKeyVersion, pdpParentLocalName, pdpLocalName, parentLocalName, localName, parentKeyVersion, parentKeyName)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0500-pdpsubgroup.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpsubgroup (CURRENTINSTANCECOUNT INT DEFAULT NULL, DESIREDINSTANCECOUNT INT DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_PDPSUBGROUP (parentLocalName, localName, parentKeyVersion, parentKeyName)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0510-toscacapabilityassignment.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilityassignment (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, type_name VARCHAR(255) NULL, type_version VARCHAR(255) NULL, PRIMARY KEY PK_TOSCACAPABILITYASSIGNMENT(name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0520-toscacapabilityassignments.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilityassignments (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCACAPABILITYASSIGNMENTS (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0530-toscacapabilityassignments_toscacapabilityassignment.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilityassignments_toscacapabilityassignment (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCACAPABILITYASSIGNMENTS_TOSCACAPABILITYASSIGNMENT (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0540-toscacapabilitytype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilitytype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCACAPABILITYTYPE (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0550-toscacapabilitytypes.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilitytypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCACAPABILITYTYPES (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0560-toscacapabilitytypes_toscacapabilitytype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilitytypes_toscacapabilitytype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCACAPABILITYTYPES_TOSCACAPABILITYTYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0570-toscadatatype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscadatatype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCADATATYPE (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0580-toscadatatypes.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscadatatypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCADATATYPES (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0590-toscadatatypes_toscadatatype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscadatatypes_toscadatatype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCADATATYPES_TOSCADATATYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0600-toscanodetemplate.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetemplate (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, type_name VARCHAR(255) NULL, type_version VARCHAR(255) NULL, capabilitiesName VARCHAR(120) NULL, capabilitiesVersion VARCHAR(20) NULL, requirementsVersion VARCHAR(20) NULL, requirementsName VARCHAR(120) NULL, PRIMARY KEY PK_TOSCANODETEMPLATE (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0610-toscanodetemplates.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetemplates (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCANODETEMPLATES (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0620-toscanodetemplates_toscanodetemplate.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetemplates_toscanodetemplate (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCANODETEMPLATES_TOSCANODETEMPLATE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0630-toscanodetype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, requirementsVersion VARCHAR(20) NULL, requirementsName VARCHAR(120) NULL, PRIMARY KEY PK_TOSCANODETYPE (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0640-toscanodetypes.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCANODETYPES (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0650-toscanodetypes_toscanodetype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetypes_toscanodetype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCANODETYPES_TOSCANODETYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0660-toscaparameter.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscaparameter (VALUE VARCHAR(255) NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCAPARAMETER (parentLocalName, localName, parentKeyVersion, parentKeyName)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0670-toscapolicies.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicies (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCAPOLICIES (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0680-toscapolicies_toscapolicy.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicies_toscapolicy (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCAPOLICIES_TOSCAPOLICY (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0690-toscapolicy.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicy (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, type_name VARCHAR(255) NULL, type_version VARCHAR(255) NULL, PRIMARY KEY PK_TOSCAPOLICY (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0700-toscapolicytype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicytype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCAPOLICYTYPE (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0710-toscapolicytypes.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicytypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCAPOLICYTYPES (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0720-toscapolicytypes_toscapolicytype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicytypes_toscapolicytype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCAPOLICYTYPES_TOSCAPOLICYTYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0730-toscaproperty.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscaproperty (DEFAULTVALUE VARCHAR(255) DEFAULT NULL, `DESCRIPTION` VARCHAR(255) DEFAULT NULL, ENTRYSCHEMA LONGBLOB DEFAULT NULL, REQUIRED BOOLEAN DEFAULT 0, STATUS INT DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, PRIMARY KEY PK_TOSCAPROPERTY (parentLocalName, localName, parentKeyVersion, parentKeyName)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0740-toscarelationshiptype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarelationshiptype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCARELATIONSHIPTYPE (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0750-toscarelationshiptypes.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarelationshiptypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCARELATIONSHIPTYPES (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0760-toscarelationshiptypes_toscarelationshiptype.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarelationshiptypes_toscarelationshiptype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCARELATIONSHIPTYPES_TOSCARELATIONSHIPTYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0770-toscarequirement.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarequirement (CAPABILITY VARCHAR(255) NULL, `DESCRIPTION` VARCHAR(255) NULL, NODE VARCHAR(255) NULL, RELATIONSHIP VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, type_name VARCHAR(255) NULL, type_version VARCHAR(255) NULL, PRIMARY KEY PK_TOSCAREQUIREMENT (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0780-toscarequirements.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarequirements (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCAREQUIREMENTS (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0790-toscarequirements_toscarequirement.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarequirements_toscarequirement (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCAREQUIREMENTS_TOSCAREQUIREMENT (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0800-toscaservicetemplate.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscaservicetemplate (`DESCRIPTION` VARCHAR(255) NULL, TOSCADEFINITIONSVERSION VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, capabilityTypesVersion VARCHAR(20) NULL, capabilityTypesName VARCHAR(120) NULL, dataTypesName VARCHAR(120) NULL, dataTypesVersion VARCHAR(20) NULL, nodeTypesVersion VARCHAR(20) NULL, nodeTypesName VARCHAR(120) NULL, policyTypesName VARCHAR(120) NULL, policyTypesVersion VARCHAR(20) NULL, relationshipTypesVersion VARCHAR(20) NULL, relationshipTypesName VARCHAR(120) NULL, topologyTemplateLocalName VARCHAR(120) NULL, topologyTemplateParentKeyName VARCHAR(120) NULL, topologyTemplateParentKeyVersion VARCHAR(15) NULL, topologyTemplateParentLocalName VARCHAR(120) NULL, PRIMARY KEY PK_TOSCASERVICETEMPLATE (name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0810-toscatopologytemplate.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscatopologytemplate (`description` VARCHAR(255) NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, nodeTemplatessVersion VARCHAR(20) NULL, nodeTemplatesName VARCHAR(120) NULL, policyVersion VARCHAR(20) NULL, policyName VARCHAR(120) NULL, PRIMARY KEY PK_TOSCATOPOLOGYTEMPLATE (parentLocalName, localName, parentKeyVersion, parentKeyName)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0820-toscatrigger.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscatrigger (ACTION VARCHAR(255) NULL, toscaCondition LONGBLOB DEFAULT NULL, toscaConstraint LONGBLOB DEFAULT NULL, `DESCRIPTION` VARCHAR(255) NULL, EVALUATIONS INT DEFAULT NULL, EVENTTYPE VARCHAR(255) NULL, METHOD VARCHAR(255) NULL, `PERIOD` LONGBLOB DEFAULT NULL, SCHEDULE LONGBLOB DEFAULT NULL, TARGETFILTER LONGBLOB DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_TOSCATRIGGER (parentLocalName, localName, parentKeyVersion, parentKeyName)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0830-FK_ToscaNodeTemplate_capabilitiesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaNodeTemplate_capabilitiesName ON toscanodetemplate(capabilitiesName, capabilitiesVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0840-FK_ToscaNodeTemplate_requirementsName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaNodeTemplate_requirementsName ON toscanodetemplate(requirementsName, requirementsVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0850-FK_ToscaNodeType_requirementsName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaNodeType_requirementsName ON toscanodetype(requirementsName, requirementsVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0860-FK_ToscaServiceTemplate_capabilityTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_capabilityTypesName ON toscaservicetemplate(capabilityTypesName, capabilityTypesVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0870-FK_ToscaServiceTemplate_dataTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_dataTypesName ON toscaservicetemplate(dataTypesName, dataTypesVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0880-FK_ToscaServiceTemplate_nodeTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_nodeTypesName ON toscaservicetemplate(nodeTypesName, nodeTypesVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0890-FK_ToscaServiceTemplate_policyTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_policyTypesName ON toscaservicetemplate(policyTypesName, policyTypesVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0900-FK_ToscaServiceTemplate_relationshipTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_relationshipTypesName ON toscaservicetemplate(relationshipTypesName, relationshipTypesVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0910-FK_ToscaTopologyTemplate_nodeTemplatesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaTopologyTemplate_nodeTemplatesName ON toscatopologytemplate(nodeTemplatesName, nodeTemplatessVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0920-FK_ToscaTopologyTemplate_policyName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX FK_ToscaTopologyTemplate_policyName ON toscatopologytemplate(policyName, policyVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0940-PdpPolicyStatus_PdpGroup.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX PdpPolicyStatus_PdpGroup ON pdppolicystatus(PDPGROUP) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0950-TscaServiceTemplatetopologyTemplateParentLocalName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX TscaServiceTemplatetopologyTemplateParentLocalName ON toscaservicetemplate(topologyTemplateParentLocalName, topologyTemplateLocalName, topologyTemplateParentKeyVersion, topologyTemplateParentKeyName) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0960-FK_ToscaNodeTemplate_capabilitiesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscanodetemplate ADD CONSTRAINT FK_ToscaNodeTemplate_capabilitiesName FOREIGN KEY (capabilitiesName, capabilitiesVersion) REFERENCES toscacapabilityassignments (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0970-FK_ToscaNodeTemplate_requirementsName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscanodetemplate ADD CONSTRAINT FK_ToscaNodeTemplate_requirementsName FOREIGN KEY (requirementsName, requirementsVersion) REFERENCES toscarequirements (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0980-FK_ToscaNodeType_requirementsName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscanodetype ADD CONSTRAINT FK_ToscaNodeType_requirementsName FOREIGN KEY (requirementsName, requirementsVersion) REFERENCES toscarequirements (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0990-FK_ToscaServiceTemplate_capabilityTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_capabilityTypesName FOREIGN KEY (capabilityTypesName, capabilityTypesVersion) REFERENCES toscacapabilitytypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 1000-FK_ToscaServiceTemplate_dataTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_dataTypesName FOREIGN KEY (dataTypesName, dataTypesVersion) REFERENCES toscadatatypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 1010-FK_ToscaServiceTemplate_nodeTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_nodeTypesName FOREIGN KEY (nodeTypesName, nodeTypesVersion) REFERENCES toscanodetypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 1020-FK_ToscaServiceTemplate_policyTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_policyTypesName FOREIGN KEY (policyTypesName, policyTypesVersion) REFERENCES toscapolicytypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 1030-FK_ToscaServiceTemplate_relationshipTypesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_relationshipTypesName FOREIGN KEY (relationshipTypesName, relationshipTypesVersion) REFERENCES toscarelationshiptypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 1040-FK_ToscaTopologyTemplate_nodeTemplatesName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscatopologytemplate ADD CONSTRAINT FK_ToscaTopologyTemplate_nodeTemplatesName FOREIGN KEY (nodeTemplatesName, nodeTemplatessVersion) REFERENCES toscanodetemplates (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 1050-FK_ToscaTopologyTemplate_policyName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscatopologytemplate ADD CONSTRAINT FK_ToscaTopologyTemplate_policyName FOREIGN KEY (policyName, policyVersion) REFERENCES toscapolicies (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 1060-TscaServiceTemplatetopologyTemplateParentLocalName.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT TscaServiceTemplatetopologyTemplateParentLocalName FOREIGN KEY (topologyTemplateParentLocalName, topologyTemplateLocalName, topologyTemplateParentKeyVersion, topologyTemplateParentKeyName) REFERENCES toscatopologytemplate (parentLocalName, localName, parentKeyVersion, parentKeyName) ON UPDATE RESTRICT ON DELETE RESTRICT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0100-pdp.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE pdp ADD COLUMN LASTUPDATE datetime NOT NULL DEFAULT CURRENT_TIMESTAMP AFTER HEALTHY 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0110-idx_tsidx1.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX IDX_TSIDX1 ON pdpstatistics(timeStamp, name, version) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0120-pk_pdpstatistics.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE pdpstatistics DROP PRIMARY KEY 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0130-pdpstatistics.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE pdpstatistics ADD COLUMN POLICYUNDEPLOYCOUNT BIGINT DEFAULT NULL AFTER POLICYEXECUTEDSUCCESSCOUNT, ADD COLUMN POLICYUNDEPLOYFAILCOUNT BIGINT DEFAULT NULL, ADD COLUMN POLICYUNDEPLOYSUCCESSCOUNT BIGINT DEFAULT NULL, ADD COLUMN ID BIGINT NOT NULL 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0140-pk_pdpstatistics.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | UPDATE pdpstatistics as p JOIN (SELECT name, version, timeStamp, ROW_NUMBER() OVER (ORDER BY timeStamp ASC) AS row_num FROM pdpstatistics GROUP BY name, version, timeStamp) AS t ON (p.name=t.name AND p.version=t.version AND p.timeStamp = t.timeStamp) SET p.id=t.row_num 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE pdpstatistics ADD CONSTRAINT PK_PDPSTATISTICS PRIMARY KEY (ID, name, version) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0150-pdpstatistics.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE pdpstatistics MODIFY COLUMN timeStamp datetime(6) NULL 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0160-jpapdpstatistics_enginestats.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE jpapdpstatistics_enginestats ADD COLUMN ID BIGINT DEFAULT NULL AFTER UPTIME 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0170-jpapdpstatistics_enginestats.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | UPDATE jpapdpstatistics_enginestats a 09:44:48 policy-db-migrator | JOIN pdpstatistics b 09:44:48 policy-db-migrator | ON a.name = b.name AND a.version = b.version AND a.timeStamp = b.timeStamp 09:44:48 policy-db-migrator | SET a.id = b.id 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0180-jpapdpstatistics_enginestats.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE jpapdpstatistics_enginestats DROP COLUMN timeStamp 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0190-jpapolicyaudit.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapolicyaudit (ACTION INT DEFAULT NULL, PDPGROUP VARCHAR(255) NULL, PDPTYPE VARCHAR(255) NULL, TIMESTAMP datetime DEFAULT NULL, USER VARCHAR(255) NULL, ID BIGINT NOT NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_JPAPOLICYAUDIT (ID, name, version)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0200-JpaPolicyAuditIndex_timestamp.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX JpaPolicyAuditIndex_timestamp ON jpapolicyaudit(TIMESTAMP) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0210-sequence.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS sequence (SEQ_NAME VARCHAR(50) NOT NULL, SEQ_COUNT DECIMAL(38) DEFAULT NULL, PRIMARY KEY PK_SEQUENCE (SEQ_NAME)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0220-sequence.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | INSERT INTO sequence(SEQ_NAME, SEQ_COUNT) VALUES('SEQ_GEN', (SELECT IFNULL(max(id),0) FROM pdpstatistics)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0100-jpatoscapolicy_targets.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE jpatoscapolicy_targets ADD COLUMN toscaPolicyName VARCHAR(120) NOT NULL, ADD COLUMN toscaPolicyVersion VARCHAR(20) NOT NULL, ADD CONSTRAINT PK_JPATOSCAPOLICY_TARGETS PRIMARY KEY (toscaPolicyName, toscaPolicyVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0110-jpatoscapolicytype_targets.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE jpatoscapolicytype_targets ADD COLUMN toscaPolicyTypeName VARCHAR(120) NOT NULL, ADD COLUMN toscaPolicyTypeVersion VARCHAR(20) NOT NULL, ADD CONSTRAINT PK_JPATOSCAPOLICYTYPE_TARGETS PRIMARY KEY (toscaPolicyTypeName, toscaPolicyTypeVersion) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0120-toscatrigger.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP TABLE IF EXISTS toscatrigger 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0130-jpatoscapolicytype_triggers.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE jpatoscapolicytype_triggers MODIFY COLUMN triggers LONGBLOB 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0140-toscaparameter.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP TABLE IF EXISTS toscaparameter 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0150-toscaproperty.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP TABLE IF EXISTS jpatoscaproperty_constraints 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP TABLE IF EXISTS jpatoscaproperty_metadata 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP TABLE IF EXISTS toscaproperty 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0160-jpapolicyaudit_pk.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE jpapolicyaudit DROP PRIMARY KEY 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE jpapolicyaudit ADD CONSTRAINT PK_JPAPOLICYAUDIT PRIMARY KEY (ID) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0170-pdpstatistics_pk.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE pdpstatistics DROP PRIMARY KEY 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE pdpstatistics ADD CONSTRAINT PK_PDPSTATISTICS PRIMARY KEY (ID) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0180-jpatoscanodetemplate_metadata.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE jpatoscanodetemplate_metadata MODIFY COLUMN METADATA LONGTEXT 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0100-upgrade.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | select 'upgrade to 1100 completed' as msg 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | msg 09:44:48 policy-db-migrator | upgrade to 1100 completed 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0100-jpapolicyaudit_renameuser.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | ALTER TABLE jpapolicyaudit RENAME COLUMN USER TO USERNAME 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0110-idx_tsidx1.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP INDEX IDX_TSIDX1 ON pdpstatistics 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE INDEX IDXTSIDX1 ON pdpstatistics(timeStamp, name, version) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0120-audit_sequence.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS audit_sequence (SEQ_NAME VARCHAR(50) NOT NULL, SEQ_COUNT DECIMAL(38) DEFAULT NULL, PRIMARY KEY PK_SEQUENCE (SEQ_NAME)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | INSERT INTO audit_sequence(SEQ_NAME, SEQ_COUNT) VALUES('SEQ_GEN', (SELECT IFNULL(max(id),0) FROM jpapolicyaudit)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0130-statistics_sequence.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | CREATE TABLE IF NOT EXISTS statistics_sequence (SEQ_NAME VARCHAR(50) NOT NULL, SEQ_COUNT DECIMAL(38) DEFAULT NULL, PRIMARY KEY PK_SEQUENCE (SEQ_NAME)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | INSERT INTO statistics_sequence(SEQ_NAME, SEQ_COUNT) VALUES('SEQ_GEN', (SELECT IFNULL(max(id),0) FROM pdpstatistics)) 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | TRUNCATE TABLE sequence 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0100-pdpstatistics.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP INDEX IDXTSIDX1 ON pdpstatistics 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP TABLE pdpstatistics 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0110-jpapdpstatistics_enginestats.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP TABLE jpapdpstatistics_enginestats 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | > upgrade 0120-statistics_sequence.sql 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | DROP TABLE statistics_sequence 09:44:48 policy-db-migrator | -------------- 09:44:48 policy-db-migrator | 09:44:48 policy-db-migrator | policyadmin: OK: upgrade (1300) 09:44:48 policy-db-migrator | name version 09:44:48 policy-db-migrator | policyadmin 1300 09:44:48 policy-db-migrator | ID script operation from_version to_version tag success atTime 09:44:48 policy-db-migrator | 1 0100-jpapdpgroup_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:02 09:44:48 policy-db-migrator | 2 0110-jpapdpstatistics_enginestats.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:02 09:44:48 policy-db-migrator | 3 0120-jpapdpsubgroup_policies.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:02 09:44:48 policy-db-migrator | 4 0130-jpapdpsubgroup_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:02 09:44:48 policy-db-migrator | 5 0140-jpapdpsubgroup_supportedpolicytypes.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:02 09:44:48 policy-db-migrator | 6 0150-jpatoscacapabilityassignment_attributes.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 7 0160-jpatoscacapabilityassignment_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 8 0170-jpatoscacapabilityassignment_occurrences.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 9 0180-jpatoscacapabilityassignment_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 10 0190-jpatoscacapabilitytype_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 11 0200-jpatoscacapabilitytype_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 12 0210-jpatoscadatatype_constraints.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 13 0220-jpatoscadatatype_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 14 0230-jpatoscadatatype_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 15 0240-jpatoscanodetemplate_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 16 0250-jpatoscanodetemplate_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 17 0260-jpatoscanodetype_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 18 0270-jpatoscanodetype_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 19 0280-jpatoscapolicy_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 20 0290-jpatoscapolicy_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 21 0300-jpatoscapolicy_targets.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:03 09:44:48 policy-db-migrator | 22 0310-jpatoscapolicytype_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 23 0320-jpatoscapolicytype_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 24 0330-jpatoscapolicytype_targets.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 25 0340-jpatoscapolicytype_triggers.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 26 0350-jpatoscaproperty_constraints.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 27 0360-jpatoscaproperty_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 28 0370-jpatoscarelationshiptype_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 29 0380-jpatoscarelationshiptype_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 30 0390-jpatoscarequirement_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 31 0400-jpatoscarequirement_occurrences.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 32 0410-jpatoscarequirement_properties.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 33 0420-jpatoscaservicetemplate_metadata.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:04 09:44:48 policy-db-migrator | 34 0430-jpatoscatopologytemplate_inputs.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:07 09:44:48 policy-db-migrator | 35 0440-pdpgroup_pdpsubgroup.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:07 09:44:48 policy-db-migrator | 36 0450-pdpgroup.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:07 09:44:48 policy-db-migrator | 37 0460-pdppolicystatus.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:07 09:44:48 policy-db-migrator | 38 0470-pdp.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 39 0480-pdpstatistics.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 40 0490-pdpsubgroup_pdp.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 41 0500-pdpsubgroup.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 42 0510-toscacapabilityassignment.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 43 0520-toscacapabilityassignments.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 44 0530-toscacapabilityassignments_toscacapabilityassignment.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 45 0540-toscacapabilitytype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 46 0550-toscacapabilitytypes.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 47 0560-toscacapabilitytypes_toscacapabilitytype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 48 0570-toscadatatype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 49 0580-toscadatatypes.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 50 0590-toscadatatypes_toscadatatype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 51 0600-toscanodetemplate.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 52 0610-toscanodetemplates.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 53 0620-toscanodetemplates_toscanodetemplate.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 54 0630-toscanodetype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 55 0640-toscanodetypes.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 56 0650-toscanodetypes_toscanodetype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 57 0660-toscaparameter.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:08 09:44:48 policy-db-migrator | 58 0670-toscapolicies.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 59 0680-toscapolicies_toscapolicy.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 60 0690-toscapolicy.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 61 0700-toscapolicytype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 62 0710-toscapolicytypes.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 63 0720-toscapolicytypes_toscapolicytype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 64 0730-toscaproperty.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 65 0740-toscarelationshiptype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 66 0750-toscarelationshiptypes.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 67 0760-toscarelationshiptypes_toscarelationshiptype.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 68 0770-toscarequirement.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 69 0780-toscarequirements.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 70 0790-toscarequirements_toscarequirement.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 71 0800-toscaservicetemplate.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 72 0810-toscatopologytemplate.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 73 0820-toscatrigger.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 74 0830-FK_ToscaNodeTemplate_capabilitiesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 75 0840-FK_ToscaNodeTemplate_requirementsName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:09 09:44:48 policy-db-migrator | 76 0850-FK_ToscaNodeType_requirementsName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 77 0860-FK_ToscaServiceTemplate_capabilityTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 78 0870-FK_ToscaServiceTemplate_dataTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 79 0880-FK_ToscaServiceTemplate_nodeTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 80 0890-FK_ToscaServiceTemplate_policyTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 81 0900-FK_ToscaServiceTemplate_relationshipTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 82 0910-FK_ToscaTopologyTemplate_nodeTemplatesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 83 0920-FK_ToscaTopologyTemplate_policyName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 84 0940-PdpPolicyStatus_PdpGroup.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 85 0950-TscaServiceTemplatetopologyTemplateParentLocalName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 86 0960-FK_ToscaNodeTemplate_capabilitiesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 87 0970-FK_ToscaNodeTemplate_requirementsName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 88 0980-FK_ToscaNodeType_requirementsName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 89 0990-FK_ToscaServiceTemplate_capabilityTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 90 1000-FK_ToscaServiceTemplate_dataTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 91 1010-FK_ToscaServiceTemplate_nodeTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:10 09:44:48 policy-db-migrator | 92 1020-FK_ToscaServiceTemplate_policyTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 93 1030-FK_ToscaServiceTemplate_relationshipTypesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 94 1040-FK_ToscaTopologyTemplate_nodeTemplatesName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 95 1050-FK_ToscaTopologyTemplate_policyName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 96 1060-TscaServiceTemplatetopologyTemplateParentLocalName.sql upgrade 0 0800 0407240942020800u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 97 0100-pdp.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 98 0110-idx_tsidx1.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 99 0120-pk_pdpstatistics.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 100 0130-pdpstatistics.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 101 0140-pk_pdpstatistics.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 102 0150-pdpstatistics.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 103 0160-jpapdpstatistics_enginestats.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 104 0170-jpapdpstatistics_enginestats.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 105 0180-jpapdpstatistics_enginestats.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 106 0190-jpapolicyaudit.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:11 09:44:48 policy-db-migrator | 107 0200-JpaPolicyAuditIndex_timestamp.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 108 0210-sequence.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 109 0220-sequence.sql upgrade 0800 0900 0407240942020900u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 110 0100-jpatoscapolicy_targets.sql upgrade 0900 1000 0407240942021000u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 111 0110-jpatoscapolicytype_targets.sql upgrade 0900 1000 0407240942021000u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 112 0120-toscatrigger.sql upgrade 0900 1000 0407240942021000u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 113 0130-jpatoscapolicytype_triggers.sql upgrade 0900 1000 0407240942021000u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 114 0140-toscaparameter.sql upgrade 0900 1000 0407240942021000u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 115 0150-toscaproperty.sql upgrade 0900 1000 0407240942021000u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 116 0160-jpapolicyaudit_pk.sql upgrade 0900 1000 0407240942021000u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 117 0170-pdpstatistics_pk.sql upgrade 0900 1000 0407240942021000u 1 2024-07-04 09:42:12 09:44:48 policy-db-migrator | 118 0180-jpatoscanodetemplate_metadata.sql upgrade 0900 1000 0407240942021000u 1 2024-07-04 09:42:13 09:44:48 policy-db-migrator | 119 0100-upgrade.sql upgrade 1000 1100 0407240942021100u 1 2024-07-04 09:42:13 09:44:48 policy-db-migrator | 120 0100-jpapolicyaudit_renameuser.sql upgrade 1100 1200 0407240942021200u 1 2024-07-04 09:42:13 09:44:48 policy-db-migrator | 121 0110-idx_tsidx1.sql upgrade 1100 1200 0407240942021200u 1 2024-07-04 09:42:13 09:44:48 policy-db-migrator | 122 0120-audit_sequence.sql upgrade 1100 1200 0407240942021200u 1 2024-07-04 09:42:13 09:44:48 policy-db-migrator | 123 0130-statistics_sequence.sql upgrade 1100 1200 0407240942021200u 1 2024-07-04 09:42:13 09:44:48 policy-db-migrator | 124 0100-pdpstatistics.sql upgrade 1200 1300 0407240942021300u 1 2024-07-04 09:42:13 09:44:48 policy-db-migrator | 125 0110-jpapdpstatistics_enginestats.sql upgrade 1200 1300 0407240942021300u 1 2024-07-04 09:42:13 09:44:48 policy-db-migrator | 126 0120-statistics_sequence.sql upgrade 1200 1300 0407240942021300u 1 2024-07-04 09:42:13 09:44:48 policy-db-migrator | policyadmin: OK @ 1300 09:44:48 =================================== 09:44:48 ======== Logs from pap ======== 09:44:48 policy-pap | Waiting for mariadb port 3306... 09:44:48 policy-pap | mariadb (172.17.0.3:3306) open 09:44:48 policy-pap | kafka (172.17.0.6:9092) open 09:44:48 policy-pap | Waiting for kafka port 9092... 09:44:48 policy-pap | Waiting for api port 6969... 09:44:48 policy-pap | api (172.17.0.8:6969) open 09:44:48 policy-pap | Policy pap config file: /opt/app/policy/pap/etc/papParameters.yaml 09:44:48 policy-pap | PDP group configuration file: /opt/app/policy/pap/etc/mounted/groups.json 09:44:48 policy-pap | 09:44:48 policy-pap | . ____ _ __ _ _ 09:44:48 policy-pap | /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \ 09:44:48 policy-pap | ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \ 09:44:48 policy-pap | \\/ ___)| |_)| | | | | || (_| | ) ) ) ) 09:44:48 policy-pap | ' |____| .__|_| |_|_| |_\__, | / / / / 09:44:48 policy-pap | =========|_|==============|___/=/_/_/_/ 09:44:48 policy-pap | :: Spring Boot :: (v3.1.10) 09:44:48 policy-pap | 09:44:48 policy-pap | [2024-07-04T09:42:28.136+00:00|INFO|Version|background-preinit] HV000001: Hibernate Validator 8.0.1.Final 09:44:48 policy-pap | [2024-07-04T09:42:28.197+00:00|INFO|PolicyPapApplication|main] Starting PolicyPapApplication using Java 17.0.11 with PID 38 (/app/pap.jar started by policy in /opt/app/policy/pap/bin) 09:44:48 policy-pap | [2024-07-04T09:42:28.198+00:00|INFO|PolicyPapApplication|main] No active profile set, falling back to 1 default profile: "default" 09:44:48 policy-pap | [2024-07-04T09:42:30.203+00:00|INFO|RepositoryConfigurationDelegate|main] Bootstrapping Spring Data JPA repositories in DEFAULT mode. 09:44:48 policy-pap | [2024-07-04T09:42:30.299+00:00|INFO|RepositoryConfigurationDelegate|main] Finished Spring Data repository scanning in 87 ms. Found 7 JPA repository interfaces. 09:44:48 policy-pap | [2024-07-04T09:42:30.740+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.exception.ServiceExceptionHandler 09:44:48 policy-pap | [2024-07-04T09:42:30.741+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.exception.ServiceExceptionHandler 09:44:48 policy-pap | [2024-07-04T09:42:31.422+00:00|INFO|TomcatWebServer|main] Tomcat initialized with port(s): 6969 (http) 09:44:48 policy-pap | [2024-07-04T09:42:31.435+00:00|INFO|Http11NioProtocol|main] Initializing ProtocolHandler ["http-nio-6969"] 09:44:48 policy-pap | [2024-07-04T09:42:31.438+00:00|INFO|StandardService|main] Starting service [Tomcat] 09:44:48 policy-pap | [2024-07-04T09:42:31.439+00:00|INFO|StandardEngine|main] Starting Servlet engine: [Apache Tomcat/10.1.19] 09:44:48 policy-pap | [2024-07-04T09:42:31.571+00:00|INFO|[/policy/pap/v1]|main] Initializing Spring embedded WebApplicationContext 09:44:48 policy-pap | [2024-07-04T09:42:31.571+00:00|INFO|ServletWebServerApplicationContext|main] Root WebApplicationContext: initialization completed in 3303 ms 09:44:48 policy-pap | [2024-07-04T09:42:32.057+00:00|INFO|LogHelper|main] HHH000204: Processing PersistenceUnitInfo [name: default] 09:44:48 policy-pap | [2024-07-04T09:42:32.141+00:00|INFO|Version|main] HHH000412: Hibernate ORM core version 5.6.15.Final 09:44:48 policy-pap | [2024-07-04T09:42:32.547+00:00|INFO|HikariDataSource|main] HikariPool-1 - Starting... 09:44:48 policy-pap | [2024-07-04T09:42:32.666+00:00|INFO|HikariPool|main] HikariPool-1 - Added connection org.mariadb.jdbc.Connection@4ee5b2d9 09:44:48 policy-pap | [2024-07-04T09:42:32.668+00:00|INFO|HikariDataSource|main] HikariPool-1 - Start completed. 09:44:48 policy-pap | [2024-07-04T09:42:32.695+00:00|INFO|Dialect|main] HHH000400: Using dialect: org.hibernate.dialect.MariaDB106Dialect 09:44:48 policy-pap | [2024-07-04T09:42:34.223+00:00|INFO|JtaPlatformInitiator|main] HHH000490: Using JtaPlatform implementation: [org.hibernate.engine.transaction.jta.platform.internal.NoJtaPlatform] 09:44:48 policy-pap | [2024-07-04T09:42:34.233+00:00|INFO|LocalContainerEntityManagerFactoryBean|main] Initialized JPA EntityManagerFactory for persistence unit 'default' 09:44:48 policy-pap | [2024-07-04T09:42:34.769+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.repository.PdpGroupRepository 09:44:48 policy-pap | [2024-07-04T09:42:35.224+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.repository.PolicyStatusRepository 09:44:48 policy-pap | [2024-07-04T09:42:35.336+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.repository.PolicyAuditRepository 09:44:48 policy-pap | [2024-07-04T09:42:35.603+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: 09:44:48 policy-pap | allow.auto.create.topics = true 09:44:48 policy-pap | auto.commit.interval.ms = 5000 09:44:48 policy-pap | auto.include.jmx.reporter = true 09:44:48 policy-pap | auto.offset.reset = latest 09:44:48 policy-pap | bootstrap.servers = [kafka:9092] 09:44:48 policy-pap | check.crcs = true 09:44:48 policy-pap | client.dns.lookup = use_all_dns_ips 09:44:48 policy-pap | client.id = consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-1 09:44:48 policy-pap | client.rack = 09:44:48 policy-pap | connections.max.idle.ms = 540000 09:44:48 policy-pap | default.api.timeout.ms = 60000 09:44:48 policy-pap | enable.auto.commit = true 09:44:48 policy-pap | exclude.internal.topics = true 09:44:48 policy-pap | fetch.max.bytes = 52428800 09:44:48 policy-pap | fetch.max.wait.ms = 500 09:44:48 policy-pap | fetch.min.bytes = 1 09:44:48 policy-pap | group.id = c181ccfa-a5ba-49d2-a57d-7770c93062fb 09:44:48 policy-pap | group.instance.id = null 09:44:48 policy-pap | heartbeat.interval.ms = 3000 09:44:48 policy-pap | interceptor.classes = [] 09:44:48 policy-pap | internal.leave.group.on.close = true 09:44:48 policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false 09:44:48 policy-pap | isolation.level = read_uncommitted 09:44:48 policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-pap | max.partition.fetch.bytes = 1048576 09:44:48 policy-pap | max.poll.interval.ms = 300000 09:44:48 policy-pap | max.poll.records = 500 09:44:48 policy-pap | metadata.max.age.ms = 300000 09:44:48 policy-pap | metric.reporters = [] 09:44:48 policy-pap | metrics.num.samples = 2 09:44:48 policy-pap | metrics.recording.level = INFO 09:44:48 policy-pap | metrics.sample.window.ms = 30000 09:44:48 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] 09:44:48 policy-pap | receive.buffer.bytes = 65536 09:44:48 policy-pap | reconnect.backoff.max.ms = 1000 09:44:48 policy-pap | reconnect.backoff.ms = 50 09:44:48 policy-pap | request.timeout.ms = 30000 09:44:48 policy-pap | retry.backoff.ms = 100 09:44:48 policy-pap | sasl.client.callback.handler.class = null 09:44:48 policy-pap | sasl.jaas.config = null 09:44:48 policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 policy-pap | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 policy-pap | sasl.kerberos.service.name = null 09:44:48 policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 policy-pap | sasl.login.callback.handler.class = null 09:44:48 policy-pap | sasl.login.class = null 09:44:48 policy-pap | sasl.login.connect.timeout.ms = null 09:44:48 policy-pap | sasl.login.read.timeout.ms = null 09:44:48 policy-pap | sasl.login.refresh.buffer.seconds = 300 09:44:48 policy-pap | sasl.login.refresh.min.period.seconds = 60 09:44:48 policy-pap | sasl.login.refresh.window.factor = 0.8 09:44:48 policy-pap | sasl.login.refresh.window.jitter = 0.05 09:44:48 policy-pap | sasl.login.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.login.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.mechanism = GSSAPI 09:44:48 policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 policy-pap | sasl.oauthbearer.expected.audience = null 09:44:48 policy-pap | sasl.oauthbearer.expected.issuer = null 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 policy-pap | sasl.oauthbearer.scope.claim.name = scope 09:44:48 policy-pap | sasl.oauthbearer.sub.claim.name = sub 09:44:48 policy-pap | sasl.oauthbearer.token.endpoint.url = null 09:44:48 policy-pap | security.protocol = PLAINTEXT 09:44:48 policy-pap | security.providers = null 09:44:48 policy-pap | send.buffer.bytes = 131072 09:44:48 policy-pap | session.timeout.ms = 45000 09:44:48 policy-pap | socket.connection.setup.timeout.max.ms = 30000 09:44:48 policy-pap | socket.connection.setup.timeout.ms = 10000 09:44:48 policy-pap | ssl.cipher.suites = null 09:44:48 policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 policy-pap | ssl.endpoint.identification.algorithm = https 09:44:48 policy-pap | ssl.engine.factory.class = null 09:44:48 policy-pap | ssl.key.password = null 09:44:48 policy-pap | ssl.keymanager.algorithm = SunX509 09:44:48 policy-pap | ssl.keystore.certificate.chain = null 09:44:48 policy-pap | ssl.keystore.key = null 09:44:48 policy-pap | ssl.keystore.location = null 09:44:48 policy-pap | ssl.keystore.password = null 09:44:48 policy-pap | ssl.keystore.type = JKS 09:44:48 policy-pap | ssl.protocol = TLSv1.3 09:44:48 policy-pap | ssl.provider = null 09:44:48 policy-pap | ssl.secure.random.implementation = null 09:44:48 policy-pap | ssl.trustmanager.algorithm = PKIX 09:44:48 policy-pap | ssl.truststore.certificates = null 09:44:48 policy-pap | ssl.truststore.location = null 09:44:48 policy-pap | ssl.truststore.password = null 09:44:48 policy-pap | ssl.truststore.type = JKS 09:44:48 policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-pap | 09:44:48 policy-pap | [2024-07-04T09:42:35.771+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 09:44:48 policy-pap | [2024-07-04T09:42:35.771+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 09:44:48 policy-pap | [2024-07-04T09:42:35.771+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1720086155770 09:44:48 policy-pap | [2024-07-04T09:42:35.774+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-1, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Subscribed to topic(s): policy-pdp-pap 09:44:48 policy-pap | [2024-07-04T09:42:35.774+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: 09:44:48 policy-pap | allow.auto.create.topics = true 09:44:48 policy-pap | auto.commit.interval.ms = 5000 09:44:48 policy-pap | auto.include.jmx.reporter = true 09:44:48 policy-pap | auto.offset.reset = latest 09:44:48 policy-pap | bootstrap.servers = [kafka:9092] 09:44:48 policy-pap | check.crcs = true 09:44:48 policy-pap | client.dns.lookup = use_all_dns_ips 09:44:48 policy-pap | client.id = consumer-policy-pap-2 09:44:48 policy-pap | client.rack = 09:44:48 policy-pap | connections.max.idle.ms = 540000 09:44:48 policy-pap | default.api.timeout.ms = 60000 09:44:48 policy-pap | enable.auto.commit = true 09:44:48 policy-pap | exclude.internal.topics = true 09:44:48 policy-pap | fetch.max.bytes = 52428800 09:44:48 policy-pap | fetch.max.wait.ms = 500 09:44:48 policy-pap | fetch.min.bytes = 1 09:44:48 policy-pap | group.id = policy-pap 09:44:48 policy-pap | group.instance.id = null 09:44:48 policy-pap | heartbeat.interval.ms = 3000 09:44:48 policy-pap | interceptor.classes = [] 09:44:48 policy-pap | internal.leave.group.on.close = true 09:44:48 policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false 09:44:48 policy-pap | isolation.level = read_uncommitted 09:44:48 policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-pap | max.partition.fetch.bytes = 1048576 09:44:48 policy-pap | max.poll.interval.ms = 300000 09:44:48 policy-pap | max.poll.records = 500 09:44:48 policy-pap | metadata.max.age.ms = 300000 09:44:48 policy-pap | metric.reporters = [] 09:44:48 policy-pap | metrics.num.samples = 2 09:44:48 policy-pap | metrics.recording.level = INFO 09:44:48 policy-pap | metrics.sample.window.ms = 30000 09:44:48 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] 09:44:48 policy-pap | receive.buffer.bytes = 65536 09:44:48 policy-pap | reconnect.backoff.max.ms = 1000 09:44:48 policy-pap | reconnect.backoff.ms = 50 09:44:48 policy-pap | request.timeout.ms = 30000 09:44:48 policy-pap | retry.backoff.ms = 100 09:44:48 policy-pap | sasl.client.callback.handler.class = null 09:44:48 policy-pap | sasl.jaas.config = null 09:44:48 policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 policy-pap | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 policy-pap | sasl.kerberos.service.name = null 09:44:48 policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 policy-pap | sasl.login.callback.handler.class = null 09:44:48 policy-pap | sasl.login.class = null 09:44:48 policy-pap | sasl.login.connect.timeout.ms = null 09:44:48 policy-pap | sasl.login.read.timeout.ms = null 09:44:48 policy-pap | sasl.login.refresh.buffer.seconds = 300 09:44:48 policy-pap | sasl.login.refresh.min.period.seconds = 60 09:44:48 policy-pap | sasl.login.refresh.window.factor = 0.8 09:44:48 policy-pap | sasl.login.refresh.window.jitter = 0.05 09:44:48 policy-pap | sasl.login.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.login.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.mechanism = GSSAPI 09:44:48 policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 policy-pap | sasl.oauthbearer.expected.audience = null 09:44:48 policy-pap | sasl.oauthbearer.expected.issuer = null 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 policy-pap | sasl.oauthbearer.scope.claim.name = scope 09:44:48 policy-pap | sasl.oauthbearer.sub.claim.name = sub 09:44:48 policy-pap | sasl.oauthbearer.token.endpoint.url = null 09:44:48 policy-pap | security.protocol = PLAINTEXT 09:44:48 policy-pap | security.providers = null 09:44:48 policy-pap | send.buffer.bytes = 131072 09:44:48 policy-pap | session.timeout.ms = 45000 09:44:48 policy-pap | socket.connection.setup.timeout.max.ms = 30000 09:44:48 policy-pap | socket.connection.setup.timeout.ms = 10000 09:44:48 policy-pap | ssl.cipher.suites = null 09:44:48 policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 policy-pap | ssl.endpoint.identification.algorithm = https 09:44:48 policy-pap | ssl.engine.factory.class = null 09:44:48 policy-pap | ssl.key.password = null 09:44:48 policy-pap | ssl.keymanager.algorithm = SunX509 09:44:48 policy-pap | ssl.keystore.certificate.chain = null 09:44:48 policy-pap | ssl.keystore.key = null 09:44:48 policy-pap | ssl.keystore.location = null 09:44:48 policy-pap | ssl.keystore.password = null 09:44:48 policy-pap | ssl.keystore.type = JKS 09:44:48 policy-pap | ssl.protocol = TLSv1.3 09:44:48 policy-pap | ssl.provider = null 09:44:48 policy-pap | ssl.secure.random.implementation = null 09:44:48 policy-pap | ssl.trustmanager.algorithm = PKIX 09:44:48 policy-pap | ssl.truststore.certificates = null 09:44:48 policy-pap | ssl.truststore.location = null 09:44:48 policy-pap | ssl.truststore.password = null 09:44:48 policy-pap | ssl.truststore.type = JKS 09:44:48 policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-pap | 09:44:48 policy-pap | [2024-07-04T09:42:35.780+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 09:44:48 policy-pap | [2024-07-04T09:42:35.780+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 09:44:48 policy-pap | [2024-07-04T09:42:35.780+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1720086155780 09:44:48 policy-pap | [2024-07-04T09:42:35.780+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-policy-pap-2, groupId=policy-pap] Subscribed to topic(s): policy-pdp-pap 09:44:48 policy-pap | [2024-07-04T09:42:36.071+00:00|INFO|PapDatabaseInitializer|main] Created initial pdpGroup in DB - PdpGroups(groups=[PdpGroup(name=defaultGroup, description=The default group that registers all supported policy types and pdps., pdpGroupState=ACTIVE, properties=null, pdpSubgroups=[PdpSubGroup(pdpType=apex, supportedPolicyTypes=[onap.policies.controlloop.operational.common.Apex 1.0.0, onap.policies.native.Apex 1.0.0], policies=[], currentInstanceCount=0, desiredInstanceCount=1, properties=null, pdpInstances=null)])]) from /opt/app/policy/pap/etc/mounted/groups.json 09:44:48 policy-pap | [2024-07-04T09:42:36.222+00:00|WARN|JpaBaseConfiguration$JpaWebConfiguration|main] spring.jpa.open-in-view is enabled by default. Therefore, database queries may be performed during view rendering. Explicitly configure spring.jpa.open-in-view to disable this warning 09:44:48 policy-pap | [2024-07-04T09:42:36.459+00:00|INFO|DefaultSecurityFilterChain|main] Will secure any request with [org.springframework.security.web.session.DisableEncodeUrlFilter@21ba0d33, org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter@afb7b03, org.springframework.security.web.context.SecurityContextHolderFilter@76e2a621, org.springframework.security.web.header.HeaderWriterFilter@18b58c77, org.springframework.security.web.authentication.logout.LogoutFilter@6719f206, org.springframework.security.web.authentication.www.BasicAuthenticationFilter@9825465, org.springframework.security.web.savedrequest.RequestCacheAwareFilter@2e7517aa, org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter@76105ac0, org.springframework.security.web.authentication.AnonymousAuthenticationFilter@4fd63c43, org.springframework.security.web.access.ExceptionTranslationFilter@5ccc971e, org.springframework.security.web.access.intercept.AuthorizationFilter@cd93621] 09:44:48 policy-pap | [2024-07-04T09:42:37.251+00:00|INFO|EndpointLinksResolver|main] Exposing 3 endpoint(s) beneath base path '' 09:44:48 policy-pap | [2024-07-04T09:42:37.350+00:00|INFO|Http11NioProtocol|main] Starting ProtocolHandler ["http-nio-6969"] 09:44:48 policy-pap | [2024-07-04T09:42:37.366+00:00|INFO|TomcatWebServer|main] Tomcat started on port(s): 6969 (http) with context path '/policy/pap/v1' 09:44:48 policy-pap | [2024-07-04T09:42:37.385+00:00|INFO|ServiceManager|main] Policy PAP starting 09:44:48 policy-pap | [2024-07-04T09:42:37.385+00:00|INFO|ServiceManager|main] Policy PAP starting Meter Registry 09:44:48 policy-pap | [2024-07-04T09:42:37.386+00:00|INFO|ServiceManager|main] Policy PAP starting PAP parameters 09:44:48 policy-pap | [2024-07-04T09:42:37.386+00:00|INFO|ServiceManager|main] Policy PAP starting Pdp Heartbeat Listener 09:44:48 policy-pap | [2024-07-04T09:42:37.386+00:00|INFO|ServiceManager|main] Policy PAP starting Response Request ID Dispatcher 09:44:48 policy-pap | [2024-07-04T09:42:37.386+00:00|INFO|ServiceManager|main] Policy PAP starting Heartbeat Request ID Dispatcher 09:44:48 policy-pap | [2024-07-04T09:42:37.386+00:00|INFO|ServiceManager|main] Policy PAP starting Response Message Dispatcher 09:44:48 policy-pap | [2024-07-04T09:42:37.388+00:00|INFO|TopicBase|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=c181ccfa-a5ba-49d2-a57d-7770c93062fb, consumerInstance=policy-pap, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: registering org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher@4e9695cf 09:44:48 policy-pap | [2024-07-04T09:42:37.399+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=c181ccfa-a5ba-49d2-a57d-7770c93062fb, consumerInstance=policy-pap, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting 09:44:48 policy-pap | [2024-07-04T09:42:37.400+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: 09:44:48 policy-pap | allow.auto.create.topics = true 09:44:48 policy-pap | auto.commit.interval.ms = 5000 09:44:48 policy-pap | auto.include.jmx.reporter = true 09:44:48 policy-pap | auto.offset.reset = latest 09:44:48 policy-pap | bootstrap.servers = [kafka:9092] 09:44:48 policy-pap | check.crcs = true 09:44:48 policy-pap | client.dns.lookup = use_all_dns_ips 09:44:48 policy-pap | client.id = consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3 09:44:48 policy-pap | client.rack = 09:44:48 policy-pap | connections.max.idle.ms = 540000 09:44:48 policy-pap | default.api.timeout.ms = 60000 09:44:48 policy-pap | enable.auto.commit = true 09:44:48 policy-pap | exclude.internal.topics = true 09:44:48 policy-pap | fetch.max.bytes = 52428800 09:44:48 policy-pap | fetch.max.wait.ms = 500 09:44:48 policy-pap | fetch.min.bytes = 1 09:44:48 policy-pap | group.id = c181ccfa-a5ba-49d2-a57d-7770c93062fb 09:44:48 policy-pap | group.instance.id = null 09:44:48 policy-pap | heartbeat.interval.ms = 3000 09:44:48 policy-pap | interceptor.classes = [] 09:44:48 policy-pap | internal.leave.group.on.close = true 09:44:48 policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false 09:44:48 policy-pap | isolation.level = read_uncommitted 09:44:48 policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-pap | max.partition.fetch.bytes = 1048576 09:44:48 policy-pap | max.poll.interval.ms = 300000 09:44:48 policy-pap | max.poll.records = 500 09:44:48 policy-pap | metadata.max.age.ms = 300000 09:44:48 policy-pap | metric.reporters = [] 09:44:48 policy-pap | metrics.num.samples = 2 09:44:48 policy-pap | metrics.recording.level = INFO 09:44:48 policy-pap | metrics.sample.window.ms = 30000 09:44:48 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] 09:44:48 policy-pap | receive.buffer.bytes = 65536 09:44:48 policy-pap | reconnect.backoff.max.ms = 1000 09:44:48 policy-pap | reconnect.backoff.ms = 50 09:44:48 policy-pap | request.timeout.ms = 30000 09:44:48 policy-pap | retry.backoff.ms = 100 09:44:48 policy-pap | sasl.client.callback.handler.class = null 09:44:48 policy-pap | sasl.jaas.config = null 09:44:48 policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 policy-pap | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 policy-pap | sasl.kerberos.service.name = null 09:44:48 policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 policy-pap | sasl.login.callback.handler.class = null 09:44:48 policy-pap | sasl.login.class = null 09:44:48 policy-pap | sasl.login.connect.timeout.ms = null 09:44:48 policy-pap | sasl.login.read.timeout.ms = null 09:44:48 policy-pap | sasl.login.refresh.buffer.seconds = 300 09:44:48 policy-pap | sasl.login.refresh.min.period.seconds = 60 09:44:48 policy-pap | sasl.login.refresh.window.factor = 0.8 09:44:48 policy-pap | sasl.login.refresh.window.jitter = 0.05 09:44:48 policy-pap | sasl.login.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.login.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.mechanism = GSSAPI 09:44:48 policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 policy-pap | sasl.oauthbearer.expected.audience = null 09:44:48 policy-pap | sasl.oauthbearer.expected.issuer = null 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 policy-pap | sasl.oauthbearer.scope.claim.name = scope 09:44:48 policy-pap | sasl.oauthbearer.sub.claim.name = sub 09:44:48 policy-pap | sasl.oauthbearer.token.endpoint.url = null 09:44:48 policy-pap | security.protocol = PLAINTEXT 09:44:48 policy-pap | security.providers = null 09:44:48 policy-pap | send.buffer.bytes = 131072 09:44:48 policy-pap | session.timeout.ms = 45000 09:44:48 policy-pap | socket.connection.setup.timeout.max.ms = 30000 09:44:48 policy-pap | socket.connection.setup.timeout.ms = 10000 09:44:48 policy-pap | ssl.cipher.suites = null 09:44:48 policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 policy-pap | ssl.endpoint.identification.algorithm = https 09:44:48 policy-pap | ssl.engine.factory.class = null 09:44:48 policy-pap | ssl.key.password = null 09:44:48 policy-pap | ssl.keymanager.algorithm = SunX509 09:44:48 policy-pap | ssl.keystore.certificate.chain = null 09:44:48 policy-pap | ssl.keystore.key = null 09:44:48 policy-pap | ssl.keystore.location = null 09:44:48 policy-pap | ssl.keystore.password = null 09:44:48 policy-pap | ssl.keystore.type = JKS 09:44:48 policy-pap | ssl.protocol = TLSv1.3 09:44:48 policy-pap | ssl.provider = null 09:44:48 policy-pap | ssl.secure.random.implementation = null 09:44:48 policy-pap | ssl.trustmanager.algorithm = PKIX 09:44:48 policy-pap | ssl.truststore.certificates = null 09:44:48 policy-pap | ssl.truststore.location = null 09:44:48 policy-pap | ssl.truststore.password = null 09:44:48 policy-pap | ssl.truststore.type = JKS 09:44:48 policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-pap | 09:44:48 policy-pap | [2024-07-04T09:42:37.406+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 09:44:48 policy-pap | [2024-07-04T09:42:37.406+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 09:44:48 policy-pap | [2024-07-04T09:42:37.406+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1720086157406 09:44:48 policy-pap | [2024-07-04T09:42:37.406+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Subscribed to topic(s): policy-pdp-pap 09:44:48 policy-pap | [2024-07-04T09:42:37.406+00:00|INFO|ServiceManager|main] Policy PAP starting Heartbeat Message Dispatcher 09:44:48 policy-pap | [2024-07-04T09:42:37.406+00:00|INFO|TopicBase|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=policy-pap, consumerInstance=05955a93-f210-479e-b9f4-e33eaf580067, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-heartbeat, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: registering org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher@77978658 09:44:48 policy-pap | [2024-07-04T09:42:37.407+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=policy-pap, consumerInstance=05955a93-f210-479e-b9f4-e33eaf580067, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-heartbeat, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting 09:44:48 policy-pap | [2024-07-04T09:42:37.407+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: 09:44:48 policy-pap | allow.auto.create.topics = true 09:44:48 policy-pap | auto.commit.interval.ms = 5000 09:44:48 policy-pap | auto.include.jmx.reporter = true 09:44:48 policy-pap | auto.offset.reset = latest 09:44:48 policy-pap | bootstrap.servers = [kafka:9092] 09:44:48 policy-pap | check.crcs = true 09:44:48 policy-pap | client.dns.lookup = use_all_dns_ips 09:44:48 policy-pap | client.id = consumer-policy-pap-4 09:44:48 policy-pap | client.rack = 09:44:48 policy-pap | connections.max.idle.ms = 540000 09:44:48 policy-pap | default.api.timeout.ms = 60000 09:44:48 policy-pap | enable.auto.commit = true 09:44:48 policy-pap | exclude.internal.topics = true 09:44:48 policy-pap | fetch.max.bytes = 52428800 09:44:48 policy-pap | fetch.max.wait.ms = 500 09:44:48 policy-pap | fetch.min.bytes = 1 09:44:48 policy-pap | group.id = policy-pap 09:44:48 policy-pap | group.instance.id = null 09:44:48 policy-pap | heartbeat.interval.ms = 3000 09:44:48 policy-pap | interceptor.classes = [] 09:44:48 policy-pap | internal.leave.group.on.close = true 09:44:48 policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false 09:44:48 policy-pap | isolation.level = read_uncommitted 09:44:48 policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-pap | max.partition.fetch.bytes = 1048576 09:44:48 policy-pap | max.poll.interval.ms = 300000 09:44:48 policy-pap | max.poll.records = 500 09:44:48 policy-pap | metadata.max.age.ms = 300000 09:44:48 policy-pap | metric.reporters = [] 09:44:48 policy-pap | metrics.num.samples = 2 09:44:48 policy-pap | metrics.recording.level = INFO 09:44:48 policy-pap | metrics.sample.window.ms = 30000 09:44:48 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] 09:44:48 policy-pap | receive.buffer.bytes = 65536 09:44:48 policy-pap | reconnect.backoff.max.ms = 1000 09:44:48 policy-pap | reconnect.backoff.ms = 50 09:44:48 policy-pap | request.timeout.ms = 30000 09:44:48 policy-pap | retry.backoff.ms = 100 09:44:48 policy-pap | sasl.client.callback.handler.class = null 09:44:48 policy-pap | sasl.jaas.config = null 09:44:48 policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 policy-pap | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 policy-pap | sasl.kerberos.service.name = null 09:44:48 policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 policy-pap | sasl.login.callback.handler.class = null 09:44:48 policy-pap | sasl.login.class = null 09:44:48 policy-pap | sasl.login.connect.timeout.ms = null 09:44:48 policy-pap | sasl.login.read.timeout.ms = null 09:44:48 policy-pap | sasl.login.refresh.buffer.seconds = 300 09:44:48 policy-pap | sasl.login.refresh.min.period.seconds = 60 09:44:48 policy-pap | sasl.login.refresh.window.factor = 0.8 09:44:48 policy-pap | sasl.login.refresh.window.jitter = 0.05 09:44:48 policy-pap | sasl.login.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.login.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.mechanism = GSSAPI 09:44:48 policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 policy-pap | sasl.oauthbearer.expected.audience = null 09:44:48 policy-pap | sasl.oauthbearer.expected.issuer = null 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 policy-pap | sasl.oauthbearer.scope.claim.name = scope 09:44:48 policy-pap | sasl.oauthbearer.sub.claim.name = sub 09:44:48 policy-pap | sasl.oauthbearer.token.endpoint.url = null 09:44:48 policy-pap | security.protocol = PLAINTEXT 09:44:48 policy-pap | security.providers = null 09:44:48 policy-pap | send.buffer.bytes = 131072 09:44:48 policy-pap | session.timeout.ms = 45000 09:44:48 policy-pap | socket.connection.setup.timeout.max.ms = 30000 09:44:48 policy-pap | socket.connection.setup.timeout.ms = 10000 09:44:48 policy-pap | ssl.cipher.suites = null 09:44:48 policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 policy-pap | ssl.endpoint.identification.algorithm = https 09:44:48 policy-pap | ssl.engine.factory.class = null 09:44:48 policy-pap | ssl.key.password = null 09:44:48 policy-pap | ssl.keymanager.algorithm = SunX509 09:44:48 policy-pap | ssl.keystore.certificate.chain = null 09:44:48 policy-pap | ssl.keystore.key = null 09:44:48 policy-pap | ssl.keystore.location = null 09:44:48 policy-pap | ssl.keystore.password = null 09:44:48 policy-pap | ssl.keystore.type = JKS 09:44:48 policy-pap | ssl.protocol = TLSv1.3 09:44:48 policy-pap | ssl.provider = null 09:44:48 policy-pap | ssl.secure.random.implementation = null 09:44:48 policy-pap | ssl.trustmanager.algorithm = PKIX 09:44:48 policy-pap | ssl.truststore.certificates = null 09:44:48 policy-pap | ssl.truststore.location = null 09:44:48 policy-pap | ssl.truststore.password = null 09:44:48 policy-pap | ssl.truststore.type = JKS 09:44:48 policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer 09:44:48 policy-pap | 09:44:48 policy-pap | [2024-07-04T09:42:37.430+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 09:44:48 policy-pap | [2024-07-04T09:42:37.430+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 09:44:48 policy-pap | [2024-07-04T09:42:37.430+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1720086157430 09:44:48 policy-pap | [2024-07-04T09:42:37.431+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Subscribed to topic(s): policy-pdp-pap 09:44:48 policy-pap | [2024-07-04T09:42:37.431+00:00|INFO|ServiceManager|main] Policy PAP starting topics 09:44:48 policy-pap | [2024-07-04T09:42:37.431+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=policy-pap, consumerInstance=05955a93-f210-479e-b9f4-e33eaf580067, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-heartbeat,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-heartbeat, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting 09:44:48 policy-pap | [2024-07-04T09:42:37.431+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=c181ccfa-a5ba-49d2-a57d-7770c93062fb, consumerInstance=policy-pap, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting 09:44:48 policy-pap | [2024-07-04T09:42:37.431+00:00|INFO|InlineBusTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=6bf70c8f-e16d-4e02-aea7-93c43fe94953, alive=false, publisher=null]]: starting 09:44:48 policy-pap | [2024-07-04T09:42:37.470+00:00|INFO|ProducerConfig|main] ProducerConfig values: 09:44:48 policy-pap | acks = -1 09:44:48 policy-pap | auto.include.jmx.reporter = true 09:44:48 policy-pap | batch.size = 16384 09:44:48 policy-pap | bootstrap.servers = [kafka:9092] 09:44:48 policy-pap | buffer.memory = 33554432 09:44:48 policy-pap | client.dns.lookup = use_all_dns_ips 09:44:48 policy-pap | client.id = producer-1 09:44:48 policy-pap | compression.type = none 09:44:48 policy-pap | connections.max.idle.ms = 540000 09:44:48 policy-pap | delivery.timeout.ms = 120000 09:44:48 policy-pap | enable.idempotence = true 09:44:48 policy-pap | interceptor.classes = [] 09:44:48 policy-pap | key.serializer = class org.apache.kafka.common.serialization.StringSerializer 09:44:48 policy-pap | linger.ms = 0 09:44:48 policy-pap | max.block.ms = 60000 09:44:48 policy-pap | max.in.flight.requests.per.connection = 5 09:44:48 policy-pap | max.request.size = 1048576 09:44:48 policy-pap | metadata.max.age.ms = 300000 09:44:48 policy-pap | metadata.max.idle.ms = 300000 09:44:48 policy-pap | metric.reporters = [] 09:44:48 policy-pap | metrics.num.samples = 2 09:44:48 policy-pap | metrics.recording.level = INFO 09:44:48 policy-pap | metrics.sample.window.ms = 30000 09:44:48 policy-pap | partitioner.adaptive.partitioning.enable = true 09:44:48 policy-pap | partitioner.availability.timeout.ms = 0 09:44:48 policy-pap | partitioner.class = null 09:44:48 policy-pap | partitioner.ignore.keys = false 09:44:48 policy-pap | receive.buffer.bytes = 32768 09:44:48 policy-pap | reconnect.backoff.max.ms = 1000 09:44:48 policy-pap | reconnect.backoff.ms = 50 09:44:48 policy-pap | request.timeout.ms = 30000 09:44:48 policy-pap | retries = 2147483647 09:44:48 policy-pap | retry.backoff.ms = 100 09:44:48 policy-pap | sasl.client.callback.handler.class = null 09:44:48 policy-pap | sasl.jaas.config = null 09:44:48 policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 policy-pap | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 policy-pap | sasl.kerberos.service.name = null 09:44:48 policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 policy-pap | sasl.login.callback.handler.class = null 09:44:48 policy-pap | sasl.login.class = null 09:44:48 policy-pap | sasl.login.connect.timeout.ms = null 09:44:48 policy-pap | sasl.login.read.timeout.ms = null 09:44:48 policy-pap | sasl.login.refresh.buffer.seconds = 300 09:44:48 policy-pap | sasl.login.refresh.min.period.seconds = 60 09:44:48 policy-pap | sasl.login.refresh.window.factor = 0.8 09:44:48 policy-pap | sasl.login.refresh.window.jitter = 0.05 09:44:48 policy-pap | sasl.login.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.login.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.mechanism = GSSAPI 09:44:48 policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 policy-pap | sasl.oauthbearer.expected.audience = null 09:44:48 policy-pap | sasl.oauthbearer.expected.issuer = null 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 policy-pap | sasl.oauthbearer.scope.claim.name = scope 09:44:48 policy-pap | sasl.oauthbearer.sub.claim.name = sub 09:44:48 policy-pap | sasl.oauthbearer.token.endpoint.url = null 09:44:48 policy-pap | security.protocol = PLAINTEXT 09:44:48 policy-pap | security.providers = null 09:44:48 policy-pap | send.buffer.bytes = 131072 09:44:48 policy-pap | socket.connection.setup.timeout.max.ms = 30000 09:44:48 policy-pap | socket.connection.setup.timeout.ms = 10000 09:44:48 policy-pap | ssl.cipher.suites = null 09:44:48 policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 policy-pap | ssl.endpoint.identification.algorithm = https 09:44:48 policy-pap | ssl.engine.factory.class = null 09:44:48 policy-pap | ssl.key.password = null 09:44:48 policy-pap | ssl.keymanager.algorithm = SunX509 09:44:48 policy-pap | ssl.keystore.certificate.chain = null 09:44:48 policy-pap | ssl.keystore.key = null 09:44:48 policy-pap | ssl.keystore.location = null 09:44:48 policy-pap | ssl.keystore.password = null 09:44:48 policy-pap | ssl.keystore.type = JKS 09:44:48 policy-pap | ssl.protocol = TLSv1.3 09:44:48 policy-pap | ssl.provider = null 09:44:48 policy-pap | ssl.secure.random.implementation = null 09:44:48 policy-pap | ssl.trustmanager.algorithm = PKIX 09:44:48 policy-pap | ssl.truststore.certificates = null 09:44:48 policy-pap | ssl.truststore.location = null 09:44:48 policy-pap | ssl.truststore.password = null 09:44:48 policy-pap | ssl.truststore.type = JKS 09:44:48 policy-pap | transaction.timeout.ms = 60000 09:44:48 policy-pap | transactional.id = null 09:44:48 policy-pap | value.serializer = class org.apache.kafka.common.serialization.StringSerializer 09:44:48 policy-pap | 09:44:48 policy-pap | [2024-07-04T09:42:37.480+00:00|INFO|KafkaProducer|main] [Producer clientId=producer-1] Instantiated an idempotent producer. 09:44:48 policy-pap | [2024-07-04T09:42:37.496+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 09:44:48 policy-pap | [2024-07-04T09:42:37.496+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 09:44:48 policy-pap | [2024-07-04T09:42:37.496+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1720086157495 09:44:48 policy-pap | [2024-07-04T09:42:37.496+00:00|INFO|InlineKafkaTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=6bf70c8f-e16d-4e02-aea7-93c43fe94953, alive=false, publisher=KafkaPublisherWrapper []]]: KAFKA SINK created 09:44:48 policy-pap | [2024-07-04T09:42:37.496+00:00|INFO|InlineBusTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=48e26709-9c42-428f-a472-ee5078f014d2, alive=false, publisher=null]]: starting 09:44:48 policy-pap | [2024-07-04T09:42:37.497+00:00|INFO|ProducerConfig|main] ProducerConfig values: 09:44:48 policy-pap | acks = -1 09:44:48 policy-pap | auto.include.jmx.reporter = true 09:44:48 policy-pap | batch.size = 16384 09:44:48 policy-pap | bootstrap.servers = [kafka:9092] 09:44:48 policy-pap | buffer.memory = 33554432 09:44:48 policy-pap | client.dns.lookup = use_all_dns_ips 09:44:48 policy-pap | client.id = producer-2 09:44:48 policy-pap | compression.type = none 09:44:48 policy-pap | connections.max.idle.ms = 540000 09:44:48 policy-pap | delivery.timeout.ms = 120000 09:44:48 policy-pap | enable.idempotence = true 09:44:48 policy-pap | interceptor.classes = [] 09:44:48 policy-pap | key.serializer = class org.apache.kafka.common.serialization.StringSerializer 09:44:48 policy-pap | linger.ms = 0 09:44:48 policy-pap | max.block.ms = 60000 09:44:48 policy-pap | max.in.flight.requests.per.connection = 5 09:44:48 policy-pap | max.request.size = 1048576 09:44:48 policy-pap | metadata.max.age.ms = 300000 09:44:48 policy-pap | metadata.max.idle.ms = 300000 09:44:48 policy-pap | metric.reporters = [] 09:44:48 policy-pap | metrics.num.samples = 2 09:44:48 policy-pap | metrics.recording.level = INFO 09:44:48 policy-pap | metrics.sample.window.ms = 30000 09:44:48 policy-pap | partitioner.adaptive.partitioning.enable = true 09:44:48 policy-pap | partitioner.availability.timeout.ms = 0 09:44:48 policy-pap | partitioner.class = null 09:44:48 policy-pap | partitioner.ignore.keys = false 09:44:48 policy-pap | receive.buffer.bytes = 32768 09:44:48 policy-pap | reconnect.backoff.max.ms = 1000 09:44:48 policy-pap | reconnect.backoff.ms = 50 09:44:48 policy-pap | request.timeout.ms = 30000 09:44:48 policy-pap | retries = 2147483647 09:44:48 policy-pap | retry.backoff.ms = 100 09:44:48 policy-pap | sasl.client.callback.handler.class = null 09:44:48 policy-pap | sasl.jaas.config = null 09:44:48 policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit 09:44:48 policy-pap | sasl.kerberos.min.time.before.relogin = 60000 09:44:48 policy-pap | sasl.kerberos.service.name = null 09:44:48 policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 09:44:48 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 09:44:48 policy-pap | sasl.login.callback.handler.class = null 09:44:48 policy-pap | sasl.login.class = null 09:44:48 policy-pap | sasl.login.connect.timeout.ms = null 09:44:48 policy-pap | sasl.login.read.timeout.ms = null 09:44:48 policy-pap | sasl.login.refresh.buffer.seconds = 300 09:44:48 policy-pap | sasl.login.refresh.min.period.seconds = 60 09:44:48 policy-pap | sasl.login.refresh.window.factor = 0.8 09:44:48 policy-pap | sasl.login.refresh.window.jitter = 0.05 09:44:48 policy-pap | sasl.login.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.login.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.mechanism = GSSAPI 09:44:48 policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 09:44:48 policy-pap | sasl.oauthbearer.expected.audience = null 09:44:48 policy-pap | sasl.oauthbearer.expected.issuer = null 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 09:44:48 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null 09:44:48 policy-pap | sasl.oauthbearer.scope.claim.name = scope 09:44:48 policy-pap | sasl.oauthbearer.sub.claim.name = sub 09:44:48 policy-pap | sasl.oauthbearer.token.endpoint.url = null 09:44:48 policy-pap | security.protocol = PLAINTEXT 09:44:48 policy-pap | security.providers = null 09:44:48 policy-pap | send.buffer.bytes = 131072 09:44:48 policy-pap | socket.connection.setup.timeout.max.ms = 30000 09:44:48 policy-pap | socket.connection.setup.timeout.ms = 10000 09:44:48 policy-pap | ssl.cipher.suites = null 09:44:48 policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] 09:44:48 policy-pap | ssl.endpoint.identification.algorithm = https 09:44:48 policy-pap | ssl.engine.factory.class = null 09:44:48 policy-pap | ssl.key.password = null 09:44:48 policy-pap | ssl.keymanager.algorithm = SunX509 09:44:48 policy-pap | ssl.keystore.certificate.chain = null 09:44:48 policy-pap | ssl.keystore.key = null 09:44:48 policy-pap | ssl.keystore.location = null 09:44:48 policy-pap | ssl.keystore.password = null 09:44:48 policy-pap | ssl.keystore.type = JKS 09:44:48 policy-pap | ssl.protocol = TLSv1.3 09:44:48 policy-pap | ssl.provider = null 09:44:48 policy-pap | ssl.secure.random.implementation = null 09:44:48 policy-pap | ssl.trustmanager.algorithm = PKIX 09:44:48 policy-pap | ssl.truststore.certificates = null 09:44:48 policy-pap | ssl.truststore.location = null 09:44:48 policy-pap | ssl.truststore.password = null 09:44:48 policy-pap | ssl.truststore.type = JKS 09:44:48 policy-pap | transaction.timeout.ms = 60000 09:44:48 policy-pap | transactional.id = null 09:44:48 policy-pap | value.serializer = class org.apache.kafka.common.serialization.StringSerializer 09:44:48 policy-pap | 09:44:48 policy-pap | [2024-07-04T09:42:37.497+00:00|INFO|KafkaProducer|main] [Producer clientId=producer-2] Instantiated an idempotent producer. 09:44:48 policy-pap | [2024-07-04T09:42:37.500+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 09:44:48 policy-pap | [2024-07-04T09:42:37.500+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 09:44:48 policy-pap | [2024-07-04T09:42:37.500+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1720086157500 09:44:48 policy-pap | [2024-07-04T09:42:37.501+00:00|INFO|InlineKafkaTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=48e26709-9c42-428f-a472-ee5078f014d2, alive=false, publisher=KafkaPublisherWrapper []]]: KAFKA SINK created 09:44:48 policy-pap | [2024-07-04T09:42:37.501+00:00|INFO|ServiceManager|main] Policy PAP starting PAP Activator 09:44:48 policy-pap | [2024-07-04T09:42:37.501+00:00|INFO|ServiceManager|main] Policy PAP starting PDP publisher 09:44:48 policy-pap | [2024-07-04T09:42:37.505+00:00|INFO|ServiceManager|main] Policy PAP starting Policy Notification publisher 09:44:48 policy-pap | [2024-07-04T09:42:37.505+00:00|INFO|ServiceManager|main] Policy PAP starting PDP update timers 09:44:48 policy-pap | [2024-07-04T09:42:37.508+00:00|INFO|ServiceManager|main] Policy PAP starting PDP state-change timers 09:44:48 policy-pap | [2024-07-04T09:42:37.509+00:00|INFO|ServiceManager|main] Policy PAP starting PDP modification lock 09:44:48 policy-pap | [2024-07-04T09:42:37.509+00:00|INFO|ServiceManager|main] Policy PAP starting PDP modification requests 09:44:48 policy-pap | [2024-07-04T09:42:37.509+00:00|INFO|TimerManager|Thread-10] timer manager state-change started 09:44:48 policy-pap | [2024-07-04T09:42:37.509+00:00|INFO|ServiceManager|main] Policy PAP starting PDP expiration timer 09:44:48 policy-pap | [2024-07-04T09:42:37.510+00:00|INFO|TimerManager|Thread-9] timer manager update started 09:44:48 policy-pap | [2024-07-04T09:42:37.510+00:00|INFO|ServiceManager|main] Policy PAP started 09:44:48 policy-pap | [2024-07-04T09:42:37.511+00:00|INFO|PolicyPapApplication|main] Started PolicyPapApplication in 10.116 seconds (process running for 10.725) 09:44:48 policy-pap | [2024-07-04T09:42:37.937+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 2 : {policy-pdp-pap=UNKNOWN_TOPIC_OR_PARTITION} 09:44:48 policy-pap | [2024-07-04T09:42:37.938+00:00|INFO|Metadata|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Cluster ID: 5fWDTH-mSmOqNb5HKZw_Tw 09:44:48 policy-pap | [2024-07-04T09:42:37.939+00:00|INFO|Metadata|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] Cluster ID: 5fWDTH-mSmOqNb5HKZw_Tw 09:44:48 policy-pap | [2024-07-04T09:42:37.939+00:00|INFO|Metadata|kafka-producer-network-thread | producer-2] [Producer clientId=producer-2] Cluster ID: 5fWDTH-mSmOqNb5HKZw_Tw 09:44:48 policy-pap | [2024-07-04T09:42:38.009+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Error while fetching metadata with correlation id 2 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.009+00:00|INFO|Metadata|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Cluster ID: 5fWDTH-mSmOqNb5HKZw_Tw 09:44:48 policy-pap | [2024-07-04T09:42:38.056+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 4 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.063+00:00|INFO|TransactionManager|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] ProducerId set to 0 with epoch 0 09:44:48 policy-pap | [2024-07-04T09:42:38.073+00:00|INFO|TransactionManager|kafka-producer-network-thread | producer-2] [Producer clientId=producer-2] ProducerId set to 1 with epoch 0 09:44:48 policy-pap | [2024-07-04T09:42:38.126+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Error while fetching metadata with correlation id 4 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.230+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Error while fetching metadata with correlation id 6 : {policy-pdp-pap=UNKNOWN_TOPIC_OR_PARTITION} 09:44:48 policy-pap | [2024-07-04T09:42:38.262+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 6 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.340+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Error while fetching metadata with correlation id 8 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.379+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 8 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.446+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Error while fetching metadata with correlation id 10 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.484+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 10 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.551+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Error while fetching metadata with correlation id 12 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.591+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 12 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.658+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Error while fetching metadata with correlation id 14 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.697+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 14 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.766+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Error while fetching metadata with correlation id 16 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.804+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 16 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.876+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Error while fetching metadata with correlation id 18 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.907+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 18 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} 09:44:48 policy-pap | [2024-07-04T09:42:38.988+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Discovered group coordinator kafka:9092 (id: 2147483646 rack: null) 09:44:48 policy-pap | [2024-07-04T09:42:38.994+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] (Re-)joining group 09:44:48 policy-pap | [2024-07-04T09:42:39.012+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Discovered group coordinator kafka:9092 (id: 2147483646 rack: null) 09:44:48 policy-pap | [2024-07-04T09:42:39.016+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Request joining group due to: need to re-join with the given member-id: consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3-6b61b2ad-4dad-4d27-a110-dd1a2dddc971 09:44:48 policy-pap | [2024-07-04T09:42:39.016+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) 09:44:48 policy-pap | [2024-07-04T09:42:39.016+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] (Re-)joining group 09:44:48 policy-pap | [2024-07-04T09:42:39.020+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] (Re-)joining group 09:44:48 policy-pap | [2024-07-04T09:42:39.028+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Request joining group due to: need to re-join with the given member-id: consumer-policy-pap-4-bc3a0567-01a8-42b4-92fb-cd90febcdb27 09:44:48 policy-pap | [2024-07-04T09:42:39.028+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) 09:44:48 policy-pap | [2024-07-04T09:42:39.028+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] (Re-)joining group 09:44:48 policy-pap | [2024-07-04T09:42:41.593+00:00|INFO|[/policy/pap/v1]|http-nio-6969-exec-3] Initializing Spring DispatcherServlet 'dispatcherServlet' 09:44:48 policy-pap | [2024-07-04T09:42:41.593+00:00|INFO|DispatcherServlet|http-nio-6969-exec-3] Initializing Servlet 'dispatcherServlet' 09:44:48 policy-pap | [2024-07-04T09:42:41.595+00:00|INFO|DispatcherServlet|http-nio-6969-exec-3] Completed initialization in 2 ms 09:44:48 policy-pap | [2024-07-04T09:42:42.044+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Successfully joined group with generation Generation{generationId=1, memberId='consumer-policy-pap-4-bc3a0567-01a8-42b4-92fb-cd90febcdb27', protocol='range'} 09:44:48 policy-pap | [2024-07-04T09:42:42.044+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Successfully joined group with generation Generation{generationId=1, memberId='consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3-6b61b2ad-4dad-4d27-a110-dd1a2dddc971', protocol='range'} 09:44:48 policy-pap | [2024-07-04T09:42:42.050+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Finished assignment for group at generation 1: {consumer-policy-pap-4-bc3a0567-01a8-42b4-92fb-cd90febcdb27=Assignment(partitions=[policy-pdp-pap-0])} 09:44:48 policy-pap | [2024-07-04T09:42:42.050+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Finished assignment for group at generation 1: {consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3-6b61b2ad-4dad-4d27-a110-dd1a2dddc971=Assignment(partitions=[policy-pdp-pap-0])} 09:44:48 policy-pap | [2024-07-04T09:42:42.073+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Successfully synced group in generation Generation{generationId=1, memberId='consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3-6b61b2ad-4dad-4d27-a110-dd1a2dddc971', protocol='range'} 09:44:48 policy-pap | [2024-07-04T09:42:42.073+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Successfully synced group in generation Generation{generationId=1, memberId='consumer-policy-pap-4-bc3a0567-01a8-42b4-92fb-cd90febcdb27', protocol='range'} 09:44:48 policy-pap | [2024-07-04T09:42:42.073+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Notifying assignor about the new Assignment(partitions=[policy-pdp-pap-0]) 09:44:48 policy-pap | [2024-07-04T09:42:42.074+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Notifying assignor about the new Assignment(partitions=[policy-pdp-pap-0]) 09:44:48 policy-pap | [2024-07-04T09:42:42.077+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Adding newly assigned partitions: policy-pdp-pap-0 09:44:48 policy-pap | [2024-07-04T09:42:42.077+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Adding newly assigned partitions: policy-pdp-pap-0 09:44:48 policy-pap | [2024-07-04T09:42:42.097+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Found no committed offset for partition policy-pdp-pap-0 09:44:48 policy-pap | [2024-07-04T09:42:42.098+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Found no committed offset for partition policy-pdp-pap-0 09:44:48 policy-pap | [2024-07-04T09:42:42.116+00:00|INFO|SubscriptionState|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-c181ccfa-a5ba-49d2-a57d-7770c93062fb-3, groupId=c181ccfa-a5ba-49d2-a57d-7770c93062fb] Resetting offset for partition policy-pdp-pap-0 to position FetchPosition{offset=1, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:9092 (id: 1 rack: null)], epoch=0}}. 09:44:48 policy-pap | [2024-07-04T09:42:42.116+00:00|INFO|SubscriptionState|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Resetting offset for partition policy-pdp-pap-0 to position FetchPosition{offset=1, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:9092 (id: 1 rack: null)], epoch=0}}. 09:44:48 policy-pap | [2024-07-04T09:42:59.349+00:00|INFO|OrderedServiceImpl|KAFKA-source-policy-heartbeat] ***** OrderedServiceImpl implementers: 09:44:48 policy-pap | [] 09:44:48 policy-pap | [2024-07-04T09:42:59.350+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] 09:44:48 policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"908a25c1-f405-4bad-b634-83eb08f9182d","timestampMs":1720086179313,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup"} 09:44:48 policy-pap | [2024-07-04T09:42:59.350+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"908a25c1-f405-4bad-b634-83eb08f9182d","timestampMs":1720086179313,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup"} 09:44:48 policy-pap | [2024-07-04T09:42:59.359+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-pdp-pap] no listeners for autonomous message of type PdpStatus 09:44:48 policy-pap | [2024-07-04T09:42:59.453+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate starting 09:44:48 policy-pap | [2024-07-04T09:42:59.454+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate starting listener 09:44:48 policy-pap | [2024-07-04T09:42:59.454+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate starting timer 09:44:48 policy-pap | [2024-07-04T09:42:59.455+00:00|INFO|TimerManager|KAFKA-source-policy-heartbeat] update timer registered Timer [name=83f05f71-f951-4e81-8af2-344adb821102, expireMs=1720086209454] 09:44:48 policy-pap | [2024-07-04T09:42:59.456+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate starting enqueue 09:44:48 policy-pap | [2024-07-04T09:42:59.457+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate started 09:44:48 policy-pap | [2024-07-04T09:42:59.457+00:00|INFO|TimerManager|Thread-9] update timer waiting 29997ms Timer [name=83f05f71-f951-4e81-8af2-344adb821102, expireMs=1720086209454] 09:44:48 policy-pap | [2024-07-04T09:42:59.463+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"messageName":"PDP_UPDATE","requestId":"83f05f71-f951-4e81-8af2-344adb821102","timestampMs":1720086179433,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.502+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] 09:44:48 policy-pap | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"messageName":"PDP_UPDATE","requestId":"83f05f71-f951-4e81-8af2-344adb821102","timestampMs":1720086179433,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.504+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_UPDATE 09:44:48 policy-pap | [2024-07-04T09:42:59.507+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"messageName":"PDP_UPDATE","requestId":"83f05f71-f951-4e81-8af2-344adb821102","timestampMs":1720086179433,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.507+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_UPDATE 09:44:48 policy-pap | [2024-07-04T09:42:59.520+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] 09:44:48 policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"03bc4065-c817-4394-9c0d-26791ee3b6ff","timestampMs":1720086179510,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup"} 09:44:48 policy-pap | [2024-07-04T09:42:59.525+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"03bc4065-c817-4394-9c0d-26791ee3b6ff","timestampMs":1720086179510,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup"} 09:44:48 policy-pap | [2024-07-04T09:42:59.526+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-pdp-pap] no listeners for autonomous message of type PdpStatus 09:44:48 policy-pap | [2024-07-04T09:42:59.530+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"83f05f71-f951-4e81-8af2-344adb821102","responseStatus":"SUCCESS","responseMessage":"Pdp update successful."},"messageName":"PDP_STATUS","requestId":"8cf42592-1b3a-46c7-8ffb-56547c206e64","timestampMs":1720086179511,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.556+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopping 09:44:48 policy-pap | [2024-07-04T09:42:59.556+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopping enqueue 09:44:48 policy-pap | [2024-07-04T09:42:59.557+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopping timer 09:44:48 policy-pap | [2024-07-04T09:42:59.557+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer cancelled Timer [name=83f05f71-f951-4e81-8af2-344adb821102, expireMs=1720086209454] 09:44:48 policy-pap | [2024-07-04T09:42:59.557+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopping listener 09:44:48 policy-pap | [2024-07-04T09:42:59.557+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopped 09:44:48 policy-pap | [2024-07-04T09:42:59.561+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] 09:44:48 policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"83f05f71-f951-4e81-8af2-344adb821102","responseStatus":"SUCCESS","responseMessage":"Pdp update successful."},"messageName":"PDP_STATUS","requestId":"8cf42592-1b3a-46c7-8ffb-56547c206e64","timestampMs":1720086179511,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.562+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id 83f05f71-f951-4e81-8af2-344adb821102 09:44:48 policy-pap | [2024-07-04T09:42:59.563+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate successful 09:44:48 policy-pap | [2024-07-04T09:42:59.564+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 start publishing next request 09:44:48 policy-pap | [2024-07-04T09:42:59.564+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange starting 09:44:48 policy-pap | [2024-07-04T09:42:59.564+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange starting listener 09:44:48 policy-pap | [2024-07-04T09:42:59.564+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange starting timer 09:44:48 policy-pap | [2024-07-04T09:42:59.564+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] state-change timer registered Timer [name=bc1c9506-3f24-4c95-b519-1c7dbca5b7b0, expireMs=1720086209564] 09:44:48 policy-pap | [2024-07-04T09:42:59.564+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange starting enqueue 09:44:48 policy-pap | [2024-07-04T09:42:59.564+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange started 09:44:48 policy-pap | [2024-07-04T09:42:59.564+00:00|INFO|TimerManager|Thread-10] state-change timer waiting 30000ms Timer [name=bc1c9506-3f24-4c95-b519-1c7dbca5b7b0, expireMs=1720086209564] 09:44:48 policy-pap | [2024-07-04T09:42:59.564+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"bc1c9506-3f24-4c95-b519-1c7dbca5b7b0","timestampMs":1720086179434,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.579+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] 09:44:48 policy-pap | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"bc1c9506-3f24-4c95-b519-1c7dbca5b7b0","timestampMs":1720086179434,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.580+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_STATE_CHANGE 09:44:48 policy-pap | [2024-07-04T09:42:59.584+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] 09:44:48 policy-pap | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpStateChange","policies":[],"response":{"responseTo":"bc1c9506-3f24-4c95-b519-1c7dbca5b7b0","responseStatus":"SUCCESS","responseMessage":"State changed to active. No policies found."},"messageName":"PDP_STATUS","requestId":"c6eccbf2-6525-456b-98c1-d9ba1f8088ea","timestampMs":1720086179574,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.585+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id bc1c9506-3f24-4c95-b519-1c7dbca5b7b0 09:44:48 policy-pap | [2024-07-04T09:42:59.603+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"bc1c9506-3f24-4c95-b519-1c7dbca5b7b0","timestampMs":1720086179434,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.604+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATE_CHANGE 09:44:48 policy-pap | [2024-07-04T09:42:59.608+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpStateChange","policies":[],"response":{"responseTo":"bc1c9506-3f24-4c95-b519-1c7dbca5b7b0","responseStatus":"SUCCESS","responseMessage":"State changed to active. No policies found."},"messageName":"PDP_STATUS","requestId":"c6eccbf2-6525-456b-98c1-d9ba1f8088ea","timestampMs":1720086179574,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.608+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange stopping 09:44:48 policy-pap | [2024-07-04T09:42:59.608+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange stopping enqueue 09:44:48 policy-pap | [2024-07-04T09:42:59.608+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange stopping timer 09:44:48 policy-pap | [2024-07-04T09:42:59.608+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] state-change timer cancelled Timer [name=bc1c9506-3f24-4c95-b519-1c7dbca5b7b0, expireMs=1720086209564] 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange stopping listener 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange stopped 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpStateChange successful 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 start publishing next request 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate starting 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate starting listener 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate starting timer 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer registered Timer [name=68fa5061-72d2-4f2d-ba41-4a7727b3c604, expireMs=1720086209609] 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate starting enqueue 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate started 09:44:48 policy-pap | [2024-07-04T09:42:59.609+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"68fa5061-72d2-4f2d-ba41-4a7727b3c604","timestampMs":1720086179596,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.620+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"68fa5061-72d2-4f2d-ba41-4a7727b3c604","timestampMs":1720086179596,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.620+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_UPDATE 09:44:48 policy-pap | [2024-07-04T09:42:59.623+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] 09:44:48 policy-pap | {"source":"pap-ac7da8c6-cb5e-4ff7-b0d3-615c2f4ae3ce","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"68fa5061-72d2-4f2d-ba41-4a7727b3c604","timestampMs":1720086179596,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.624+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_UPDATE 09:44:48 policy-pap | [2024-07-04T09:42:59.628+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] 09:44:48 policy-pap | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"68fa5061-72d2-4f2d-ba41-4a7727b3c604","responseStatus":"SUCCESS","responseMessage":"Pdp already updated"},"messageName":"PDP_STATUS","requestId":"c6d6d97d-4b4b-476f-966f-02ef731a732f","timestampMs":1720086179616,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.628+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] 09:44:48 policy-pap | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"68fa5061-72d2-4f2d-ba41-4a7727b3c604","responseStatus":"SUCCESS","responseMessage":"Pdp already updated"},"messageName":"PDP_STATUS","requestId":"c6d6d97d-4b4b-476f-966f-02ef731a732f","timestampMs":1720086179616,"name":"apex-84299ad1-3ac8-4023-82b2-31178178cc71","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} 09:44:48 policy-pap | [2024-07-04T09:42:59.629+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopping 09:44:48 policy-pap | [2024-07-04T09:42:59.629+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopping enqueue 09:44:48 policy-pap | [2024-07-04T09:42:59.629+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopping timer 09:44:48 policy-pap | [2024-07-04T09:42:59.629+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer cancelled Timer [name=68fa5061-72d2-4f2d-ba41-4a7727b3c604, expireMs=1720086209609] 09:44:48 policy-pap | [2024-07-04T09:42:59.629+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopping listener 09:44:48 policy-pap | [2024-07-04T09:42:59.629+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate stopped 09:44:48 policy-pap | [2024-07-04T09:42:59.629+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id 68fa5061-72d2-4f2d-ba41-4a7727b3c604 09:44:48 policy-pap | [2024-07-04T09:42:59.632+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 PdpUpdate successful 09:44:48 policy-pap | [2024-07-04T09:42:59.632+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] apex-84299ad1-3ac8-4023-82b2-31178178cc71 has no more requests 09:44:48 policy-pap | [2024-07-04T09:43:29.454+00:00|INFO|TimerManager|Thread-9] update timer discarded (expired) Timer [name=83f05f71-f951-4e81-8af2-344adb821102, expireMs=1720086209454] 09:44:48 policy-pap | [2024-07-04T09:43:29.565+00:00|INFO|TimerManager|Thread-10] state-change timer discarded (expired) Timer [name=bc1c9506-3f24-4c95-b519-1c7dbca5b7b0, expireMs=1720086209564] 09:44:48 policy-pap | [2024-07-04T09:44:37.510+00:00|INFO|PdpModifyRequestMap|pool-3-thread-1] check for PDP records older than 360000ms 09:44:48 =================================== 09:44:48 ======== Logs from prometheus ======== 09:44:48 prometheus | ts=2024-07-04T09:41:57.513Z caller=main.go:589 level=info msg="No time or size retention was set so using the default time retention" duration=15d 09:44:48 prometheus | ts=2024-07-04T09:41:57.513Z caller=main.go:633 level=info msg="Starting Prometheus Server" mode=server version="(version=2.53.0, branch=HEAD, revision=4c35b9250afefede41c5f5acd76191f90f625898)" 09:44:48 prometheus | ts=2024-07-04T09:41:57.513Z caller=main.go:638 level=info build_context="(go=go1.22.4, platform=linux/amd64, user=root@7f8d89cbbd64, date=20240619-07:39:12, tags=netgo,builtinassets,stringlabels)" 09:44:48 prometheus | ts=2024-07-04T09:41:57.513Z caller=main.go:639 level=info host_details="(Linux 4.15.0-192-generic #203-Ubuntu SMP Wed Aug 10 17:40:03 UTC 2022 x86_64 prometheus (none))" 09:44:48 prometheus | ts=2024-07-04T09:41:57.513Z caller=main.go:640 level=info fd_limits="(soft=1048576, hard=1048576)" 09:44:48 prometheus | ts=2024-07-04T09:41:57.513Z caller=main.go:641 level=info vm_limits="(soft=unlimited, hard=unlimited)" 09:44:48 prometheus | ts=2024-07-04T09:41:57.523Z caller=web.go:568 level=info component=web msg="Start listening for connections" address=0.0.0.0:9090 09:44:48 prometheus | ts=2024-07-04T09:41:57.535Z caller=tls_config.go:313 level=info component=web msg="Listening on" address=[::]:9090 09:44:48 prometheus | ts=2024-07-04T09:41:57.535Z caller=tls_config.go:316 level=info component=web msg="TLS is disabled." http2=false address=[::]:9090 09:44:48 prometheus | ts=2024-07-04T09:41:57.531Z caller=main.go:1148 level=info msg="Starting TSDB ..." 09:44:48 prometheus | ts=2024-07-04T09:41:57.549Z caller=head.go:626 level=info component=tsdb msg="Replaying on-disk memory mappable chunks if any" 09:44:48 prometheus | ts=2024-07-04T09:41:57.549Z caller=head.go:713 level=info component=tsdb msg="On-disk memory mappable chunks replay completed" duration=2.09µs 09:44:48 prometheus | ts=2024-07-04T09:41:57.549Z caller=head.go:721 level=info component=tsdb msg="Replaying WAL, this may take a while" 09:44:48 prometheus | ts=2024-07-04T09:41:57.550Z caller=head.go:793 level=info component=tsdb msg="WAL segment loaded" segment=0 maxSegment=0 09:44:48 prometheus | ts=2024-07-04T09:41:57.550Z caller=head.go:830 level=info component=tsdb msg="WAL replay completed" checkpoint_replay_duration=115.011µs wal_replay_duration=336.314µs wbl_replay_duration=220ns chunk_snapshot_load_duration=0s mmap_chunk_replay_duration=2.09µs total_replay_duration=663.718µs 09:44:48 prometheus | ts=2024-07-04T09:41:57.553Z caller=main.go:1169 level=info fs_type=EXT4_SUPER_MAGIC 09:44:48 prometheus | ts=2024-07-04T09:41:57.553Z caller=main.go:1172 level=info msg="TSDB started" 09:44:48 prometheus | ts=2024-07-04T09:41:57.553Z caller=main.go:1354 level=info msg="Loading configuration file" filename=/etc/prometheus/prometheus.yml 09:44:48 prometheus | ts=2024-07-04T09:41:57.554Z caller=main.go:1391 level=info msg="updated GOGC" old=100 new=75 09:44:48 prometheus | ts=2024-07-04T09:41:57.554Z caller=main.go:1402 level=info msg="Completed loading of configuration file" filename=/etc/prometheus/prometheus.yml totalDuration=1.253575ms db_storage=1.3µs remote_storage=1.43µs web_handler=600ns query_engine=1.14µs scrape=385.035µs scrape_sd=139.872µs notify=23.47µs notify_sd=12.34µs rules=1.65µs tracing=42.971µs 09:44:48 prometheus | ts=2024-07-04T09:41:57.554Z caller=main.go:1133 level=info msg="Server is ready to receive web requests." 09:44:48 prometheus | ts=2024-07-04T09:41:57.554Z caller=manager.go:164 level=info component="rule manager" msg="Starting rule manager..." 09:44:48 =================================== 09:44:48 ======== Logs from simulator ======== 09:44:48 simulator | Policy simulator config file: /opt/app/policy/simulators/etc/mounted/simParameters.json 09:44:48 simulator | overriding logback.xml 09:44:48 simulator | 2024-07-04 09:41:56,799 INFO replacing 'HOST_NAME' with simulator in /opt/app/policy/simulators/etc/mounted/simParameters.json 09:44:48 simulator | 2024-07-04 09:41:56,905 INFO org.onap.policy.models.simulators starting 09:44:48 simulator | 2024-07-04 09:41:56,905 INFO org.onap.policy.models.simulators starting CDS gRPC Server Properties 09:44:48 simulator | 2024-07-04 09:41:57,079 INFO org.onap.policy.models.simulators starting org.onap.policy.simulators.AaiSimulatorJaxRs_RESOURCE_LOCATION 09:44:48 simulator | 2024-07-04 09:41:57,080 INFO org.onap.policy.models.simulators starting A&AI simulator 09:44:48 simulator | 2024-07-04 09:41:57,184 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=A&AI simulator, host=0.0.0.0, port=6666, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@64a8c844{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@3f6db3fb{/,null,STOPPED}, connector=A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: WAITED-START 09:44:48 simulator | 2024-07-04 09:41:57,203 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=A&AI simulator, host=0.0.0.0, port=6666, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@64a8c844{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@3f6db3fb{/,null,STOPPED}, connector=A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 simulator | 2024-07-04 09:41:57,206 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=A&AI simulator, host=0.0.0.0, port=6666, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@64a8c844{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@3f6db3fb{/,null,STOPPED}, connector=A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666}, jettyThread=Thread[A&AI simulator-6666,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 simulator | 2024-07-04 09:41:57,212 INFO jetty-11.0.20; built: 2024-01-29T21:04:22.394Z; git: 922f8dc188f7011e60d0361de585fd4ac4d63064; jvm 17.0.11+9-alpine-r0 09:44:48 simulator | 2024-07-04 09:41:57,292 INFO Session workerName=node0 09:44:48 simulator | 2024-07-04 09:41:57,877 INFO Using GSON for REST calls 09:44:48 simulator | 2024-07-04 09:41:57,984 INFO Started o.e.j.s.ServletContextHandler@3f6db3fb{/,null,AVAILABLE} 09:44:48 simulator | 2024-07-04 09:41:57,990 INFO Started A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666} 09:44:48 simulator | 2024-07-04 09:41:57,997 INFO Started Server@64a8c844{STARTING}[11.0.20,sto=0] @1668ms 09:44:48 simulator | 2024-07-04 09:41:57,997 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}}, swaggerId=null, toString()=JettyServletServer(name=A&AI simulator, host=0.0.0.0, port=6666, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@64a8c844{STARTED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@3f6db3fb{/,null,AVAILABLE}, connector=A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666}, jettyThread=Thread[A&AI simulator-6666,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}})]: pending time is 4208 ms. 09:44:48 simulator | 2024-07-04 09:41:58,010 INFO org.onap.policy.models.simulators starting SDNC simulator 09:44:48 simulator | 2024-07-04 09:41:58,017 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SDNC simulator, host=0.0.0.0, port=6668, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@70efb718{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@b70da4c{/,null,STOPPED}, connector=SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: WAITED-START 09:44:48 simulator | 2024-07-04 09:41:58,019 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SDNC simulator, host=0.0.0.0, port=6668, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@70efb718{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@b70da4c{/,null,STOPPED}, connector=SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 simulator | 2024-07-04 09:41:58,023 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SDNC simulator, host=0.0.0.0, port=6668, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@70efb718{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@b70da4c{/,null,STOPPED}, connector=SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668}, jettyThread=Thread[SDNC simulator-6668,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 simulator | 2024-07-04 09:41:58,024 INFO jetty-11.0.20; built: 2024-01-29T21:04:22.394Z; git: 922f8dc188f7011e60d0361de585fd4ac4d63064; jvm 17.0.11+9-alpine-r0 09:44:48 simulator | 2024-07-04 09:41:58,030 INFO Session workerName=node0 09:44:48 simulator | 2024-07-04 09:41:58,083 INFO Using GSON for REST calls 09:44:48 simulator | 2024-07-04 09:41:58,092 INFO Started o.e.j.s.ServletContextHandler@b70da4c{/,null,AVAILABLE} 09:44:48 simulator | 2024-07-04 09:41:58,093 INFO Started SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668} 09:44:48 simulator | 2024-07-04 09:41:58,093 INFO Started Server@70efb718{STARTING}[11.0.20,sto=0] @1765ms 09:44:48 simulator | 2024-07-04 09:41:58,094 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}}, swaggerId=null, toString()=JettyServletServer(name=SDNC simulator, host=0.0.0.0, port=6668, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@70efb718{STARTED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@b70da4c{/,null,AVAILABLE}, connector=SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668}, jettyThread=Thread[SDNC simulator-6668,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}})]: pending time is 4930 ms. 09:44:48 simulator | 2024-07-04 09:41:58,095 INFO org.onap.policy.models.simulators starting SO simulator 09:44:48 simulator | 2024-07-04 09:41:58,099 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SO simulator, host=0.0.0.0, port=6669, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@b7838a9{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@696f0212{/,null,STOPPED}, connector=SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: WAITED-START 09:44:48 simulator | 2024-07-04 09:41:58,100 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SO simulator, host=0.0.0.0, port=6669, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@b7838a9{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@696f0212{/,null,STOPPED}, connector=SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 simulator | 2024-07-04 09:41:58,101 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SO simulator, host=0.0.0.0, port=6669, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@b7838a9{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@696f0212{/,null,STOPPED}, connector=SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669}, jettyThread=Thread[SO simulator-6669,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 simulator | 2024-07-04 09:41:58,101 INFO jetty-11.0.20; built: 2024-01-29T21:04:22.394Z; git: 922f8dc188f7011e60d0361de585fd4ac4d63064; jvm 17.0.11+9-alpine-r0 09:44:48 simulator | 2024-07-04 09:41:58,104 INFO Session workerName=node0 09:44:48 simulator | 2024-07-04 09:41:58,175 INFO Using GSON for REST calls 09:44:48 simulator | 2024-07-04 09:41:58,187 INFO Started o.e.j.s.ServletContextHandler@696f0212{/,null,AVAILABLE} 09:44:48 simulator | 2024-07-04 09:41:58,188 INFO Started SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669} 09:44:48 simulator | 2024-07-04 09:41:58,188 INFO Started Server@b7838a9{STARTING}[11.0.20,sto=0] @1860ms 09:44:48 simulator | 2024-07-04 09:41:58,188 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}}, swaggerId=null, toString()=JettyServletServer(name=SO simulator, host=0.0.0.0, port=6669, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@b7838a9{STARTED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@696f0212{/,null,AVAILABLE}, connector=SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669}, jettyThread=Thread[SO simulator-6669,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}})]: pending time is 4912 ms. 09:44:48 simulator | 2024-07-04 09:41:58,189 INFO org.onap.policy.models.simulators starting VFC simulator 09:44:48 simulator | 2024-07-04 09:41:58,192 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=VFC simulator, host=0.0.0.0, port=6670, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@f478a81{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@19553973{/,null,STOPPED}, connector=VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: WAITED-START 09:44:48 simulator | 2024-07-04 09:41:58,193 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=VFC simulator, host=0.0.0.0, port=6670, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@f478a81{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@19553973{/,null,STOPPED}, connector=VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 simulator | 2024-07-04 09:41:58,194 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=VFC simulator, host=0.0.0.0, port=6670, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@f478a81{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@19553973{/,null,STOPPED}, connector=VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670}, jettyThread=Thread[VFC simulator-6670,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING 09:44:48 simulator | 2024-07-04 09:41:58,195 INFO jetty-11.0.20; built: 2024-01-29T21:04:22.394Z; git: 922f8dc188f7011e60d0361de585fd4ac4d63064; jvm 17.0.11+9-alpine-r0 09:44:48 simulator | 2024-07-04 09:41:58,199 INFO Session workerName=node0 09:44:48 simulator | 2024-07-04 09:41:58,236 INFO Using GSON for REST calls 09:44:48 simulator | 2024-07-04 09:41:58,244 INFO Started o.e.j.s.ServletContextHandler@19553973{/,null,AVAILABLE} 09:44:48 simulator | 2024-07-04 09:41:58,245 INFO Started VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670} 09:44:48 simulator | 2024-07-04 09:41:58,245 INFO Started Server@f478a81{STARTING}[11.0.20,sto=0] @1917ms 09:44:48 simulator | 2024-07-04 09:41:58,245 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}}, swaggerId=null, toString()=JettyServletServer(name=VFC simulator, host=0.0.0.0, port=6670, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@f478a81{STARTED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@19553973{/,null,AVAILABLE}, connector=VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670}, jettyThread=Thread[VFC simulator-6670,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}})]: pending time is 4949 ms. 09:44:48 simulator | 2024-07-04 09:41:58,246 INFO org.onap.policy.models.simulators started 09:44:48 =================================== 09:44:48 ======== Logs from zookeeper ======== 09:44:48 zookeeper | ===> User 09:44:48 zookeeper | uid=1000(appuser) gid=1000(appuser) groups=1000(appuser) 09:44:48 zookeeper | ===> Configuring ... 09:44:48 zookeeper | ===> Running preflight checks ... 09:44:48 zookeeper | ===> Check if /var/lib/zookeeper/data is writable ... 09:44:48 zookeeper | ===> Check if /var/lib/zookeeper/log is writable ... 09:44:48 zookeeper | ===> Launching ... 09:44:48 zookeeper | ===> Launching zookeeper ... 09:44:48 zookeeper | [2024-07-04 09:41:55,960] INFO Reading configuration from: /etc/kafka/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,967] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,967] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,967] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,967] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,969] INFO autopurge.snapRetainCount set to 3 (org.apache.zookeeper.server.DatadirCleanupManager) 09:44:48 zookeeper | [2024-07-04 09:41:55,969] INFO autopurge.purgeInterval set to 0 (org.apache.zookeeper.server.DatadirCleanupManager) 09:44:48 zookeeper | [2024-07-04 09:41:55,969] INFO Purge task is not scheduled. (org.apache.zookeeper.server.DatadirCleanupManager) 09:44:48 zookeeper | [2024-07-04 09:41:55,969] WARN Either no config or no quorum defined in config, running in standalone mode (org.apache.zookeeper.server.quorum.QuorumPeerMain) 09:44:48 zookeeper | [2024-07-04 09:41:55,970] INFO Log4j 1.2 jmx support not found; jmx disabled. (org.apache.zookeeper.jmx.ManagedUtil) 09:44:48 zookeeper | [2024-07-04 09:41:55,970] INFO Reading configuration from: /etc/kafka/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,971] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,971] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,971] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,971] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) 09:44:48 zookeeper | [2024-07-04 09:41:55,971] INFO Starting server (org.apache.zookeeper.server.ZooKeeperServerMain) 09:44:48 zookeeper | [2024-07-04 09:41:55,983] INFO ServerMetrics initialized with provider org.apache.zookeeper.metrics.impl.DefaultMetricsProvider@77eca502 (org.apache.zookeeper.server.ServerMetrics) 09:44:48 zookeeper | [2024-07-04 09:41:55,985] INFO ACL digest algorithm is: SHA1 (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) 09:44:48 zookeeper | [2024-07-04 09:41:55,985] INFO zookeeper.DigestAuthenticationProvider.enabled = true (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) 09:44:48 zookeeper | [2024-07-04 09:41:55,987] INFO zookeeper.snapshot.trust.empty : false (org.apache.zookeeper.server.persistence.FileTxnSnapLog) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO ______ _ (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO |___ / | | (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO / / ___ ___ | | __ ___ ___ _ __ ___ _ __ (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO / / / _ \ / _ \ | |/ / / _ \ / _ \ | '_ \ / _ \ | '__| (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO / /__ | (_) | | (_) | | < | __/ | __/ | |_) | | __/ | | (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO /_____| \___/ \___/ |_|\_\ \___| \___| | .__/ \___| |_| (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO | | (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO |_| (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,996] INFO (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:zookeeper.version=3.8.4-9316c2a7a97e1666d8f4593f34dd6fc36ecc436c, built on 2024-02-12 22:16 UTC (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:host.name=zookeeper (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:java.version=11.0.22 (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:java.vendor=Azul Systems, Inc. (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:java.home=/usr/lib/jvm/java-11-zulu-openjdk-ca (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:java.class.path=/usr/bin/../share/java/kafka/scala-logging_2.13-3.9.4.jar:/usr/bin/../share/java/kafka/jersey-common-2.39.1.jar:/usr/bin/../share/java/kafka/swagger-annotations-2.2.8.jar:/usr/bin/../share/java/kafka/connect-json-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-storage-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/commons-validator-1.7.jar:/usr/bin/../share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/bin/../share/java/kafka/aopalliance-repackaged-2.6.1.jar:/usr/bin/../share/java/kafka/netty-transport-4.1.100.Final.jar:/usr/bin/../share/java/kafka/rocksdbjni-7.9.2.jar:/usr/bin/../share/java/kafka/jetty-http-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka-log4j-appender-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/commons-io-2.11.0.jar:/usr/bin/../share/java/kafka/kafka-clients-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/javax.activation-api-1.2.0.jar:/usr/bin/../share/java/kafka/jetty-security-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/commons-cli-1.4.jar:/usr/bin/../share/java/kafka/jetty-server-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/slf4j-reload4j-1.7.36.jar:/usr/bin/../share/java/kafka/scala-reflect-2.13.11.jar:/usr/bin/../share/java/kafka/kafka-server-common-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-group-coordinator-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-buffer-4.1.100.Final.jar:/usr/bin/../share/java/kafka/connect-runtime-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/trogdor-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.ws.rs-api-2.1.6.jar:/usr/bin/../share/java/kafka/jakarta.annotation-api-1.3.5.jar:/usr/bin/../share/java/kafka/kafka-streams-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/scala-java8-compat_2.13-1.0.2.jar:/usr/bin/../share/java/kafka/javax.ws.rs-api-2.1.1.jar:/usr/bin/../share/java/kafka/zookeeper-jute-3.8.4.jar:/usr/bin/../share/java/kafka/netty-resolver-4.1.100.Final.jar:/usr/bin/../share/java/kafka/hk2-api-2.6.1.jar:/usr/bin/../share/java/kafka/kafka-storage-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/zstd-jni-1.5.5-1.jar:/usr/bin/../share/java/kafka/jackson-dataformat-csv-2.13.5.jar:/usr/bin/../share/java/kafka/kafka.jar:/usr/bin/../share/java/kafka/scala-library-2.13.11.jar:/usr/bin/../share/java/kafka/jakarta.inject-2.6.1.jar:/usr/bin/../share/java/kafka/jakarta.xml.bind-api-2.3.3.jar:/usr/bin/../share/java/kafka/jetty-continuation-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/connect-transforms-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jose4j-0.9.4.jar:/usr/bin/../share/java/kafka/hk2-locator-2.6.1.jar:/usr/bin/../share/java/kafka/reflections-0.10.2.jar:/usr/bin/../share/java/kafka/slf4j-api-1.7.36.jar:/usr/bin/../share/java/kafka/paranamer-2.8.jar:/usr/bin/../share/java/kafka/commons-beanutils-1.9.4.jar:/usr/bin/../share/java/kafka/jaxb-api-2.3.1.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-2.39.1.jar:/usr/bin/../share/java/kafka/hk2-utils-2.6.1.jar:/usr/bin/../share/java/kafka/jackson-module-scala_2.13-2.13.5.jar:/usr/bin/../share/java/kafka/reload4j-1.2.25.jar:/usr/bin/../share/java/kafka/connect-mirror-client-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-servlet-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jackson-core-2.13.5.jar:/usr/bin/../share/java/kafka/jersey-hk2-2.39.1.jar:/usr/bin/../share/java/kafka/jackson-databind-2.13.5.jar:/usr/bin/../share/java/kafka/jetty-servlets-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jersey-client-2.39.1.jar:/usr/bin/../share/java/kafka/osgi-resource-locator-1.0.3.jar:/usr/bin/../share/java/kafka/commons-digester-2.1.jar:/usr/bin/../share/java/kafka/netty-transport-native-epoll-4.1.100.Final.jar:/usr/bin/../share/java/kafka/argparse4j-0.7.0.jar:/usr/bin/../share/java/kafka/connect-mirror-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-datatype-jdk8-2.13.5.jar:/usr/bin/../share/java/kafka/audience-annotations-0.12.0.jar:/usr/bin/../share/java/kafka/jackson-module-jaxb-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/kafka-raft-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/javax.annotation-api-1.3.2.jar:/usr/bin/../share/java/kafka/maven-artifact-3.8.8.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-json-provider-2.13.5.jar:/usr/bin/../share/java/kafka/jakarta.validation-api-2.0.2.jar:/usr/bin/../share/java/kafka/zookeeper-3.8.4.jar:/usr/bin/../share/java/kafka/jersey-server-2.39.1.jar:/usr/bin/../share/java/kafka/commons-lang3-3.8.1.jar:/usr/bin/../share/java/kafka/jopt-simple-5.0.4.jar:/usr/bin/../share/java/kafka/error_prone_annotations-2.10.0.jar:/usr/bin/../share/java/kafka/lz4-java-1.8.0.jar:/usr/bin/../share/java/kafka/jetty-util-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka-tools-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.activation-api-1.2.2.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-core-2.39.1.jar:/usr/bin/../share/java/kafka/checker-qual-3.19.0.jar:/usr/bin/../share/java/kafka/kafka-metadata-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/pcollections-4.0.1.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-base-2.13.5.jar:/usr/bin/../share/java/kafka/commons-logging-1.2.jar:/usr/bin/../share/java/kafka/jsr305-3.0.2.jar:/usr/bin/../share/java/kafka/jetty-io-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/netty-codec-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-transport-native-unix-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/scala-collection-compat_2.13-2.10.0.jar:/usr/bin/../share/java/kafka/metrics-core-2.2.0.jar:/usr/bin/../share/java/kafka/kafka-streams-test-utils-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-streams-examples-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-handler-4.1.100.Final.jar:/usr/bin/../share/java/kafka/commons-collections-3.2.2.jar:/usr/bin/../share/java/kafka/javassist-3.29.2-GA.jar:/usr/bin/../share/java/kafka/caffeine-2.9.3.jar:/usr/bin/../share/java/kafka/plexus-utils-3.3.1.jar:/usr/bin/../share/java/kafka/kafka-tools-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/activation-1.1.1.jar:/usr/bin/../share/java/kafka/netty-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/kafka-streams-scala_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/metrics-core-4.1.12.1.jar:/usr/bin/../share/java/kafka/jline-3.25.1.jar:/usr/bin/../share/java/kafka/jetty-client-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-basic-auth-extension-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-util-ajax-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/snappy-java-1.1.10.5.jar:/usr/bin/../share/java/kafka/kafka-shell-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-transport-classes-epoll-4.1.100.Final.jar:/usr/bin/../share/java/confluent-telemetry/* (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:java.io.tmpdir=/tmp (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:java.compiler= (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:os.name=Linux (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:os.arch=amd64 (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:os.version=4.15.0-192-generic (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:user.name=appuser (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:user.home=/home/appuser (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:user.dir=/home/appuser (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:os.memory.free=491MB (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:os.memory.max=512MB (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO Server environment:os.memory.total=512MB (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO zookeeper.enableEagerACLCheck = false (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,997] INFO zookeeper.digest.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,998] INFO zookeeper.closeSessionTxn.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,998] INFO zookeeper.flushDelay = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,998] INFO zookeeper.maxWriteQueuePollTime = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,998] INFO zookeeper.maxBatchSize=1000 (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,998] INFO zookeeper.intBufferStartingSizeBytes = 1024 (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,999] INFO Weighed connection throttling is disabled (org.apache.zookeeper.server.BlueThrottle) 09:44:48 zookeeper | [2024-07-04 09:41:55,999] INFO minSessionTimeout set to 6000 ms (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:55,999] INFO maxSessionTimeout set to 60000 ms (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:56,000] INFO getData response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) 09:44:48 zookeeper | [2024-07-04 09:41:56,000] INFO getChildren response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) 09:44:48 zookeeper | [2024-07-04 09:41:56,001] INFO zookeeper.pathStats.slotCapacity = 60 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) 09:44:48 zookeeper | [2024-07-04 09:41:56,001] INFO zookeeper.pathStats.slotDuration = 15 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) 09:44:48 zookeeper | [2024-07-04 09:41:56,001] INFO zookeeper.pathStats.maxDepth = 6 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) 09:44:48 zookeeper | [2024-07-04 09:41:56,001] INFO zookeeper.pathStats.initialDelay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) 09:44:48 zookeeper | [2024-07-04 09:41:56,001] INFO zookeeper.pathStats.delay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) 09:44:48 zookeeper | [2024-07-04 09:41:56,001] INFO zookeeper.pathStats.enabled = false (org.apache.zookeeper.server.util.RequestPathMetricsCollector) 09:44:48 zookeeper | [2024-07-04 09:41:56,003] INFO The max bytes for all large requests are set to 104857600 (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:56,003] INFO The large request threshold is set to -1 (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:56,004] INFO zookeeper.enforce.auth.enabled = false (org.apache.zookeeper.server.AuthenticationHelper) 09:44:48 zookeeper | [2024-07-04 09:41:56,004] INFO zookeeper.enforce.auth.schemes = [] (org.apache.zookeeper.server.AuthenticationHelper) 09:44:48 zookeeper | [2024-07-04 09:41:56,004] INFO Created server with tickTime 3000 ms minSessionTimeout 6000 ms maxSessionTimeout 60000 ms clientPortListenBacklog -1 datadir /var/lib/zookeeper/log/version-2 snapdir /var/lib/zookeeper/data/version-2 (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:56,034] INFO Logging initialized @544ms to org.eclipse.jetty.util.log.Slf4jLog (org.eclipse.jetty.util.log) 09:44:48 zookeeper | [2024-07-04 09:41:56,137] WARN o.e.j.s.ServletContextHandler@6d5620ce{/,null,STOPPED} contextPath ends with /* (org.eclipse.jetty.server.handler.ContextHandler) 09:44:48 zookeeper | [2024-07-04 09:41:56,137] WARN Empty contextPath (org.eclipse.jetty.server.handler.ContextHandler) 09:44:48 zookeeper | [2024-07-04 09:41:56,162] INFO jetty-9.4.54.v20240208; built: 2024-02-08T19:42:39.027Z; git: cef3fbd6d736a21e7d541a5db490381d95a2047d; jvm 11.0.22+7-LTS (org.eclipse.jetty.server.Server) 09:44:48 zookeeper | [2024-07-04 09:41:56,192] INFO DefaultSessionIdManager workerName=node0 (org.eclipse.jetty.server.session) 09:44:48 zookeeper | [2024-07-04 09:41:56,192] INFO No SessionScavenger set, using defaults (org.eclipse.jetty.server.session) 09:44:48 zookeeper | [2024-07-04 09:41:56,194] INFO node0 Scavenging every 600000ms (org.eclipse.jetty.server.session) 09:44:48 zookeeper | [2024-07-04 09:41:56,197] WARN ServletContext@o.e.j.s.ServletContextHandler@6d5620ce{/,null,STARTING} has uncovered http methods for path: /* (org.eclipse.jetty.security.SecurityHandler) 09:44:48 zookeeper | [2024-07-04 09:41:56,205] INFO Started o.e.j.s.ServletContextHandler@6d5620ce{/,null,AVAILABLE} (org.eclipse.jetty.server.handler.ContextHandler) 09:44:48 zookeeper | [2024-07-04 09:41:56,219] INFO Started ServerConnector@4d1bf319{HTTP/1.1, (http/1.1)}{0.0.0.0:8080} (org.eclipse.jetty.server.AbstractConnector) 09:44:48 zookeeper | [2024-07-04 09:41:56,219] INFO Started @730ms (org.eclipse.jetty.server.Server) 09:44:48 zookeeper | [2024-07-04 09:41:56,219] INFO Started AdminServer on address 0.0.0.0, port 8080 and command URL /commands (org.apache.zookeeper.server.admin.JettyAdminServer) 09:44:48 zookeeper | [2024-07-04 09:41:56,223] INFO Using org.apache.zookeeper.server.NIOServerCnxnFactory as server connection factory (org.apache.zookeeper.server.ServerCnxnFactory) 09:44:48 zookeeper | [2024-07-04 09:41:56,224] WARN maxCnxns is not configured, using default value 0. (org.apache.zookeeper.server.ServerCnxnFactory) 09:44:48 zookeeper | [2024-07-04 09:41:56,225] INFO Configuring NIO connection handler with 10s sessionless connection timeout, 2 selector thread(s), 16 worker threads, and 64 kB direct buffers. (org.apache.zookeeper.server.NIOServerCnxnFactory) 09:44:48 zookeeper | [2024-07-04 09:41:56,226] INFO binding to port 0.0.0.0/0.0.0.0:2181 (org.apache.zookeeper.server.NIOServerCnxnFactory) 09:44:48 zookeeper | [2024-07-04 09:41:56,242] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) 09:44:48 zookeeper | [2024-07-04 09:41:56,242] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) 09:44:48 zookeeper | [2024-07-04 09:41:56,243] INFO zookeeper.snapshotSizeFactor = 0.33 (org.apache.zookeeper.server.ZKDatabase) 09:44:48 zookeeper | [2024-07-04 09:41:56,243] INFO zookeeper.commitLogCount=500 (org.apache.zookeeper.server.ZKDatabase) 09:44:48 zookeeper | [2024-07-04 09:41:56,247] INFO zookeeper.snapshot.compression.method = CHECKED (org.apache.zookeeper.server.persistence.SnapStream) 09:44:48 zookeeper | [2024-07-04 09:41:56,247] INFO Snapshotting: 0x0 to /var/lib/zookeeper/data/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) 09:44:48 zookeeper | [2024-07-04 09:41:56,250] INFO Snapshot loaded in 8 ms, highest zxid is 0x0, digest is 1371985504 (org.apache.zookeeper.server.ZKDatabase) 09:44:48 zookeeper | [2024-07-04 09:41:56,251] INFO Snapshotting: 0x0 to /var/lib/zookeeper/data/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) 09:44:48 zookeeper | [2024-07-04 09:41:56,251] INFO Snapshot taken in 1 ms (org.apache.zookeeper.server.ZooKeeperServer) 09:44:48 zookeeper | [2024-07-04 09:41:56,259] INFO PrepRequestProcessor (sid:0) started, reconfigEnabled=false (org.apache.zookeeper.server.PrepRequestProcessor) 09:44:48 zookeeper | [2024-07-04 09:41:56,260] INFO zookeeper.request_throttler.shutdownTimeout = 10000 ms (org.apache.zookeeper.server.RequestThrottler) 09:44:48 zookeeper | [2024-07-04 09:41:56,272] INFO Using checkIntervalMs=60000 maxPerMinute=10000 maxNeverUsedIntervalMs=0 (org.apache.zookeeper.server.ContainerManager) 09:44:48 zookeeper | [2024-07-04 09:41:56,273] INFO ZooKeeper audit is disabled. (org.apache.zookeeper.audit.ZKAuditProvider) 09:44:48 zookeeper | [2024-07-04 09:41:59,525] INFO Creating new log file: log.1 (org.apache.zookeeper.server.persistence.FileTxnLog) 09:44:48 =================================== 09:44:48 Tearing down containers... 09:44:48 time="2024-07-04T09:44:48Z" level=warning msg="The \"TEST_ENV\" variable is not set. Defaulting to a blank string." 09:44:48 Container grafana Stopping 09:44:48 Container policy-csit Stopping 09:44:48 Container policy-apex-pdp Stopping 09:44:48 Container policy-csit Stopped 09:44:48 Container policy-csit Removing 09:44:48 Container policy-csit Removed 09:44:48 Container grafana Stopped 09:44:48 Container grafana Removing 09:44:48 Container grafana Removed 09:44:48 Container prometheus Stopping 09:44:48 Container prometheus Stopped 09:44:48 Container prometheus Removing 09:44:48 Container prometheus Removed 09:44:58 Container policy-apex-pdp Stopped 09:44:58 Container policy-apex-pdp Removing 09:44:58 Container policy-apex-pdp Removed 09:44:58 Container simulator Stopping 09:44:58 Container policy-pap Stopping 09:45:09 Container simulator Stopped 09:45:09 Container simulator Removing 09:45:09 Container simulator Removed 09:45:09 Container policy-pap Stopped 09:45:09 Container policy-pap Removing 09:45:09 Container policy-pap Removed 09:45:09 Container kafka Stopping 09:45:09 Container policy-api Stopping 09:45:10 Container kafka Stopped 09:45:10 Container kafka Removing 09:45:10 Container kafka Removed 09:45:10 Container zookeeper Stopping 09:45:10 Container zookeeper Stopped 09:45:10 Container zookeeper Removing 09:45:10 Container zookeeper Removed 09:45:19 Container policy-api Stopped 09:45:19 Container policy-api Removing 09:45:19 Container policy-api Removed 09:45:19 Container policy-db-migrator Stopping 09:45:19 Container policy-db-migrator Stopped 09:45:19 Container policy-db-migrator Removing 09:45:19 Container policy-db-migrator Removed 09:45:19 Container mariadb Stopping 09:45:20 Container mariadb Stopped 09:45:20 Container mariadb Removing 09:45:20 Container mariadb Removed 09:45:20 Network compose_default Removing 09:45:20 Network compose_default Removed 09:45:20 $ ssh-agent -k 09:45:20 unset SSH_AUTH_SOCK; 09:45:20 unset SSH_AGENT_PID; 09:45:20 echo Agent pid 2147 killed; 09:45:20 [ssh-agent] Stopped. 09:45:21 Robot results publisher started... 09:45:21 INFO: Checking test criticality is deprecated and will be dropped in a future release! 09:45:21 -Parsing output xml: 09:45:21 Done! 09:45:21 -Copying log files to build dir: 09:45:21 Done! 09:45:21 -Assigning results to build: 09:45:21 Done! 09:45:21 -Checking thresholds: 09:45:21 Done! 09:45:21 Done publishing Robot results. 09:45:21 Build step 'Publish Robot Framework test results' changed build result to UNSTABLE 09:45:21 [PostBuildScript] - [INFO] Executing post build scripts. 09:45:21 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash /tmp/jenkins1793288325613390663.sh 09:45:21 ---> sysstat.sh 09:45:22 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash /tmp/jenkins12669198460770766156.sh 09:45:22 ---> package-listing.sh 09:45:22 ++ facter osfamily 09:45:22 ++ tr '[:upper:]' '[:lower:]' 09:45:22 + OS_FAMILY=debian 09:45:22 + workspace=/w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp 09:45:22 + START_PACKAGES=/tmp/packages_start.txt 09:45:22 + END_PACKAGES=/tmp/packages_end.txt 09:45:22 + DIFF_PACKAGES=/tmp/packages_diff.txt 09:45:22 + PACKAGES=/tmp/packages_start.txt 09:45:22 + '[' /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp ']' 09:45:22 + PACKAGES=/tmp/packages_end.txt 09:45:22 + case "${OS_FAMILY}" in 09:45:22 + dpkg -l 09:45:22 + grep '^ii' 09:45:22 + '[' -f /tmp/packages_start.txt ']' 09:45:22 + '[' -f /tmp/packages_end.txt ']' 09:45:22 + diff /tmp/packages_start.txt /tmp/packages_end.txt 09:45:22 + '[' /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp ']' 09:45:22 + mkdir -p /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp/archives/ 09:45:22 + cp -f /tmp/packages_diff.txt /tmp/packages_end.txt /tmp/packages_start.txt /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp/archives/ 09:45:22 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash /tmp/jenkins4650626352762568027.sh 09:45:22 ---> capture-instance-metadata.sh 09:45:22 Setup pyenv: 09:45:22 system 09:45:22 3.8.13 09:45:22 3.9.13 09:45:22 * 3.10.6 (set by /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp/.python-version) 09:45:22 lf-activate-venv(): INFO: Reuse venv:/tmp/venv-O2Gi from file:/tmp/.os_lf_venv 09:45:23 lf-activate-venv(): INFO: Installing: lftools 09:45:31 lf-activate-venv(): INFO: Adding /tmp/venv-O2Gi/bin to PATH 09:45:31 INFO: Running in OpenStack, capturing instance metadata 09:45:32 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash /tmp/jenkins411151357675070922.sh 09:45:32 provisioning config files... 09:45:32 copy managed file [jenkins-log-archives-settings] to file:/w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp@tmp/config6444636741840304009tmp 09:45:32 Regular expression run condition: Expression=[^.*logs-s3.*], Label=[] 09:45:32 Run condition [Regular expression match] preventing perform for step [Provide Configuration files] 09:45:32 [EnvInject] - Injecting environment variables from a build step. 09:45:32 [EnvInject] - Injecting as environment variables the properties content 09:45:32 SERVER_ID=logs 09:45:32 09:45:32 [EnvInject] - Variables injected successfully. 09:45:32 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash /tmp/jenkins3615890254173172972.sh 09:45:32 ---> create-netrc.sh 09:45:32 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash /tmp/jenkins15608429396824755413.sh 09:45:32 ---> python-tools-install.sh 09:45:32 Setup pyenv: 09:45:32 system 09:45:32 3.8.13 09:45:32 3.9.13 09:45:32 * 3.10.6 (set by /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp/.python-version) 09:45:32 lf-activate-venv(): INFO: Reuse venv:/tmp/venv-O2Gi from file:/tmp/.os_lf_venv 09:45:33 lf-activate-venv(): INFO: Installing: lftools 09:45:40 lf-activate-venv(): INFO: Adding /tmp/venv-O2Gi/bin to PATH 09:45:40 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash /tmp/jenkins3205436489079244234.sh 09:45:40 ---> sudo-logs.sh 09:45:40 Archiving 'sudo' log.. 09:45:41 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash /tmp/jenkins11071683102034424586.sh 09:45:41 ---> job-cost.sh 09:45:41 Setup pyenv: 09:45:41 system 09:45:41 3.8.13 09:45:41 3.9.13 09:45:41 * 3.10.6 (set by /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp/.python-version) 09:45:41 lf-activate-venv(): INFO: Reuse venv:/tmp/venv-O2Gi from file:/tmp/.os_lf_venv 09:45:42 lf-activate-venv(): INFO: Installing: zipp==1.1.0 python-openstackclient urllib3~=1.26.15 09:45:47 lf-activate-venv(): INFO: Adding /tmp/venv-O2Gi/bin to PATH 09:45:47 INFO: No Stack... 09:45:47 INFO: Retrieving Pricing Info for: v3-standard-8 09:45:47 INFO: Archiving Costs 09:45:47 [policy-apex-pdp-master-project-csit-verify-apex-pdp] $ /bin/bash -l /tmp/jenkins17714488965296724727.sh 09:45:47 ---> logs-deploy.sh 09:45:47 Setup pyenv: 09:45:47 system 09:45:47 3.8.13 09:45:47 3.9.13 09:45:47 * 3.10.6 (set by /w/workspace/policy-apex-pdp-master-project-csit-verify-apex-pdp/.python-version) 09:45:47 lf-activate-venv(): INFO: Reuse venv:/tmp/venv-O2Gi from file:/tmp/.os_lf_venv 09:45:48 lf-activate-venv(): INFO: Installing: lftools 09:45:56 lf-activate-venv(): INFO: Adding /tmp/venv-O2Gi/bin to PATH 09:45:56 INFO: Nexus URL https://nexus.onap.org path production/vex-yul-ecomp-jenkins-1/policy-apex-pdp-master-project-csit-verify-apex-pdp/550 09:45:56 INFO: archiving workspace using pattern(s): -p **/target/surefire-reports/*-output.txt 09:45:58 Archives upload complete. 09:45:58 INFO: archiving logs to Nexus 09:45:59 ---> uname -a: 09:45:59 Linux prd-ubuntu1804-docker-8c-8g-21289 4.15.0-192-generic #203-Ubuntu SMP Wed Aug 10 17:40:03 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux 09:45:59 09:45:59 09:45:59 ---> lscpu: 09:45:59 Architecture: x86_64 09:45:59 CPU op-mode(s): 32-bit, 64-bit 09:45:59 Byte Order: Little Endian 09:45:59 CPU(s): 8 09:45:59 On-line CPU(s) list: 0-7 09:45:59 Thread(s) per core: 1 09:45:59 Core(s) per socket: 1 09:45:59 Socket(s): 8 09:45:59 NUMA node(s): 1 09:45:59 Vendor ID: AuthenticAMD 09:45:59 CPU family: 23 09:45:59 Model: 49 09:45:59 Model name: AMD EPYC-Rome Processor 09:45:59 Stepping: 0 09:45:59 CPU MHz: 2800.000 09:45:59 BogoMIPS: 5600.00 09:45:59 Virtualization: AMD-V 09:45:59 Hypervisor vendor: KVM 09:45:59 Virtualization type: full 09:45:59 L1d cache: 32K 09:45:59 L1i cache: 32K 09:45:59 L2 cache: 512K 09:45:59 L3 cache: 16384K 09:45:59 NUMA node0 CPU(s): 0-7 09:45:59 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm rep_good nopl xtopology cpuid extd_apicid tsc_known_freq pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext perfctr_core ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves clzero xsaveerptr arat npt nrip_save umip rdpid arch_capabilities 09:45:59 09:45:59 09:45:59 ---> nproc: 09:45:59 8 09:45:59 09:45:59 09:45:59 ---> df -h: 09:45:59 Filesystem Size Used Avail Use% Mounted on 09:45:59 udev 16G 0 16G 0% /dev 09:45:59 tmpfs 3.2G 708K 3.2G 1% /run 09:45:59 /dev/vda1 155G 14G 141G 9% / 09:45:59 tmpfs 16G 0 16G 0% /dev/shm 09:45:59 tmpfs 5.0M 0 5.0M 0% /run/lock 09:45:59 tmpfs 16G 0 16G 0% /sys/fs/cgroup 09:45:59 /dev/vda15 105M 4.4M 100M 5% /boot/efi 09:45:59 tmpfs 3.2G 0 3.2G 0% /run/user/1001 09:45:59 09:45:59 09:45:59 ---> free -m: 09:45:59 total used free shared buff/cache available 09:45:59 Mem: 32167 881 24728 0 6557 30830 09:45:59 Swap: 1023 0 1023 09:45:59 09:45:59 09:45:59 ---> ip addr: 09:45:59 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 09:45:59 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 09:45:59 inet 127.0.0.1/8 scope host lo 09:45:59 valid_lft forever preferred_lft forever 09:45:59 inet6 ::1/128 scope host 09:45:59 valid_lft forever preferred_lft forever 09:45:59 2: ens3: mtu 1458 qdisc mq state UP group default qlen 1000 09:45:59 link/ether fa:16:3e:4e:7c:ab brd ff:ff:ff:ff:ff:ff 09:45:59 inet 10.30.107.74/23 brd 10.30.107.255 scope global dynamic ens3 09:45:59 valid_lft 85956sec preferred_lft 85956sec 09:45:59 inet6 fe80::f816:3eff:fe4e:7cab/64 scope link 09:45:59 valid_lft forever preferred_lft forever 09:45:59 3: docker0: mtu 1500 qdisc noqueue state DOWN group default 09:45:59 link/ether 02:42:ff:5d:12:73 brd ff:ff:ff:ff:ff:ff 09:45:59 inet 10.250.0.254/24 brd 10.250.0.255 scope global docker0 09:45:59 valid_lft forever preferred_lft forever 09:45:59 inet6 fe80::42:ffff:fe5d:1273/64 scope link 09:45:59 valid_lft forever preferred_lft forever 09:45:59 09:45:59 09:45:59 ---> sar -b -r -n DEV: 09:45:59 Linux 4.15.0-192-generic (prd-ubuntu1804-docker-8c-8g-21289) 07/04/24 _x86_64_ (8 CPU) 09:45:59 09:45:59 09:38:36 LINUX RESTART (8 CPU) 09:45:59 09:45:59 09:39:01 tps rtps wtps bread/s bwrtn/s 09:45:59 09:40:01 300.18 71.32 228.86 4608.83 51261.99 09:45:59 09:41:01 236.43 18.96 217.46 2270.55 85668.39 09:45:59 09:42:01 296.03 11.80 284.24 757.64 110741.51 09:45:59 09:43:01 167.07 0.20 166.87 33.33 26817.33 09:45:59 09:44:01 137.29 0.28 137.01 22.26 37768.79 09:45:59 09:45:01 32.71 0.05 32.66 8.26 22286.04 09:45:59 Average: 194.95 17.10 177.85 1283.44 55756.41 09:45:59 09:45:59 09:39:01 kbmemfree kbavail kbmemused %memused kbbuffers kbcached kbcommit %commit kbactive kbinact kbdirty 09:45:59 09:40:01 30083944 31631456 2855276 8.67 66888 1791216 1497832 4.41 933084 1629132 155096 09:45:59 09:41:01 25801260 31634652 7137960 21.67 124952 5853344 1725080 5.08 1015296 5607776 2496328 09:45:59 09:42:01 25007064 30957944 7932156 24.08 135300 5943012 7225776 21.26 1817920 5538776 3720 09:45:59 09:43:01 23517008 29600676 9422212 28.60 144224 6060400 9000408 26.48 3264772 5559324 8164 09:45:59 09:44:01 23120280 29511168 9818940 29.81 172196 6302632 9258320 27.24 3414972 5765124 1672 09:45:59 09:45:01 23445044 29831164 9494176 28.82 172580 6302480 7578320 22.30 3114984 5756948 412 09:45:59 Average: 25162433 30527843 7776787 23.61 136023 5375514 6047623 17.79 2260171 4976180 444232 09:45:59 09:45:59 09:39:01 IFACE rxpck/s txpck/s rxkB/s txkB/s rxcmp/s txcmp/s rxmcst/s %ifutil 09:45:59 09:40:01 ens3 288.09 230.84 1516.85 75.53 0.00 0.00 0.00 0.00 09:45:59 09:40:01 lo 1.40 1.40 0.15 0.15 0.00 0.00 0.00 0.00 09:45:59 09:40:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:45:59 09:41:01 ens3 1413.56 660.74 33324.81 59.74 0.00 0.00 0.00 0.00 09:45:59 09:41:01 lo 14.13 14.13 1.37 1.37 0.00 0.00 0.00 0.00 09:45:59 09:41:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:45:59 09:42:01 br-9cce439c9c7f 0.08 0.15 0.00 0.01 0.00 0.00 0.00 0.00 09:45:59 09:42:01 veth53b9820 0.00 0.15 0.00 0.01 0.00 0.00 0.00 0.00 09:45:59 09:42:01 vethd9c37f1 0.45 0.60 0.02 0.04 0.00 0.00 0.00 0.00 09:45:59 09:42:01 veth1ce11c2 0.17 0.33 0.01 0.02 0.00 0.00 0.00 0.00 09:45:59 09:43:01 br-9cce439c9c7f 1.15 1.05 0.11 0.41 0.00 0.00 0.00 0.00 09:45:59 09:43:01 veth53b9820 0.63 0.78 0.07 0.38 0.00 0.00 0.00 0.00 09:45:59 09:43:01 vethd9c37f1 54.41 64.94 19.86 15.75 0.00 0.00 0.00 0.00 09:45:59 09:43:01 veth1ce11c2 24.31 22.40 10.96 16.09 0.00 0.00 0.00 0.00 09:45:59 09:44:01 br-9cce439c9c7f 0.35 0.22 0.02 0.01 0.00 0.00 0.00 0.00 09:45:59 09:44:01 veth53b9820 0.32 0.27 0.02 0.02 0.00 0.00 0.00 0.00 09:45:59 09:44:01 vethd9c37f1 33.14 41.63 60.11 12.97 0.00 0.00 0.00 0.00 09:45:59 09:44:01 veth1ce11c2 44.43 36.61 21.62 79.37 0.00 0.00 0.00 0.01 09:45:59 09:45:01 br-9cce439c9c7f 0.02 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:45:59 09:45:01 vethd9c37f1 0.22 0.40 0.11 0.09 0.00 0.00 0.00 0.00 09:45:59 09:45:01 veth1ce11c2 0.40 0.58 0.58 0.04 0.00 0.00 0.00 0.00 09:45:59 09:45:01 ens3 1795.40 965.16 36095.80 159.61 0.00 0.00 0.00 0.00 09:45:59 Average: br-9cce439c9c7f 0.27 0.24 0.02 0.07 0.00 0.00 0.00 0.00 09:45:59 Average: vethd9c37f1 14.70 17.93 13.35 4.81 0.00 0.00 0.00 0.00 09:45:59 Average: veth1ce11c2 11.55 9.99 5.53 15.92 0.00 0.00 0.00 0.00 09:45:59 Average: ens3 297.15 159.36 6005.21 26.41 0.00 0.00 0.00 0.00 09:45:59 09:45:59 09:45:59 ---> sar -P ALL: 09:45:59 Linux 4.15.0-192-generic (prd-ubuntu1804-docker-8c-8g-21289) 07/04/24 _x86_64_ (8 CPU) 09:45:59 09:45:59 09:38:36 LINUX RESTART (8 CPU) 09:45:59 09:45:59 09:39:01 CPU %user %nice %system %iowait %steal %idle 09:45:59 09:40:01 all 9.96 0.00 1.07 5.57 0.05 83.36 09:45:59 09:40:01 0 4.54 0.00 0.73 1.11 0.03 93.58 09:45:59 09:40:01 1 10.29 0.00 0.89 1.65 0.03 87.14 09:45:59 09:40:01 2 17.65 0.00 1.16 1.77 0.05 79.38 09:45:59 09:40:01 3 12.34 0.00 1.44 2.16 0.05 84.01 09:45:59 09:40:01 4 13.55 0.00 1.43 25.73 0.05 59.24 09:45:59 09:40:01 5 8.40 0.00 0.77 2.14 0.03 88.65 09:45:59 09:40:01 6 9.91 0.00 1.14 6.31 0.08 82.56 09:45:59 09:40:01 7 2.96 0.00 0.97 3.59 0.05 92.43 09:45:59 09:41:01 all 15.39 0.00 5.79 13.09 0.08 65.65 09:45:59 09:41:01 0 12.08 0.00 6.21 6.06 0.05 75.60 09:45:59 09:41:01 1 16.30 0.00 5.74 3.37 0.05 74.54 09:45:59 09:41:01 2 30.13 0.00 6.42 15.84 0.08 47.52 09:45:59 09:41:01 3 14.56 0.00 5.64 16.45 0.10 63.25 09:45:59 09:41:01 4 13.65 0.00 4.43 23.67 0.10 58.15 09:45:59 09:41:01 5 13.21 0.00 6.21 4.48 0.07 76.03 09:45:59 09:41:01 6 11.62 0.00 5.94 27.40 0.07 54.98 09:45:59 09:41:01 7 11.59 0.00 5.73 7.49 0.14 75.05 09:45:59 09:42:01 all 5.74 0.00 1.72 22.59 0.07 69.88 09:45:59 09:42:01 0 5.37 0.00 1.40 19.06 0.03 74.14 09:45:59 09:42:01 1 6.17 0.00 1.46 1.76 0.05 90.57 09:45:59 09:42:01 2 6.71 0.00 2.19 34.80 0.08 56.21 09:45:59 09:42:01 3 6.12 0.00 1.56 4.73 0.05 87.55 09:45:59 09:42:01 4 4.20 0.00 2.11 46.07 0.05 47.57 09:45:59 09:42:01 5 5.01 0.00 1.46 2.51 0.08 90.94 09:45:59 09:42:01 6 5.38 0.00 1.95 57.48 0.18 35.01 09:45:59 09:42:01 7 6.98 0.00 1.66 14.41 0.03 76.92 09:45:59 09:43:01 all 22.36 0.00 2.69 3.69 0.07 71.19 09:45:59 09:43:01 0 18.87 0.00 2.31 2.31 0.07 76.44 09:45:59 09:43:01 1 27.90 0.00 2.94 1.44 0.08 67.63 09:45:59 09:43:01 2 23.21 0.00 3.14 1.45 0.07 72.13 09:45:59 09:43:01 3 21.13 0.00 2.30 2.49 0.08 74.00 09:45:59 09:43:01 4 19.83 0.00 2.16 13.93 0.07 64.01 09:45:59 09:43:01 5 23.09 0.00 2.73 0.25 0.08 73.84 09:45:59 09:43:01 6 23.52 0.00 2.95 0.67 0.08 72.77 09:45:59 09:43:01 7 21.34 0.00 2.96 6.95 0.07 68.69 09:45:59 09:44:01 all 7.12 0.00 1.71 4.19 0.06 86.92 09:45:59 09:44:01 0 8.30 0.00 1.69 3.16 0.07 86.78 09:45:59 09:44:01 1 8.81 0.00 1.58 1.34 0.05 88.22 09:45:59 09:44:01 2 6.10 0.00 1.73 9.01 0.07 83.10 09:45:59 09:44:01 3 5.33 0.00 1.65 4.44 0.05 88.54 09:45:59 09:44:01 4 5.10 0.00 1.52 11.73 0.08 81.56 09:45:59 09:44:01 5 8.36 0.00 2.09 1.34 0.05 88.15 09:45:59 09:44:01 6 6.33 0.00 1.58 1.41 0.08 90.60 09:45:59 09:44:01 7 8.64 0.00 1.84 1.01 0.05 88.46 09:45:59 09:45:01 all 1.73 0.00 0.48 1.32 0.04 96.43 09:45:59 09:45:01 0 1.91 0.00 0.69 0.02 0.05 97.34 09:45:59 09:45:01 1 0.78 0.00 0.32 0.05 0.02 98.83 09:45:59 09:45:01 2 1.45 0.00 0.40 10.16 0.03 87.95 09:45:59 09:45:01 3 2.23 0.00 0.47 0.03 0.03 97.23 09:45:59 09:45:01 4 1.42 0.00 0.43 0.02 0.03 98.09 09:45:59 09:45:01 5 1.79 0.00 0.57 0.12 0.03 97.49 09:45:59 09:45:01 6 2.30 0.00 0.54 0.07 0.05 97.05 09:45:59 09:45:01 7 1.89 0.00 0.44 0.10 0.05 97.52 09:45:59 Average: all 10.38 0.00 2.24 8.40 0.06 78.92 09:45:59 Average: 0 8.52 0.00 2.17 5.29 0.05 83.97 09:45:59 Average: 1 11.69 0.00 2.15 1.60 0.05 84.51 09:45:59 Average: 2 14.18 0.00 2.50 12.18 0.06 71.07 09:45:59 Average: 3 10.27 0.00 2.17 5.04 0.06 82.46 09:45:59 Average: 4 9.62 0.00 2.01 20.16 0.06 68.14 09:45:59 Average: 5 9.96 0.00 2.30 1.80 0.06 85.88 09:45:59 Average: 6 9.84 0.00 2.35 15.55 0.09 72.17 09:45:59 Average: 7 8.92 0.00 2.27 5.59 0.06 83.16 09:45:59 09:45:59 09:45:59