Started by upstream project "policy-docker-master-merge-java" build number 350 originally caused by: Triggered by Gerrit: https://gerrit.onap.org/r/c/policy/docker/+/137727 Running as SYSTEM [EnvInject] - Loading node environment variables. Building remotely on prd-ubuntu1804-docker-8c-8g-25507 (ubuntu1804-docker-8c-8g) in workspace /w/workspace/policy-pap-master-project-csit-pap [ssh-agent] Looking for ssh-agent implementation... [ssh-agent] Exec ssh-agent (binary ssh-agent on a remote machine) $ ssh-agent SSH_AUTH_SOCK=/tmp/ssh-JjI1I7Nl25Y4/agent.2158 SSH_AGENT_PID=2159 [ssh-agent] Started. Running ssh-add (command line suppressed) Identity added: /w/workspace/policy-pap-master-project-csit-pap@tmp/private_key_14973092772675130381.key (/w/workspace/policy-pap-master-project-csit-pap@tmp/private_key_14973092772675130381.key) [ssh-agent] Using credentials onap-jobbuiler (Gerrit user) The recommended git tool is: NONE using credential onap-jenkins-ssh Wiping out workspace first. Cloning the remote Git repository Cloning repository git://cloud.onap.org/mirror/policy/docker.git > git init /w/workspace/policy-pap-master-project-csit-pap # timeout=10 Fetching upstream changes from git://cloud.onap.org/mirror/policy/docker.git > git --version # timeout=10 > git --version # 'git version 2.17.1' using GIT_SSH to set credentials Gerrit user Verifying host key using manually-configured host key entries > git fetch --tags --progress -- git://cloud.onap.org/mirror/policy/docker.git +refs/heads/*:refs/remotes/origin/* # timeout=30 > git config remote.origin.url git://cloud.onap.org/mirror/policy/docker.git # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 Avoid second fetch > git rev-parse refs/remotes/origin/master^{commit} # timeout=10 Checking out Revision b6d1d479556b6798d8dfb70aa87db4e247327113 (refs/remotes/origin/master) > git config core.sparsecheckout # timeout=10 > git checkout -f b6d1d479556b6798d8dfb70aa87db4e247327113 # timeout=30 Commit message: "Merge "Add ACM db upgrade/downgrade to db-migrator"" > git rev-list --no-walk 427f193118436b2aa7664f72fcb16ca1b25b8061 # timeout=10 provisioning config files... copy managed file [npmrc] to file:/home/jenkins/.npmrc copy managed file [pipconf] to file:/home/jenkins/.config/pip/pip.conf [policy-pap-master-project-csit-pap] $ /bin/bash /tmp/jenkins13755176375877631448.sh ---> python-tools-install.sh Setup pyenv: * system (set by /opt/pyenv/version) * 3.8.13 (set by /opt/pyenv/version) * 3.9.13 (set by /opt/pyenv/version) * 3.10.6 (set by /opt/pyenv/version) lf-activate-venv(): INFO: Creating python3 venv at /tmp/venv-8CY4 lf-activate-venv(): INFO: Save venv in file: /tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-8CY4/bin to PATH Generating Requirements File Python 3.10.6 pip 24.0 from /tmp/venv-8CY4/lib/python3.10/site-packages/pip (python 3.10) appdirs==1.4.4 argcomplete==3.3.0 aspy.yaml==1.3.0 attrs==23.2.0 autopage==0.5.2 beautifulsoup4==4.12.3 boto3==1.34.90 botocore==1.34.90 bs4==0.0.2 cachetools==5.3.3 certifi==2024.2.2 cffi==1.16.0 cfgv==3.4.0 chardet==5.2.0 charset-normalizer==3.3.2 click==8.1.7 cliff==4.6.0 cmd2==2.4.3 cryptography==3.3.2 debtcollector==3.0.0 decorator==5.1.1 defusedxml==0.7.1 Deprecated==1.2.14 distlib==0.3.8 dnspython==2.6.1 docker==4.2.2 dogpile.cache==1.3.2 email_validator==2.1.1 filelock==3.13.4 future==1.0.0 gitdb==4.0.11 GitPython==3.1.43 google-auth==2.29.0 httplib2==0.22.0 identify==2.5.36 idna==3.7 importlib-resources==1.5.0 iso8601==2.1.0 Jinja2==3.1.3 jmespath==1.0.1 jsonpatch==1.33 jsonpointer==2.4 jsonschema==4.21.1 jsonschema-specifications==2023.12.1 keystoneauth1==5.6.0 kubernetes==29.0.0 lftools==0.37.10 lxml==5.2.1 MarkupSafe==2.1.5 msgpack==1.0.8 multi_key_dict==2.0.3 munch==4.0.0 netaddr==1.2.1 netifaces==0.11.0 niet==1.4.2 nodeenv==1.8.0 oauth2client==4.1.3 oauthlib==3.2.2 openstacksdk==3.1.0 os-client-config==2.1.0 os-service-types==1.7.0 osc-lib==3.0.1 oslo.config==9.4.0 oslo.context==5.5.0 oslo.i18n==6.3.0 oslo.log==5.5.1 oslo.serialization==5.4.0 oslo.utils==7.1.0 packaging==24.0 pbr==6.0.0 platformdirs==4.2.1 prettytable==3.10.0 pyasn1==0.6.0 pyasn1_modules==0.4.0 pycparser==2.22 pygerrit2==2.0.15 PyGithub==2.3.0 pyinotify==0.9.6 PyJWT==2.8.0 PyNaCl==1.5.0 pyparsing==2.4.7 pyperclip==1.8.2 pyrsistent==0.20.0 python-cinderclient==9.5.0 python-dateutil==2.9.0.post0 python-heatclient==3.5.0 python-jenkins==1.8.2 python-keystoneclient==5.4.0 python-magnumclient==4.4.0 python-novaclient==18.6.0 python-openstackclient==6.6.0 python-swiftclient==4.5.0 PyYAML==6.0.1 referencing==0.35.0 requests==2.31.0 requests-oauthlib==2.0.0 requestsexceptions==1.4.0 rfc3986==2.0.0 rpds-py==0.18.0 rsa==4.9 ruamel.yaml==0.18.6 ruamel.yaml.clib==0.2.8 s3transfer==0.10.1 simplejson==3.19.2 six==1.16.0 smmap==5.0.1 soupsieve==2.5 stevedore==5.2.0 tabulate==0.9.0 toml==0.10.2 tomlkit==0.12.4 tqdm==4.66.2 typing_extensions==4.11.0 tzdata==2024.1 urllib3==1.26.18 virtualenv==20.26.0 wcwidth==0.2.13 websocket-client==1.8.0 wrapt==1.16.0 xdg==6.0.0 xmltodict==0.13.0 yq==3.4.1 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content SET_JDK_VERSION=openjdk17 GIT_URL="git://cloud.onap.org/mirror" [EnvInject] - Variables injected successfully. [policy-pap-master-project-csit-pap] $ /bin/sh /tmp/jenkins3426379937651602965.sh ---> update-java-alternatives.sh ---> Updating Java version ---> Ubuntu/Debian system detected update-alternatives: using /usr/lib/jvm/java-17-openjdk-amd64/bin/java to provide /usr/bin/java (java) in manual mode update-alternatives: using /usr/lib/jvm/java-17-openjdk-amd64/bin/javac to provide /usr/bin/javac (javac) in manual mode update-alternatives: using /usr/lib/jvm/java-17-openjdk-amd64 to provide /usr/lib/jvm/java-openjdk (java_sdk_openjdk) in manual mode openjdk version "17.0.4" 2022-07-19 OpenJDK Runtime Environment (build 17.0.4+8-Ubuntu-118.04) OpenJDK 64-Bit Server VM (build 17.0.4+8-Ubuntu-118.04, mixed mode, sharing) JAVA_HOME=/usr/lib/jvm/java-17-openjdk-amd64 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path '/tmp/java.env' [EnvInject] - Variables injected successfully. [policy-pap-master-project-csit-pap] $ /bin/sh -xe /tmp/jenkins8684352219582312135.sh + /w/workspace/policy-pap-master-project-csit-pap/csit/run-project-csit.sh pap + set +u + save_set + RUN_CSIT_SAVE_SET=ehxB + RUN_CSIT_SHELLOPTS=braceexpand:errexit:hashall:interactive-comments:pipefail:xtrace + '[' 1 -eq 0 ']' + '[' -z /w/workspace/policy-pap-master-project-csit-pap ']' + export PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/w/workspace/policy-pap-master-project-csit-pap/csit:/w/workspace/policy-pap-master-project-csit-pap/scripts:/bin + PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/w/workspace/policy-pap-master-project-csit-pap/csit:/w/workspace/policy-pap-master-project-csit-pap/scripts:/bin + export SCRIPTS=/w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts + SCRIPTS=/w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts + export ROBOT_VARIABLES= + ROBOT_VARIABLES= + export PROJECT=pap + PROJECT=pap + cd /w/workspace/policy-pap-master-project-csit-pap + rm -rf /w/workspace/policy-pap-master-project-csit-pap/csit/archives/pap + mkdir -p /w/workspace/policy-pap-master-project-csit-pap/csit/archives/pap + source_safely /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/prepare-robot-env.sh + '[' -z /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/prepare-robot-env.sh ']' + relax_set + set +e + set +o pipefail + . /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/prepare-robot-env.sh ++ '[' -z /w/workspace/policy-pap-master-project-csit-pap ']' +++ mktemp -d ++ ROBOT_VENV=/tmp/tmp.IouyJTO6pY ++ echo ROBOT_VENV=/tmp/tmp.IouyJTO6pY +++ python3 --version ++ echo 'Python version is: Python 3.6.9' Python version is: Python 3.6.9 ++ python3 -m venv --clear /tmp/tmp.IouyJTO6pY ++ source /tmp/tmp.IouyJTO6pY/bin/activate +++ deactivate nondestructive +++ '[' -n '' ']' +++ '[' -n '' ']' +++ '[' -n /bin/bash -o -n '' ']' +++ hash -r +++ '[' -n '' ']' +++ unset VIRTUAL_ENV +++ '[' '!' nondestructive = nondestructive ']' +++ VIRTUAL_ENV=/tmp/tmp.IouyJTO6pY +++ export VIRTUAL_ENV +++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/w/workspace/policy-pap-master-project-csit-pap/csit:/w/workspace/policy-pap-master-project-csit-pap/scripts:/bin +++ PATH=/tmp/tmp.IouyJTO6pY/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/w/workspace/policy-pap-master-project-csit-pap/csit:/w/workspace/policy-pap-master-project-csit-pap/scripts:/bin +++ export PATH +++ '[' -n '' ']' +++ '[' -z '' ']' +++ _OLD_VIRTUAL_PS1= +++ '[' 'x(tmp.IouyJTO6pY) ' '!=' x ']' +++ PS1='(tmp.IouyJTO6pY) ' +++ export PS1 +++ '[' -n /bin/bash -o -n '' ']' +++ hash -r ++ set -exu ++ python3 -m pip install -qq --upgrade 'pip<=23.0' 'setuptools<=66.1.1' ++ echo 'Installing Python Requirements' Installing Python Requirements ++ python3 -m pip install -qq -r /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/pylibs.txt ++ python3 -m pip -qq freeze bcrypt==4.0.1 beautifulsoup4==4.12.3 bitarray==2.9.2 certifi==2024.2.2 cffi==1.15.1 charset-normalizer==2.0.12 cryptography==40.0.2 decorator==5.1.1 elasticsearch==7.17.9 elasticsearch-dsl==7.4.1 enum34==1.1.10 idna==3.7 importlib-resources==5.4.0 ipaddr==2.2.0 isodate==0.6.1 jmespath==0.10.0 jsonpatch==1.32 jsonpath-rw==1.4.0 jsonpointer==2.3 lxml==5.2.1 netaddr==0.8.0 netifaces==0.11.0 odltools==0.1.28 paramiko==3.4.0 pkg_resources==0.0.0 ply==3.11 pyang==2.6.0 pyangbind==0.8.1 pycparser==2.21 pyhocon==0.3.60 PyNaCl==1.5.0 pyparsing==3.1.2 python-dateutil==2.9.0.post0 regex==2023.8.8 requests==2.27.1 robotframework==6.1.1 robotframework-httplibrary==0.4.2 robotframework-pythonlibcore==3.0.0 robotframework-requests==0.9.4 robotframework-selenium2library==3.0.0 robotframework-seleniumlibrary==5.1.3 robotframework-sshlibrary==3.8.0 scapy==2.5.0 scp==0.14.5 selenium==3.141.0 six==1.16.0 soupsieve==2.3.2.post1 urllib3==1.26.18 waitress==2.0.0 WebOb==1.8.7 WebTest==3.0.0 zipp==3.6.0 ++ mkdir -p /tmp/tmp.IouyJTO6pY/src/onap ++ rm -rf /tmp/tmp.IouyJTO6pY/src/onap/testsuite ++ python3 -m pip install -qq --upgrade --extra-index-url=https://nexus3.onap.org/repository/PyPi.staging/simple 'robotframework-onap==0.6.0.*' --pre ++ echo 'Installing python confluent-kafka library' Installing python confluent-kafka library ++ python3 -m pip install -qq confluent-kafka ++ echo 'Uninstall docker-py and reinstall docker.' Uninstall docker-py and reinstall docker. ++ python3 -m pip uninstall -y -qq docker ++ python3 -m pip install -U -qq docker ++ python3 -m pip -qq freeze bcrypt==4.0.1 beautifulsoup4==4.12.3 bitarray==2.9.2 certifi==2024.2.2 cffi==1.15.1 charset-normalizer==2.0.12 confluent-kafka==2.3.0 cryptography==40.0.2 decorator==5.1.1 deepdiff==5.7.0 dnspython==2.2.1 docker==5.0.3 elasticsearch==7.17.9 elasticsearch-dsl==7.4.1 enum34==1.1.10 future==1.0.0 idna==3.7 importlib-resources==5.4.0 ipaddr==2.2.0 isodate==0.6.1 Jinja2==3.0.3 jmespath==0.10.0 jsonpatch==1.32 jsonpath-rw==1.4.0 jsonpointer==2.3 kafka-python==2.0.2 lxml==5.2.1 MarkupSafe==2.0.1 more-itertools==5.0.0 netaddr==0.8.0 netifaces==0.11.0 odltools==0.1.28 ordered-set==4.0.2 paramiko==3.4.0 pbr==6.0.0 pkg_resources==0.0.0 ply==3.11 protobuf==3.19.6 pyang==2.6.0 pyangbind==0.8.1 pycparser==2.21 pyhocon==0.3.60 PyNaCl==1.5.0 pyparsing==3.1.2 python-dateutil==2.9.0.post0 PyYAML==6.0.1 regex==2023.8.8 requests==2.27.1 robotframework==6.1.1 robotframework-httplibrary==0.4.2 robotframework-onap==0.6.0.dev105 robotframework-pythonlibcore==3.0.0 robotframework-requests==0.9.4 robotframework-selenium2library==3.0.0 robotframework-seleniumlibrary==5.1.3 robotframework-sshlibrary==3.8.0 robotlibcore-temp==1.0.2 scapy==2.5.0 scp==0.14.5 selenium==3.141.0 six==1.16.0 soupsieve==2.3.2.post1 urllib3==1.26.18 waitress==2.0.0 WebOb==1.8.7 websocket-client==1.3.1 WebTest==3.0.0 zipp==3.6.0 ++ uname ++ grep -q Linux ++ sudo apt-get -y -qq install libxml2-utils + load_set + _setopts=ehuxB ++ echo braceexpand:hashall:interactive-comments:nounset:xtrace ++ tr : ' ' + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o braceexpand + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o hashall + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o interactive-comments + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o nounset + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o xtrace ++ echo ehuxB ++ sed 's/./& /g' + for i in $(echo "$_setopts" | sed 's/./& /g') + set +e + for i in $(echo "$_setopts" | sed 's/./& /g') + set +h + for i in $(echo "$_setopts" | sed 's/./& /g') + set +u + for i in $(echo "$_setopts" | sed 's/./& /g') + set +x + source_safely /tmp/tmp.IouyJTO6pY/bin/activate + '[' -z /tmp/tmp.IouyJTO6pY/bin/activate ']' + relax_set + set +e + set +o pipefail + . /tmp/tmp.IouyJTO6pY/bin/activate ++ deactivate nondestructive ++ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/w/workspace/policy-pap-master-project-csit-pap/csit:/w/workspace/policy-pap-master-project-csit-pap/scripts:/bin ']' ++ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/w/workspace/policy-pap-master-project-csit-pap/csit:/w/workspace/policy-pap-master-project-csit-pap/scripts:/bin ++ export PATH ++ unset _OLD_VIRTUAL_PATH ++ '[' -n '' ']' ++ '[' -n /bin/bash -o -n '' ']' ++ hash -r ++ '[' -n '' ']' ++ unset VIRTUAL_ENV ++ '[' '!' nondestructive = nondestructive ']' ++ VIRTUAL_ENV=/tmp/tmp.IouyJTO6pY ++ export VIRTUAL_ENV ++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/w/workspace/policy-pap-master-project-csit-pap/csit:/w/workspace/policy-pap-master-project-csit-pap/scripts:/bin ++ PATH=/tmp/tmp.IouyJTO6pY/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/w/workspace/policy-pap-master-project-csit-pap/csit:/w/workspace/policy-pap-master-project-csit-pap/scripts:/bin ++ export PATH ++ '[' -n '' ']' ++ '[' -z '' ']' ++ _OLD_VIRTUAL_PS1='(tmp.IouyJTO6pY) ' ++ '[' 'x(tmp.IouyJTO6pY) ' '!=' x ']' ++ PS1='(tmp.IouyJTO6pY) (tmp.IouyJTO6pY) ' ++ export PS1 ++ '[' -n /bin/bash -o -n '' ']' ++ hash -r + load_set + _setopts=hxB ++ echo braceexpand:hashall:interactive-comments:xtrace ++ tr : ' ' + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o braceexpand + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o hashall + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o interactive-comments + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o xtrace ++ sed 's/./& /g' ++ echo hxB + for i in $(echo "$_setopts" | sed 's/./& /g') + set +h + for i in $(echo "$_setopts" | sed 's/./& /g') + set +x + export TEST_PLAN_DIR=/w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests + TEST_PLAN_DIR=/w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests + export TEST_OPTIONS= + TEST_OPTIONS= ++ mktemp -d + WORKDIR=/tmp/tmp.UcCaUoaeAm + cd /tmp/tmp.UcCaUoaeAm + docker login -u docker -p docker nexus3.onap.org:10001 WARNING! Using --password via the CLI is insecure. Use --password-stdin. WARNING! Your password will be stored unencrypted in /home/jenkins/.docker/config.json. Configure a credential helper to remove this warning. See https://docs.docker.com/engine/reference/commandline/login/#credentials-store Login Succeeded + SETUP=/w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/setup-pap.sh + '[' -f /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/setup-pap.sh ']' + echo 'Running setup script /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/setup-pap.sh' Running setup script /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/setup-pap.sh + source_safely /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/setup-pap.sh + '[' -z /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/setup-pap.sh ']' + relax_set + set +e + set +o pipefail + . /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/setup-pap.sh ++ source /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/node-templates.sh +++ '[' -z /w/workspace/policy-pap-master-project-csit-pap ']' ++++ awk -F= '$1 == "defaultbranch" { print $2 }' /w/workspace/policy-pap-master-project-csit-pap/.gitreview +++ GERRIT_BRANCH=master +++ echo GERRIT_BRANCH=master GERRIT_BRANCH=master +++ rm -rf /w/workspace/policy-pap-master-project-csit-pap/models +++ mkdir /w/workspace/policy-pap-master-project-csit-pap/models +++ git clone -b master --single-branch https://github.com/onap/policy-models.git /w/workspace/policy-pap-master-project-csit-pap/models Cloning into '/w/workspace/policy-pap-master-project-csit-pap/models'... +++ export DATA=/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/policies +++ DATA=/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/policies +++ export NODETEMPLATES=/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/nodetemplates +++ NODETEMPLATES=/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/nodetemplates +++ sed -e 's!Measurement_vGMUX!ADifferentValue!' /w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/policies/vCPE.policy.monitoring.input.tosca.json +++ sed -e 's!"version": "1.0.0"!"version": "2.0.0"!' -e 's!"policy-version": 1!"policy-version": 2!' /w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/policies/vCPE.policy.monitoring.input.tosca.json ++ source /w/workspace/policy-pap-master-project-csit-pap/compose/start-compose.sh apex-pdp --grafana +++ '[' -z /w/workspace/policy-pap-master-project-csit-pap ']' +++ COMPOSE_FOLDER=/w/workspace/policy-pap-master-project-csit-pap/compose +++ grafana=false +++ gui=false +++ [[ 2 -gt 0 ]] +++ key=apex-pdp +++ case $key in +++ echo apex-pdp apex-pdp +++ component=apex-pdp +++ shift +++ [[ 1 -gt 0 ]] +++ key=--grafana +++ case $key in +++ grafana=true +++ shift +++ [[ 0 -gt 0 ]] +++ cd /w/workspace/policy-pap-master-project-csit-pap/compose +++ echo 'Configuring docker compose...' Configuring docker compose... +++ source export-ports.sh +++ source get-versions.sh +++ '[' -z pap ']' +++ '[' -n apex-pdp ']' +++ '[' apex-pdp == logs ']' +++ '[' true = true ']' +++ echo 'Starting apex-pdp application with Grafana' Starting apex-pdp application with Grafana +++ docker-compose up -d apex-pdp grafana Creating network "compose_default" with the default driver Pulling prometheus (nexus3.onap.org:10001/prom/prometheus:latest)... latest: Pulling from prom/prometheus Digest: sha256:4f6c47e39a9064028766e8c95890ed15690c30f00c4ba14e7ce6ae1ded0295b1 Status: Downloaded newer image for nexus3.onap.org:10001/prom/prometheus:latest Pulling grafana (nexus3.onap.org:10001/grafana/grafana:latest)... latest: Pulling from grafana/grafana Digest: sha256:7d5faae481a4c6f436c99e98af11534f7fd5e8d3e35213552dd1dd02bc393d2e Status: Downloaded newer image for nexus3.onap.org:10001/grafana/grafana:latest Pulling mariadb (nexus3.onap.org:10001/mariadb:10.10.2)... 10.10.2: Pulling from mariadb Digest: sha256:bfc25a68e113de43d0d112f5a7126df8e278579c3224e3923359e1c1d8d5ce6e Status: Downloaded newer image for nexus3.onap.org:10001/mariadb:10.10.2 Pulling simulator (nexus3.onap.org:10001/onap/policy-models-simulator:3.1.2-SNAPSHOT)... 3.1.2-SNAPSHOT: Pulling from onap/policy-models-simulator Digest: sha256:d8f1d8ae67fc0b53114a44577cb43c90a3a3281908d2f2418d7fbd203413bd6a Status: Downloaded newer image for nexus3.onap.org:10001/onap/policy-models-simulator:3.1.2-SNAPSHOT Pulling zookeeper (confluentinc/cp-zookeeper:latest)... latest: Pulling from confluentinc/cp-zookeeper Digest: sha256:4dc780642bfc5ec3a2d4901e2ff1f9ddef7f7c5c0b793e1e2911cbfb4e3a3214 Status: Downloaded newer image for confluentinc/cp-zookeeper:latest Pulling kafka (confluentinc/cp-kafka:latest)... latest: Pulling from confluentinc/cp-kafka Digest: sha256:620734d9fc0bb1f9886932e5baf33806074469f40e3fe246a3fdbb59309535fa Status: Downloaded newer image for confluentinc/cp-kafka:latest Pulling policy-db-migrator (nexus3.onap.org:10001/onap/policy-db-migrator:3.1.2-SNAPSHOT)... 3.1.2-SNAPSHOT: Pulling from onap/policy-db-migrator Digest: sha256:dbedf7b51feded6305e873f6f27b64b1e808f60e39ac58976ca87dd8559e4887 Status: Downloaded newer image for nexus3.onap.org:10001/onap/policy-db-migrator:3.1.2-SNAPSHOT Pulling api (nexus3.onap.org:10001/onap/policy-api:3.1.2-SNAPSHOT)... 3.1.2-SNAPSHOT: Pulling from onap/policy-api Digest: sha256:0e8cbccfee833c5b2be68d71dd51902b884e77df24bbbac2751693f58bdc20ce Status: Downloaded newer image for nexus3.onap.org:10001/onap/policy-api:3.1.2-SNAPSHOT Pulling pap (nexus3.onap.org:10001/onap/policy-pap:3.1.2-SNAPSHOT)... 3.1.2-SNAPSHOT: Pulling from onap/policy-pap Digest: sha256:4424490684da433df5069c1f1dbbafe83fffd4c8b6a174807fb10d6443ecef06 Status: Downloaded newer image for nexus3.onap.org:10001/onap/policy-pap:3.1.2-SNAPSHOT Pulling apex-pdp (nexus3.onap.org:10001/onap/policy-apex-pdp:3.1.2-SNAPSHOT)... 3.1.2-SNAPSHOT: Pulling from onap/policy-apex-pdp Digest: sha256:75a74a87b7345e553563fbe2ececcd2285ed9500fd91489d9968ae81123b9982 Status: Downloaded newer image for nexus3.onap.org:10001/onap/policy-apex-pdp:3.1.2-SNAPSHOT Creating prometheus ... Creating mariadb ... Creating zookeeper ... Creating simulator ... Creating prometheus ... done Creating grafana ... Creating grafana ... done Creating mariadb ... done Creating policy-db-migrator ... Creating policy-db-migrator ... done Creating policy-api ... Creating zookeeper ... done Creating kafka ... Creating policy-api ... done Creating simulator ... done Creating kafka ... done Creating policy-pap ... Creating policy-pap ... done Creating policy-apex-pdp ... Creating policy-apex-pdp ... done +++ echo 'Prometheus server: http://localhost:30259' Prometheus server: http://localhost:30259 +++ echo 'Grafana server: http://localhost:30269' Grafana server: http://localhost:30269 +++ cd /w/workspace/policy-pap-master-project-csit-pap ++ sleep 10 ++ unset http_proxy https_proxy ++ bash /w/workspace/policy-pap-master-project-csit-pap/csit/resources/scripts/wait_for_rest.sh localhost 30003 Waiting for REST to come up on localhost port 30003... NAMES STATUS policy-apex-pdp Up 10 seconds policy-pap Up 11 seconds kafka Up 12 seconds policy-api Up 14 seconds grafana Up 18 seconds simulator Up 13 seconds zookeeper Up 15 seconds mariadb Up 17 seconds prometheus Up 19 seconds NAMES STATUS policy-apex-pdp Up 15 seconds policy-pap Up 16 seconds kafka Up 17 seconds policy-api Up 19 seconds grafana Up 23 seconds simulator Up 18 seconds zookeeper Up 20 seconds mariadb Up 22 seconds prometheus Up 24 seconds NAMES STATUS policy-apex-pdp Up 20 seconds policy-pap Up 21 seconds kafka Up 22 seconds policy-api Up 24 seconds grafana Up 28 seconds simulator Up 23 seconds zookeeper Up 25 seconds mariadb Up 27 seconds prometheus Up 29 seconds NAMES STATUS policy-apex-pdp Up 25 seconds policy-pap Up 26 seconds kafka Up 27 seconds policy-api Up 29 seconds grafana Up 33 seconds simulator Up 28 seconds zookeeper Up 30 seconds mariadb Up 32 seconds prometheus Up 34 seconds NAMES STATUS policy-apex-pdp Up 30 seconds policy-pap Up 31 seconds kafka Up 32 seconds policy-api Up 34 seconds grafana Up 38 seconds simulator Up 33 seconds zookeeper Up 35 seconds mariadb Up 37 seconds prometheus Up 39 seconds NAMES STATUS policy-apex-pdp Up 35 seconds policy-pap Up 36 seconds kafka Up 37 seconds policy-api Up 39 seconds grafana Up 43 seconds simulator Up 38 seconds zookeeper Up 40 seconds mariadb Up 42 seconds prometheus Up 44 seconds ++ export 'SUITES=pap-test.robot pap-slas.robot' ++ SUITES='pap-test.robot pap-slas.robot' ++ ROBOT_VARIABLES='-v POLICY_PAP_IP:localhost:30003 -v POLICY_API_IP:localhost:30002 -v PROMETHEUS_IP:localhost:30259 -v DATA:/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/policies -v NODETEMPLATES:/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/nodetemplates' + load_set + _setopts=hxB ++ echo braceexpand:hashall:interactive-comments:xtrace ++ tr : ' ' + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o braceexpand + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o hashall + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o interactive-comments + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o xtrace ++ echo hxB ++ sed 's/./& /g' + for i in $(echo "$_setopts" | sed 's/./& /g') + set +h + for i in $(echo "$_setopts" | sed 's/./& /g') + set +x + docker_stats + tee /w/workspace/policy-pap-master-project-csit-pap/csit/archives/pap/_sysinfo-1-after-setup.txt ++ uname -s + '[' Linux == Darwin ']' + sh -c 'top -bn1 | head -3' top - 09:38:21 up 6 min, 0 users, load average: 3.14, 1.94, 0.89 Tasks: 203 total, 1 running, 131 sleeping, 0 stopped, 0 zombie %Cpu(s): 10.1 us, 2.2 sy, 0.0 ni, 77.6 id, 10.0 wa, 0.0 hi, 0.0 si, 0.0 st + echo + sh -c 'free -h' total used free shared buff/cache available Mem: 31G 2.6G 22G 1.3M 6.2G 28G Swap: 1.0G 0B 1.0G + echo + docker ps --format 'table {{ .Names }}\t{{ .Status }}' NAMES STATUS policy-apex-pdp Up 35 seconds policy-pap Up 36 seconds kafka Up 37 seconds policy-api Up 39 seconds grafana Up 44 seconds simulator Up 38 seconds zookeeper Up 40 seconds mariadb Up 42 seconds prometheus Up 45 seconds + echo + docker stats --no-stream CONTAINER ID NAME CPU % MEM USAGE / LIMIT MEM % NET I/O BLOCK I/O PIDS 9e3714c08494 policy-apex-pdp 7.75% 185.9MiB / 31.41GiB 0.58% 8.98kB / 8.53kB 0B / 0B 48 285d6e7e8e35 policy-pap 27.86% 512.1MiB / 31.41GiB 1.59% 33.9kB / 36kB 0B / 149MB 62 de58d82459ba kafka 56.41% 380.8MiB / 31.41GiB 1.18% 77.1kB / 79kB 0B / 508kB 85 7b3030160d60 policy-api 0.12% 438.7MiB / 31.41GiB 1.36% 988kB / 647kB 0B / 0B 53 5216b06b0fa7 grafana 0.04% 54.73MiB / 31.41GiB 0.17% 19.7kB / 3.49kB 0B / 24.9MB 17 fe0bb1f44e93 simulator 0.08% 123.3MiB / 31.41GiB 0.38% 1.15kB / 0B 0B / 0B 76 737a1873bb73 zookeeper 0.12% 96.11MiB / 31.41GiB 0.30% 57.3kB / 52kB 0B / 369kB 59 6122865c1a80 mariadb 0.02% 102MiB / 31.41GiB 0.32% 935kB / 1.18MB 11MB / 68.2MB 37 755071c4ee29 prometheus 0.96% 19.34MiB / 31.41GiB 0.06% 1.99kB / 158B 106kB / 0B 13 + echo + cd /tmp/tmp.UcCaUoaeAm + echo 'Reading the testplan:' Reading the testplan: + echo 'pap-test.robot pap-slas.robot' + egrep -v '(^[[:space:]]*#|^[[:space:]]*$)' + sed 's|^|/w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/|' + cat testplan.txt /w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-test.robot /w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-slas.robot ++ xargs + SUITES='/w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-test.robot /w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-slas.robot' + echo 'ROBOT_VARIABLES=-v POLICY_PAP_IP:localhost:30003 -v POLICY_API_IP:localhost:30002 -v PROMETHEUS_IP:localhost:30259 -v DATA:/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/policies -v NODETEMPLATES:/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/nodetemplates' ROBOT_VARIABLES=-v POLICY_PAP_IP:localhost:30003 -v POLICY_API_IP:localhost:30002 -v PROMETHEUS_IP:localhost:30259 -v DATA:/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/policies -v NODETEMPLATES:/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/nodetemplates + echo 'Starting Robot test suites /w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-test.robot /w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-slas.robot ...' Starting Robot test suites /w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-test.robot /w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-slas.robot ... + relax_set + set +e + set +o pipefail + python3 -m robot.run -N pap -v WORKSPACE:/tmp -v POLICY_PAP_IP:localhost:30003 -v POLICY_API_IP:localhost:30002 -v PROMETHEUS_IP:localhost:30259 -v DATA:/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/policies -v NODETEMPLATES:/w/workspace/policy-pap-master-project-csit-pap/models/models-examples/src/main/resources/nodetemplates /w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-test.robot /w/workspace/policy-pap-master-project-csit-pap/csit/resources/tests/pap-slas.robot ============================================================================== pap ============================================================================== pap.Pap-Test ============================================================================== LoadPolicy :: Create a policy named 'onap.restart.tca' and version... | PASS | ------------------------------------------------------------------------------ LoadPolicyWithMetadataSet :: Create a policy named 'operational.ap... | PASS | ------------------------------------------------------------------------------ LoadNodeTemplates :: Create node templates in database using speci... | PASS | ------------------------------------------------------------------------------ Healthcheck :: Verify policy pap health check | PASS | ------------------------------------------------------------------------------ Consolidated Healthcheck :: Verify policy consolidated health check | PASS | ------------------------------------------------------------------------------ Metrics :: Verify policy pap is exporting prometheus metrics | PASS | ------------------------------------------------------------------------------ AddPdpGroup :: Add a new PdpGroup named 'testGroup' in the policy ... | PASS | ------------------------------------------------------------------------------ QueryPdpGroupsBeforeActivation :: Verify PdpGroups before activation | PASS | ------------------------------------------------------------------------------ ActivatePdpGroup :: Change the state of PdpGroup named 'testGroup'... | PASS | ------------------------------------------------------------------------------ QueryPdpGroupsAfterActivation :: Verify PdpGroups after activation | PASS | ------------------------------------------------------------------------------ DeployPdpGroups :: Deploy policies in PdpGroups | PASS | ------------------------------------------------------------------------------ QueryPdpGroupsAfterDeploy :: Verify PdpGroups after undeploy | PASS | ------------------------------------------------------------------------------ QueryPolicyAuditAfterDeploy :: Verify policy audit record after de... | PASS | ------------------------------------------------------------------------------ QueryPolicyAuditWithMetadataSetAfterDeploy :: Verify policy audit ... | PASS | ------------------------------------------------------------------------------ UndeployPolicy :: Undeploy a policy named 'onap.restart.tca' from ... | PASS | ------------------------------------------------------------------------------ UndeployPolicyWithMetadataSet :: Undeploy a policy named 'operatio... | PASS | ------------------------------------------------------------------------------ QueryPdpGroupsAfterUndeploy :: Verify PdpGroups after undeploy | PASS | ------------------------------------------------------------------------------ QueryPolicyAuditAfterUnDeploy :: Verify policy audit record after ... | PASS | ------------------------------------------------------------------------------ QueryPolicyAuditWithMetadataSetAfterUnDeploy :: Verify policy audi... | PASS | ------------------------------------------------------------------------------ DeactivatePdpGroup :: Change the state of PdpGroup named 'testGrou... | PASS | ------------------------------------------------------------------------------ DeletePdpGroups :: Delete the PdpGroup named 'testGroup' from poli... | PASS | ------------------------------------------------------------------------------ QueryPdpGroupsAfterDelete :: Verify PdpGroups after delete | PASS | ------------------------------------------------------------------------------ pap.Pap-Test | PASS | 22 tests, 22 passed, 0 failed ============================================================================== pap.Pap-Slas ============================================================================== WaitForPrometheusServer :: Wait for Prometheus server to gather al... | PASS | ------------------------------------------------------------------------------ ValidateResponseTimeForHealthcheck :: Validate component healthche... | PASS | ------------------------------------------------------------------------------ ValidateResponseTimeForSystemHealthcheck :: Validate if system hea... | PASS | ------------------------------------------------------------------------------ ValidateResponseTimeQueryPolicyAudit :: Validate query audits resp... | PASS | ------------------------------------------------------------------------------ ValidateResponseTimeUpdateGroup :: Validate pdps/group response time | PASS | ------------------------------------------------------------------------------ ValidatePolicyDeploymentTime :: Check if deployment of policy is u... | PASS | ------------------------------------------------------------------------------ ValidateResponseTimeDeletePolicy :: Check if undeployment of polic... | PASS | ------------------------------------------------------------------------------ ValidateResponseTimeDeleteGroup :: Validate delete group response ... | PASS | ------------------------------------------------------------------------------ pap.Pap-Slas | PASS | 8 tests, 8 passed, 0 failed ============================================================================== pap | PASS | 30 tests, 30 passed, 0 failed ============================================================================== Output: /tmp/tmp.UcCaUoaeAm/output.xml Log: /tmp/tmp.UcCaUoaeAm/log.html Report: /tmp/tmp.UcCaUoaeAm/report.html + RESULT=0 + load_set + _setopts=hxB ++ echo braceexpand:hashall:interactive-comments:xtrace ++ tr : ' ' + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o braceexpand + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o hashall + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o interactive-comments + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o xtrace ++ echo hxB ++ sed 's/./& /g' + for i in $(echo "$_setopts" | sed 's/./& /g') + set +h + for i in $(echo "$_setopts" | sed 's/./& /g') + set +x + echo 'RESULT: 0' RESULT: 0 + exit 0 + on_exit + rc=0 + [[ -n /w/workspace/policy-pap-master-project-csit-pap ]] + docker ps --format 'table {{ .Names }}\t{{ .Status }}' NAMES STATUS policy-apex-pdp Up 2 minutes policy-pap Up 2 minutes kafka Up 2 minutes policy-api Up 2 minutes grafana Up 2 minutes simulator Up 2 minutes zookeeper Up 2 minutes mariadb Up 2 minutes prometheus Up 2 minutes + docker_stats ++ uname -s + '[' Linux == Darwin ']' + sh -c 'top -bn1 | head -3' top - 09:40:10 up 8 min, 0 users, load average: 0.71, 1.45, 0.83 Tasks: 201 total, 1 running, 129 sleeping, 0 stopped, 0 zombie %Cpu(s): 8.7 us, 1.8 sy, 0.0 ni, 81.5 id, 7.9 wa, 0.0 hi, 0.0 si, 0.0 st + echo + sh -c 'free -h' total used free shared buff/cache available Mem: 31G 2.6G 22G 1.3M 6.2G 28G Swap: 1.0G 0B 1.0G + echo + docker ps --format 'table {{ .Names }}\t{{ .Status }}' NAMES STATUS policy-apex-pdp Up 2 minutes policy-pap Up 2 minutes kafka Up 2 minutes policy-api Up 2 minutes grafana Up 2 minutes simulator Up 2 minutes zookeeper Up 2 minutes mariadb Up 2 minutes prometheus Up 2 minutes + echo + docker stats --no-stream CONTAINER ID NAME CPU % MEM USAGE / LIMIT MEM % NET I/O BLOCK I/O PIDS 9e3714c08494 policy-apex-pdp 0.31% 182.9MiB / 31.41GiB 0.57% 56.5kB / 90.7kB 0B / 0B 52 285d6e7e8e35 policy-pap 0.55% 461.2MiB / 31.41GiB 1.43% 2.47MB / 1.04MB 0B / 149MB 65 de58d82459ba kafka 11.48% 389.1MiB / 31.41GiB 1.21% 243kB / 217kB 0B / 606kB 85 7b3030160d60 policy-api 0.10% 482.7MiB / 31.41GiB 1.50% 2.45MB / 1.1MB 0B / 0B 56 5216b06b0fa7 grafana 0.12% 64.38MiB / 31.41GiB 0.20% 20.6kB / 4.43kB 0B / 24.9MB 17 fe0bb1f44e93 simulator 0.11% 123.4MiB / 31.41GiB 0.38% 1.45kB / 0B 0B / 0B 78 737a1873bb73 zookeeper 0.10% 97.28MiB / 31.41GiB 0.30% 60.2kB / 53.5kB 0B / 369kB 59 6122865c1a80 mariadb 0.02% 103.2MiB / 31.41GiB 0.32% 2.02MB / 4.87MB 11MB / 68.4MB 28 755071c4ee29 prometheus 0.00% 25.2MiB / 31.41GiB 0.08% 185kB / 10.8kB 106kB / 0B 14 + echo + source_safely /w/workspace/policy-pap-master-project-csit-pap/compose/stop-compose.sh + '[' -z /w/workspace/policy-pap-master-project-csit-pap/compose/stop-compose.sh ']' + relax_set + set +e + set +o pipefail + . /w/workspace/policy-pap-master-project-csit-pap/compose/stop-compose.sh ++ echo 'Shut down started!' Shut down started! ++ '[' -z /w/workspace/policy-pap-master-project-csit-pap ']' ++ COMPOSE_FOLDER=/w/workspace/policy-pap-master-project-csit-pap/compose ++ cd /w/workspace/policy-pap-master-project-csit-pap/compose ++ source export-ports.sh ++ source get-versions.sh ++ echo 'Collecting logs from docker compose containers...' Collecting logs from docker compose containers... ++ docker-compose logs ++ cat docker_compose.log Attaching to policy-apex-pdp, policy-pap, kafka, policy-api, policy-db-migrator, grafana, simulator, zookeeper, mariadb, prometheus mariadb | 2024-04-24 09:37:38+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:10.10.2+maria~ubu2204 started. mariadb | 2024-04-24 09:37:38+00:00 [Note] [Entrypoint]: Switching to dedicated user 'mysql' mariadb | 2024-04-24 09:37:38+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:10.10.2+maria~ubu2204 started. mariadb | 2024-04-24 09:37:38+00:00 [Note] [Entrypoint]: Initializing database files mariadb | 2024-04-24 9:37:38 0 [Warning] mariadbd: io_uring_queue_init() failed with ENOSYS: check seccomp filters, and the kernel version (newer than 5.1 required) mariadb | 2024-04-24 9:37:38 0 [Warning] InnoDB: liburing disabled: falling back to innodb_use_native_aio=OFF mariadb | 2024-04-24 9:37:38 0 [Warning] 'default-authentication-plugin' is MySQL 5.6 / 5.7 compatible option. To be implemented in later versions. mariadb | mariadb | mariadb | PLEASE REMEMBER TO SET A PASSWORD FOR THE MariaDB root USER ! mariadb | To do so, start the server, then issue the following command: mariadb | mariadb | '/usr/bin/mysql_secure_installation' mariadb | mariadb | which will also give you the option of removing the test mariadb | databases and anonymous user created by default. This is mariadb | strongly recommended for production servers. mariadb | mariadb | See the MariaDB Knowledgebase at https://mariadb.com/kb mariadb | mariadb | Please report any problems at https://mariadb.org/jira mariadb | mariadb | The latest information about MariaDB is available at https://mariadb.org/. mariadb | mariadb | Consider joining MariaDB's strong and vibrant community: mariadb | https://mariadb.org/get-involved/ mariadb | mariadb | 2024-04-24 09:37:40+00:00 [Note] [Entrypoint]: Database files initialized mariadb | 2024-04-24 09:37:40+00:00 [Note] [Entrypoint]: Starting temporary server mariadb | 2024-04-24 09:37:40+00:00 [Note] [Entrypoint]: Waiting for server startup mariadb | 2024-04-24 9:37:40 0 [Note] mariadbd (server 10.10.2-MariaDB-1:10.10.2+maria~ubu2204) starting as process 95 ... mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: Compressed tables use zlib 1.2.11 mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: Number of transaction pools: 1 mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: Using crc32 + pclmulqdq instructions mariadb | 2024-04-24 9:37:40 0 [Note] mariadbd: O_TMPFILE is not supported on /tmp (disabling future attempts) mariadb | 2024-04-24 9:37:40 0 [Warning] mariadbd: io_uring_queue_init() failed with ENOSYS: check seccomp filters, and the kernel version (newer than 5.1 required) mariadb | 2024-04-24 9:37:40 0 [Warning] InnoDB: liburing disabled: falling back to innodb_use_native_aio=OFF mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: Completed initialization of buffer pool mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes) mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: 128 rollback segments are active. mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ... mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB. mariadb | 2024-04-24 9:37:40 0 [Note] InnoDB: log sequence number 46590; transaction id 14 mariadb | 2024-04-24 9:37:40 0 [Note] Plugin 'FEEDBACK' is disabled. mariadb | 2024-04-24 9:37:40 0 [Warning] 'default-authentication-plugin' is MySQL 5.6 / 5.7 compatible option. To be implemented in later versions. mariadb | 2024-04-24 9:37:40 0 [Warning] 'user' entry 'root@mariadb' ignored in --skip-name-resolve mode. mariadb | 2024-04-24 9:37:40 0 [Warning] 'proxies_priv' entry '@% root@mariadb' ignored in --skip-name-resolve mode. mariadb | 2024-04-24 9:37:40 0 [Note] mariadbd: ready for connections. mariadb | Version: '10.10.2-MariaDB-1:10.10.2+maria~ubu2204' socket: '/run/mysqld/mysqld.sock' port: 0 mariadb.org binary distribution mariadb | 2024-04-24 09:37:41+00:00 [Note] [Entrypoint]: Temporary server started. mariadb | 2024-04-24 09:37:43+00:00 [Note] [Entrypoint]: Creating user policy_user mariadb | 2024-04-24 09:37:43+00:00 [Note] [Entrypoint]: Securing system users (equivalent to running mysql_secure_installation) mariadb | mariadb | mariadb | 2024-04-24 09:37:43+00:00 [Warn] [Entrypoint]: /usr/local/bin/docker-entrypoint.sh: ignoring /docker-entrypoint-initdb.d/db.conf mariadb | 2024-04-24 09:37:43+00:00 [Note] [Entrypoint]: /usr/local/bin/docker-entrypoint.sh: running /docker-entrypoint-initdb.d/db.sh mariadb | #!/bin/bash -xv mariadb | # Copyright 2019,2021 AT&T Intellectual Property. All rights reserved mariadb | # Modifications Copyright (c) 2022 Nordix Foundation. mariadb | # mariadb | # Licensed under the Apache License, Version 2.0 (the "License"); mariadb | # you may not use this file except in compliance with the License. mariadb | # You may obtain a copy of the License at mariadb | # mariadb | # http://www.apache.org/licenses/LICENSE-2.0 mariadb | # mariadb | # Unless required by applicable law or agreed to in writing, software mariadb | # distributed under the License is distributed on an "AS IS" BASIS, mariadb | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. mariadb | # See the License for the specific language governing permissions and mariadb | # limitations under the License. mariadb | mariadb | for db in migration pooling policyadmin operationshistory clampacm policyclamp mariadb | do mariadb | mysql -uroot -p"${MYSQL_ROOT_PASSWORD}" --execute "CREATE DATABASE IF NOT EXISTS ${db};" mariadb | mysql -uroot -p"${MYSQL_ROOT_PASSWORD}" --execute "GRANT ALL PRIVILEGES ON \`${db}\`.* TO '${MYSQL_USER}'@'%' ;" mariadb | done mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS migration;' mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `migration`.* TO '\''policy_user'\''@'\''%'\'' ;' mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS pooling;' mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `pooling`.* TO '\''policy_user'\''@'\''%'\'' ;' mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS policyadmin;' policy-api | Waiting for mariadb port 3306... policy-api | mariadb (172.17.0.4:3306) open policy-api | Waiting for policy-db-migrator port 6824... policy-api | policy-db-migrator (172.17.0.7:6824) open policy-api | Policy api config file: /opt/app/policy/api/etc/apiParameters.yaml policy-api | policy-api | . ____ _ __ _ _ policy-api | /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \ policy-api | ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \ policy-api | \\/ ___)| |_)| | | | | || (_| | ) ) ) ) policy-api | ' |____| .__|_| |_|_| |_\__, | / / / / policy-api | =========|_|==============|___/=/_/_/_/ policy-api | :: Spring Boot :: (v3.1.10) policy-api | policy-api | [2024-04-24T09:37:55.012+00:00|INFO|Version|background-preinit] HV000001: Hibernate Validator 8.0.1.Final policy-api | [2024-04-24T09:37:55.105+00:00|INFO|PolicyApiApplication|main] Starting PolicyApiApplication using Java 17.0.10 with PID 22 (/app/api.jar started by policy in /opt/app/policy/api/bin) policy-api | [2024-04-24T09:37:55.107+00:00|INFO|PolicyApiApplication|main] No active profile set, falling back to 1 default profile: "default" policy-api | [2024-04-24T09:37:57.286+00:00|INFO|RepositoryConfigurationDelegate|main] Bootstrapping Spring Data JPA repositories in DEFAULT mode. policy-api | [2024-04-24T09:37:57.375+00:00|INFO|RepositoryConfigurationDelegate|main] Finished Spring Data repository scanning in 77 ms. Found 6 JPA repository interfaces. policy-api | [2024-04-24T09:37:57.839+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.api.main.exception.ServiceExceptionHandler policy-api | [2024-04-24T09:37:57.840+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.api.main.exception.ServiceExceptionHandler policy-api | [2024-04-24T09:37:58.550+00:00|INFO|TomcatWebServer|main] Tomcat initialized with port(s): 6969 (http) policy-api | [2024-04-24T09:37:58.561+00:00|INFO|Http11NioProtocol|main] Initializing ProtocolHandler ["http-nio-6969"] policy-api | [2024-04-24T09:37:58.563+00:00|INFO|StandardService|main] Starting service [Tomcat] policy-api | [2024-04-24T09:37:58.563+00:00|INFO|StandardEngine|main] Starting Servlet engine: [Apache Tomcat/10.1.19] policy-api | [2024-04-24T09:37:58.664+00:00|INFO|[/policy/api/v1]|main] Initializing Spring embedded WebApplicationContext policy-api | [2024-04-24T09:37:58.664+00:00|INFO|ServletWebServerApplicationContext|main] Root WebApplicationContext: initialization completed in 3475 ms policy-api | [2024-04-24T09:37:59.134+00:00|INFO|LogHelper|main] HHH000204: Processing PersistenceUnitInfo [name: default] policy-api | [2024-04-24T09:37:59.215+00:00|INFO|Version|main] HHH000412: Hibernate ORM core version 6.3.2.Final policy-api | [2024-04-24T09:37:59.267+00:00|INFO|RegionFactoryInitiator|main] HHH000026: Second-level cache disabled policy-api | [2024-04-24T09:37:59.566+00:00|INFO|SpringPersistenceUnitInfo|main] No LoadTimeWeaver setup: ignoring JPA class transformer policy-api | [2024-04-24T09:37:59.599+00:00|INFO|HikariDataSource|main] HikariPool-1 - Starting... policy-api | [2024-04-24T09:37:59.709+00:00|INFO|HikariPool|main] HikariPool-1 - Added connection org.mariadb.jdbc.Connection@65a48602 policy-api | [2024-04-24T09:37:59.711+00:00|INFO|HikariDataSource|main] HikariPool-1 - Start completed. policy-api | [2024-04-24T09:38:01.928+00:00|INFO|JtaPlatformInitiator|main] HHH000489: No JTA platform available (set 'hibernate.transaction.jta.platform' to enable JTA platform integration) policy-api | [2024-04-24T09:38:01.931+00:00|INFO|LocalContainerEntityManagerFactoryBean|main] Initialized JPA EntityManagerFactory for persistence unit 'default' policy-api | [2024-04-24T09:38:02.990+00:00|WARN|ApiDatabaseInitializer|main] Detected multi-versioned type: policytypes/onap.policies.monitoring.tcagen2.v2.yaml policy-api | [2024-04-24T09:38:03.780+00:00|INFO|ApiDatabaseInitializer|main] Multi-versioned Service Template [onap.policies.Monitoring, onap.policies.monitoring.tcagen2] policy-api | [2024-04-24T09:38:04.875+00:00|WARN|JpaBaseConfiguration$JpaWebConfiguration|main] spring.jpa.open-in-view is enabled by default. Therefore, database queries may be performed during view rendering. Explicitly configure spring.jpa.open-in-view to disable this warning policy-api | [2024-04-24T09:38:05.094+00:00|INFO|DefaultSecurityFilterChain|main] Will secure any request with [org.springframework.security.web.session.DisableEncodeUrlFilter@6e00a1ba, org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter@4639d92c, org.springframework.security.web.context.SecurityContextHolderFilter@5f966dee, org.springframework.security.web.header.HeaderWriterFilter@4d71deeb, org.springframework.security.web.authentication.logout.LogoutFilter@446717fb, org.springframework.security.web.authentication.www.BasicAuthenticationFilter@280aa1bd, org.springframework.security.web.savedrequest.RequestCacheAwareFilter@fffdd40, org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter@42ca6733, org.springframework.security.web.authentication.AnonymousAuthenticationFilter@46270641, org.springframework.security.web.access.ExceptionTranslationFilter@80ca2c3, org.springframework.security.web.access.intercept.AuthorizationFilter@2e95d163] policy-api | [2024-04-24T09:38:06.133+00:00|INFO|EndpointLinksResolver|main] Exposing 3 endpoint(s) beneath base path '' policy-api | [2024-04-24T09:38:06.248+00:00|INFO|Http11NioProtocol|main] Starting ProtocolHandler ["http-nio-6969"] policy-api | [2024-04-24T09:38:06.270+00:00|INFO|TomcatWebServer|main] Tomcat started on port(s): 6969 (http) with context path '/policy/api/v1' policy-api | [2024-04-24T09:38:06.291+00:00|INFO|PolicyApiApplication|main] Started PolicyApiApplication in 12.14 seconds (process running for 12.848) policy-api | [2024-04-24T09:38:24.289+00:00|INFO|[/policy/api/v1]|http-nio-6969-exec-2] Initializing Spring DispatcherServlet 'dispatcherServlet' policy-api | [2024-04-24T09:38:24.289+00:00|INFO|DispatcherServlet|http-nio-6969-exec-2] Initializing Servlet 'dispatcherServlet' policy-api | [2024-04-24T09:38:24.291+00:00|INFO|DispatcherServlet|http-nio-6969-exec-2] Completed initialization in 2 ms policy-api | [2024-04-24T09:38:24.609+00:00|INFO|OrderedServiceImpl|http-nio-6969-exec-2] ***** OrderedServiceImpl implementers: policy-api | [] policy-apex-pdp | Waiting for mariadb port 3306... policy-apex-pdp | mariadb (172.17.0.4:3306) open policy-apex-pdp | Waiting for kafka port 9092... policy-apex-pdp | Waiting for pap port 6969... policy-apex-pdp | kafka (172.17.0.9:9092) open policy-apex-pdp | pap (172.17.0.10:6969) open policy-apex-pdp | apexApps.sh: running application 'onappf' with command 'java -Dlogback.configurationFile=/opt/app/policy/apex-pdp/etc/logback.xml -cp /opt/app/policy/apex-pdp/etc:/opt/app/policy/apex-pdp/etc/hazelcast:/opt/app/policy/apex-pdp/etc/infinispan:/opt/app/policy/apex-pdp/lib/* -Djavax.net.ssl.keyStore=/opt/app/policy/apex-pdp/etc/ssl/policy-keystore -Djavax.net.ssl.keyStorePassword=Pol1cy_0nap -Djavax.net.ssl.trustStore=/opt/app/policy/apex-pdp/etc/ssl/policy-truststore -Djavax.net.ssl.trustStorePassword=Pol1cy_0nap -Dlogback.configurationFile=/opt/app/policy/apex-pdp/etc/logback.xml -Dhazelcast.config=/opt/app/policy/apex-pdp/etc/hazelcast.xml -Dhazelcast.mancenter.enabled=false org.onap.policy.apex.services.onappf.ApexStarterMain -c /opt/app/policy/apex-pdp/etc/onappf/config/OnapPfConfig.json' policy-apex-pdp | [2024-04-24T09:38:19.412+00:00|INFO|ApexStarterMain|main] In ApexStarter with parameters [-c, /opt/app/policy/apex-pdp/etc/onappf/config/OnapPfConfig.json] policy-apex-pdp | [2024-04-24T09:38:19.627+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-apex-pdp | allow.auto.create.topics = true policy-apex-pdp | auto.commit.interval.ms = 5000 policy-apex-pdp | auto.include.jmx.reporter = true policy-apex-pdp | auto.offset.reset = latest policy-apex-pdp | bootstrap.servers = [kafka:9092] policy-apex-pdp | check.crcs = true policy-apex-pdp | client.dns.lookup = use_all_dns_ips policy-apex-pdp | client.id = consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-1 policy-apex-pdp | client.rack = policy-apex-pdp | connections.max.idle.ms = 540000 policy-apex-pdp | default.api.timeout.ms = 60000 policy-apex-pdp | enable.auto.commit = true policy-apex-pdp | exclude.internal.topics = true policy-apex-pdp | fetch.max.bytes = 52428800 policy-apex-pdp | fetch.max.wait.ms = 500 policy-apex-pdp | fetch.min.bytes = 1 policy-apex-pdp | group.id = ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0 policy-apex-pdp | group.instance.id = null policy-apex-pdp | heartbeat.interval.ms = 3000 policy-apex-pdp | interceptor.classes = [] policy-apex-pdp | internal.leave.group.on.close = true policy-apex-pdp | internal.throw.on.fetch.stable.offset.unsupported = false policy-apex-pdp | isolation.level = read_uncommitted policy-apex-pdp | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-apex-pdp | max.partition.fetch.bytes = 1048576 policy-apex-pdp | max.poll.interval.ms = 300000 policy-apex-pdp | max.poll.records = 500 policy-apex-pdp | metadata.max.age.ms = 300000 policy-apex-pdp | metric.reporters = [] policy-apex-pdp | metrics.num.samples = 2 policy-apex-pdp | metrics.recording.level = INFO policy-apex-pdp | metrics.sample.window.ms = 30000 policy-apex-pdp | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-apex-pdp | receive.buffer.bytes = 65536 policy-apex-pdp | reconnect.backoff.max.ms = 1000 policy-apex-pdp | reconnect.backoff.ms = 50 policy-apex-pdp | request.timeout.ms = 30000 policy-apex-pdp | retry.backoff.ms = 100 mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `policyadmin`.* TO '\''policy_user'\''@'\''%'\'' ;' mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS operationshistory;' mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `operationshistory`.* TO '\''policy_user'\''@'\''%'\'' ;' mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS clampacm;' mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `clampacm`.* TO '\''policy_user'\''@'\''%'\'' ;' mariadb | + for db in migration pooling policyadmin operationshistory clampacm policyclamp mariadb | + mysql -uroot -psecret --execute 'CREATE DATABASE IF NOT EXISTS policyclamp;' mariadb | + mysql -uroot -psecret --execute 'GRANT ALL PRIVILEGES ON `policyclamp`.* TO '\''policy_user'\''@'\''%'\'' ;' mariadb | mariadb | mysql -uroot -p"${MYSQL_ROOT_PASSWORD}" --execute "FLUSH PRIVILEGES;" mariadb | + mysql -uroot -psecret --execute 'FLUSH PRIVILEGES;' policy-apex-pdp | sasl.client.callback.handler.class = null policy-apex-pdp | sasl.jaas.config = null policy-apex-pdp | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-apex-pdp | sasl.kerberos.min.time.before.relogin = 60000 policy-apex-pdp | sasl.kerberos.service.name = null policy-apex-pdp | sasl.kerberos.ticket.renew.jitter = 0.05 policy-apex-pdp | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-apex-pdp | sasl.login.callback.handler.class = null policy-apex-pdp | sasl.login.class = null policy-apex-pdp | sasl.login.connect.timeout.ms = null policy-apex-pdp | sasl.login.read.timeout.ms = null policy-apex-pdp | sasl.login.refresh.buffer.seconds = 300 policy-apex-pdp | sasl.login.refresh.min.period.seconds = 60 policy-apex-pdp | sasl.login.refresh.window.factor = 0.8 policy-apex-pdp | sasl.login.refresh.window.jitter = 0.05 policy-apex-pdp | sasl.login.retry.backoff.max.ms = 10000 policy-apex-pdp | sasl.login.retry.backoff.ms = 100 policy-apex-pdp | sasl.mechanism = GSSAPI policy-apex-pdp | sasl.oauthbearer.clock.skew.seconds = 30 policy-apex-pdp | sasl.oauthbearer.expected.audience = null policy-apex-pdp | sasl.oauthbearer.expected.issuer = null policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.url = null policy-apex-pdp | sasl.oauthbearer.scope.claim.name = scope policy-apex-pdp | sasl.oauthbearer.sub.claim.name = sub policy-apex-pdp | sasl.oauthbearer.token.endpoint.url = null policy-apex-pdp | security.protocol = PLAINTEXT policy-apex-pdp | security.providers = null policy-apex-pdp | send.buffer.bytes = 131072 policy-apex-pdp | session.timeout.ms = 45000 policy-apex-pdp | socket.connection.setup.timeout.max.ms = 30000 policy-apex-pdp | socket.connection.setup.timeout.ms = 10000 policy-apex-pdp | ssl.cipher.suites = null policy-apex-pdp | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-apex-pdp | ssl.endpoint.identification.algorithm = https policy-apex-pdp | ssl.engine.factory.class = null policy-apex-pdp | ssl.key.password = null policy-apex-pdp | ssl.keymanager.algorithm = SunX509 policy-apex-pdp | ssl.keystore.certificate.chain = null policy-apex-pdp | ssl.keystore.key = null policy-apex-pdp | ssl.keystore.location = null policy-apex-pdp | ssl.keystore.password = null policy-apex-pdp | ssl.keystore.type = JKS policy-apex-pdp | ssl.protocol = TLSv1.3 policy-apex-pdp | ssl.provider = null policy-apex-pdp | ssl.secure.random.implementation = null policy-apex-pdp | ssl.trustmanager.algorithm = PKIX policy-apex-pdp | ssl.truststore.certificates = null policy-apex-pdp | ssl.truststore.location = null policy-apex-pdp | ssl.truststore.password = null policy-apex-pdp | ssl.truststore.type = JKS policy-apex-pdp | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-apex-pdp | policy-apex-pdp | [2024-04-24T09:38:19.781+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 policy-apex-pdp | [2024-04-24T09:38:19.781+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 policy-apex-pdp | [2024-04-24T09:38:19.781+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1713951499779 policy-apex-pdp | [2024-04-24T09:38:19.784+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-1, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Subscribed to topic(s): policy-pdp-pap policy-apex-pdp | [2024-04-24T09:38:19.795+00:00|INFO|ServiceManager|main] service manager starting policy-apex-pdp | [2024-04-24T09:38:19.796+00:00|INFO|ServiceManager|main] service manager starting topics policy-apex-pdp | [2024-04-24T09:38:19.797+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0, consumerInstance=policy-apex-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: starting policy-apex-pdp | [2024-04-24T09:38:19.816+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-apex-pdp | allow.auto.create.topics = true policy-apex-pdp | auto.commit.interval.ms = 5000 policy-apex-pdp | auto.include.jmx.reporter = true policy-apex-pdp | auto.offset.reset = latest policy-apex-pdp | bootstrap.servers = [kafka:9092] policy-apex-pdp | check.crcs = true policy-apex-pdp | client.dns.lookup = use_all_dns_ips policy-apex-pdp | client.id = consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2 policy-apex-pdp | client.rack = mariadb | mysql -u"${MYSQL_USER}" -p"${MYSQL_PASSWORD}" -f policyclamp < /tmp/policy-clamp-create-tables.sql mariadb | + mysql -upolicy_user -ppolicy_user -f policyclamp mariadb | mariadb | 2024-04-24 09:37:44+00:00 [Note] [Entrypoint]: Stopping temporary server mariadb | 2024-04-24 9:37:44 0 [Note] mariadbd (initiated by: unknown): Normal shutdown mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: FTS optimize thread exiting. mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Starting shutdown... mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Dumping buffer pool(s) to /var/lib/mysql/ib_buffer_pool mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Buffer pool(s) dump completed at 240424 9:37:44 mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Removed temporary tablespace data file: "./ibtmp1" mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Shutdown completed; log sequence number 330928; transaction id 298 mariadb | 2024-04-24 9:37:44 0 [Note] mariadbd: Shutdown complete mariadb | mariadb | 2024-04-24 09:37:44+00:00 [Note] [Entrypoint]: Temporary server stopped mariadb | mariadb | 2024-04-24 09:37:44+00:00 [Note] [Entrypoint]: MariaDB init process done. Ready for start up. mariadb | mariadb | 2024-04-24 9:37:44 0 [Note] mariadbd (server 10.10.2-MariaDB-1:10.10.2+maria~ubu2204) starting as process 1 ... mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Compressed tables use zlib 1.2.11 mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Number of transaction pools: 1 mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Using crc32 + pclmulqdq instructions mariadb | 2024-04-24 9:37:44 0 [Note] mariadbd: O_TMPFILE is not supported on /tmp (disabling future attempts) mariadb | 2024-04-24 9:37:44 0 [Warning] mariadbd: io_uring_queue_init() failed with ENOSYS: check seccomp filters, and the kernel version (newer than 5.1 required) mariadb | 2024-04-24 9:37:44 0 [Warning] InnoDB: liburing disabled: falling back to innodb_use_native_aio=OFF mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Initializing buffer pool, total size = 128.000MiB, chunk size = 2.000MiB mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: Completed initialization of buffer pool mariadb | 2024-04-24 9:37:44 0 [Note] InnoDB: File system buffers for log disabled (block size=512 bytes) mariadb | 2024-04-24 9:37:45 0 [Note] InnoDB: 128 rollback segments are active. mariadb | 2024-04-24 9:37:45 0 [Note] InnoDB: Setting file './ibtmp1' size to 12.000MiB. Physically writing the file full; Please wait ... mariadb | 2024-04-24 9:37:45 0 [Note] InnoDB: File './ibtmp1' size is now 12.000MiB. mariadb | 2024-04-24 9:37:45 0 [Note] InnoDB: log sequence number 330928; transaction id 299 mariadb | 2024-04-24 9:37:45 0 [Note] Plugin 'FEEDBACK' is disabled. mariadb | 2024-04-24 9:37:45 0 [Note] InnoDB: Loading buffer pool(s) from /var/lib/mysql/ib_buffer_pool mariadb | 2024-04-24 9:37:45 0 [Warning] 'default-authentication-plugin' is MySQL 5.6 / 5.7 compatible option. To be implemented in later versions. mariadb | 2024-04-24 9:37:45 0 [Warning] You need to use --log-bin to make --expire-logs-days or --binlog-expire-logs-seconds work. mariadb | 2024-04-24 9:37:45 0 [Note] Server socket created on IP: '0.0.0.0'. mariadb | 2024-04-24 9:37:45 0 [Note] Server socket created on IP: '::'. mariadb | 2024-04-24 9:37:45 0 [Note] mariadbd: ready for connections. mariadb | Version: '10.10.2-MariaDB-1:10.10.2+maria~ubu2204' socket: '/run/mysqld/mysqld.sock' port: 3306 mariadb.org binary distribution mariadb | 2024-04-24 9:37:45 0 [Note] InnoDB: Buffer pool(s) load completed at 240424 9:37:45 mariadb | 2024-04-24 9:37:45 3 [Warning] Aborted connection 3 to db: 'unconnected' user: 'unauthenticated' host: '172.17.0.8' (This connection closed normally without authentication) mariadb | 2024-04-24 9:37:45 4 [Warning] Aborted connection 4 to db: 'unconnected' user: 'unauthenticated' host: '172.17.0.7' (This connection closed normally without authentication) grafana | logger=settings t=2024-04-24T09:37:37.289090094Z level=info msg="Starting Grafana" version=10.4.2 commit=701c851be7a930e04fbc6ebb1cd4254da80edd4c branch=v10.4.x compiled=2024-04-24T09:37:37Z grafana | logger=settings t=2024-04-24T09:37:37.289971288Z level=info msg="Config loaded from" file=/usr/share/grafana/conf/defaults.ini grafana | logger=settings t=2024-04-24T09:37:37.29000621Z level=info msg="Config loaded from" file=/etc/grafana/grafana.ini grafana | logger=settings t=2024-04-24T09:37:37.29001686Z level=info msg="Config overridden from command line" arg="default.paths.data=/var/lib/grafana" grafana | logger=settings t=2024-04-24T09:37:37.29002509Z level=info msg="Config overridden from command line" arg="default.paths.logs=/var/log/grafana" grafana | logger=settings t=2024-04-24T09:37:37.290033251Z level=info msg="Config overridden from command line" arg="default.paths.plugins=/var/lib/grafana/plugins" grafana | logger=settings t=2024-04-24T09:37:37.290102293Z level=info msg="Config overridden from command line" arg="default.paths.provisioning=/etc/grafana/provisioning" grafana | logger=settings t=2024-04-24T09:37:37.290114014Z level=info msg="Config overridden from command line" arg="default.log.mode=console" grafana | logger=settings t=2024-04-24T09:37:37.290123064Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_DATA=/var/lib/grafana" grafana | logger=settings t=2024-04-24T09:37:37.290188937Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_LOGS=/var/log/grafana" grafana | logger=settings t=2024-04-24T09:37:37.290240469Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_PLUGINS=/var/lib/grafana/plugins" grafana | logger=settings t=2024-04-24T09:37:37.29026617Z level=info msg="Config overridden from Environment variable" var="GF_PATHS_PROVISIONING=/etc/grafana/provisioning" grafana | logger=settings t=2024-04-24T09:37:37.29027525Z level=info msg=Target target=[all] grafana | logger=settings t=2024-04-24T09:37:37.290356263Z level=info msg="Path Home" path=/usr/share/grafana grafana | logger=settings t=2024-04-24T09:37:37.290435636Z level=info msg="Path Data" path=/var/lib/grafana grafana | logger=settings t=2024-04-24T09:37:37.290444277Z level=info msg="Path Logs" path=/var/log/grafana grafana | logger=settings t=2024-04-24T09:37:37.290453277Z level=info msg="Path Plugins" path=/var/lib/grafana/plugins grafana | logger=settings t=2024-04-24T09:37:37.290461067Z level=info msg="Path Provisioning" path=/etc/grafana/provisioning grafana | logger=settings t=2024-04-24T09:37:37.29052781Z level=info msg="App mode production" grafana | logger=sqlstore t=2024-04-24T09:37:37.291657834Z level=info msg="Connecting to DB" dbtype=sqlite3 grafana | logger=sqlstore t=2024-04-24T09:37:37.292191605Z level=info msg="Creating SQLite database file" path=/var/lib/grafana/grafana.db grafana | logger=migrator t=2024-04-24T09:37:37.293491145Z level=info msg="Starting DB migrations" grafana | logger=migrator t=2024-04-24T09:37:37.294864289Z level=info msg="Executing migration" id="create migration_log table" grafana | logger=migrator t=2024-04-24T09:37:37.295729243Z level=info msg="Migration successfully executed" id="create migration_log table" duration=864.724µs grafana | logger=migrator t=2024-04-24T09:37:37.301344372Z level=info msg="Executing migration" id="create user table" grafana | logger=migrator t=2024-04-24T09:37:37.301943705Z level=info msg="Migration successfully executed" id="create user table" duration=598.644µs grafana | logger=migrator t=2024-04-24T09:37:37.307902848Z level=info msg="Executing migration" id="add unique index user.login" grafana | logger=migrator t=2024-04-24T09:37:37.308497601Z level=info msg="Migration successfully executed" id="add unique index user.login" duration=592.363µs grafana | logger=migrator t=2024-04-24T09:37:37.312774237Z level=info msg="Executing migration" id="add unique index user.email" grafana | logger=migrator t=2024-04-24T09:37:37.314021396Z level=info msg="Migration successfully executed" id="add unique index user.email" duration=1.243889ms grafana | logger=migrator t=2024-04-24T09:37:37.31798595Z level=info msg="Executing migration" id="drop index UQE_user_login - v1" grafana | logger=migrator t=2024-04-24T09:37:37.318666237Z level=info msg="Migration successfully executed" id="drop index UQE_user_login - v1" duration=679.706µs grafana | logger=migrator t=2024-04-24T09:37:37.324028266Z level=info msg="Executing migration" id="drop index UQE_user_email - v1" grafana | logger=migrator t=2024-04-24T09:37:37.324786436Z level=info msg="Migration successfully executed" id="drop index UQE_user_email - v1" duration=752.85µs grafana | logger=migrator t=2024-04-24T09:37:37.328118886Z level=info msg="Executing migration" id="Rename table user to user_v1 - v1" grafana | logger=migrator t=2024-04-24T09:37:37.330252949Z level=info msg="Migration successfully executed" id="Rename table user to user_v1 - v1" duration=2.131823ms grafana | logger=migrator t=2024-04-24T09:37:37.334032986Z level=info msg="Executing migration" id="create user table v2" grafana | logger=migrator t=2024-04-24T09:37:37.334835477Z level=info msg="Migration successfully executed" id="create user table v2" duration=802.581µs grafana | logger=migrator t=2024-04-24T09:37:37.339500979Z level=info msg="Executing migration" id="create index UQE_user_login - v2" grafana | logger=migrator t=2024-04-24T09:37:37.340171825Z level=info msg="Migration successfully executed" id="create index UQE_user_login - v2" duration=670.716µs grafana | logger=migrator t=2024-04-24T09:37:37.345112928Z level=info msg="Executing migration" id="create index UQE_user_email - v2" grafana | logger=migrator t=2024-04-24T09:37:37.345747753Z level=info msg="Migration successfully executed" id="create index UQE_user_email - v2" duration=637.224µs grafana | logger=migrator t=2024-04-24T09:37:37.349220498Z level=info msg="Executing migration" id="copy data_source v1 to v2" grafana | logger=migrator t=2024-04-24T09:37:37.349650465Z level=info msg="Migration successfully executed" id="copy data_source v1 to v2" duration=427.257µs grafana | logger=migrator t=2024-04-24T09:37:37.35284901Z level=info msg="Executing migration" id="Drop old table user_v1" grafana | logger=migrator t=2024-04-24T09:37:37.353341049Z level=info msg="Migration successfully executed" id="Drop old table user_v1" duration=491.309µs grafana | logger=migrator t=2024-04-24T09:37:37.358221379Z level=info msg="Executing migration" id="Add column help_flags1 to user table" grafana | logger=migrator t=2024-04-24T09:37:37.359235038Z level=info msg="Migration successfully executed" id="Add column help_flags1 to user table" duration=1.010819ms grafana | logger=migrator t=2024-04-24T09:37:37.363194962Z level=info msg="Executing migration" id="Update user table charset" grafana | logger=migrator t=2024-04-24T09:37:37.363225664Z level=info msg="Migration successfully executed" id="Update user table charset" duration=31.412µs grafana | logger=migrator t=2024-04-24T09:37:37.366414909Z level=info msg="Executing migration" id="Add last_seen_at column to user" grafana | logger=migrator t=2024-04-24T09:37:37.367560643Z level=info msg="Migration successfully executed" id="Add last_seen_at column to user" duration=1.145314ms grafana | logger=migrator t=2024-04-24T09:37:37.371190764Z level=info msg="Executing migration" id="Add missing user data" grafana | logger=migrator t=2024-04-24T09:37:37.371500036Z level=info msg="Migration successfully executed" id="Add missing user data" duration=308.842µs grafana | logger=migrator t=2024-04-24T09:37:37.37672015Z level=info msg="Executing migration" id="Add is_disabled column to user" mariadb | 2024-04-24 9:37:45 16 [Warning] Aborted connection 16 to db: 'unconnected' user: 'unauthenticated' host: '172.17.0.11' (This connection closed normally without authentication) mariadb | 2024-04-24 9:37:45 18 [Warning] Aborted connection 18 to db: 'unconnected' user: 'unauthenticated' host: '172.17.0.10' (This connection closed normally without authentication) kafka | ===> User kafka | uid=1000(appuser) gid=1000(appuser) groups=1000(appuser) kafka | ===> Configuring ... kafka | Running in Zookeeper mode... kafka | ===> Running preflight checks ... kafka | ===> Check if /var/lib/kafka/data is writable ... kafka | ===> Check if Zookeeper is healthy ... kafka | [2024-04-24 09:37:47,799] INFO Client environment:zookeeper.version=3.8.3-6ad6d364c7c0bcf0de452d54ebefa3058098ab56, built on 2023-10-05 10:34 UTC (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,799] INFO Client environment:host.name=de58d82459ba (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,799] INFO Client environment:java.version=11.0.22 (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,799] INFO Client environment:java.vendor=Azul Systems, Inc. (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,799] INFO Client environment:java.home=/usr/lib/jvm/java-11-zulu-openjdk-ca (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,799] INFO Client environment:java.class.path=/usr/share/java/cp-base-new/scala-logging_2.13-3.9.4.jar:/usr/share/java/cp-base-new/jackson-datatype-jdk8-2.14.2.jar:/usr/share/java/cp-base-new/kafka-storage-api-7.6.1-ccs.jar:/usr/share/java/cp-base-new/commons-validator-1.7.jar:/usr/share/java/cp-base-new/logredactor-1.0.12.jar:/usr/share/java/cp-base-new/jolokia-core-1.7.1.jar:/usr/share/java/cp-base-new/re2j-1.6.jar:/usr/share/java/cp-base-new/scala-library-2.13.10.jar:/usr/share/java/cp-base-new/commons-io-2.11.0.jar:/usr/share/java/cp-base-new/kafka-clients-7.6.1-ccs.jar:/usr/share/java/cp-base-new/utility-belt-7.6.1.jar:/usr/share/java/cp-base-new/commons-cli-1.4.jar:/usr/share/java/cp-base-new/slf4j-reload4j-1.7.36.jar:/usr/share/java/cp-base-new/kafka-server-common-7.6.1-ccs.jar:/usr/share/java/cp-base-new/kafka-group-coordinator-7.6.1-ccs.jar:/usr/share/java/cp-base-new/common-utils-7.6.1.jar:/usr/share/java/cp-base-new/jackson-annotations-2.14.2.jar:/usr/share/java/cp-base-new/json-simple-1.1.1.jar:/usr/share/java/cp-base-new/jackson-module-scala_2.13-2.14.2.jar:/usr/share/java/cp-base-new/scala-java8-compat_2.13-1.0.2.jar:/usr/share/java/cp-base-new/minimal-json-0.9.5.jar:/usr/share/java/cp-base-new/kafka-storage-7.6.1-ccs.jar:/usr/share/java/cp-base-new/zstd-jni-1.5.5-1.jar:/usr/share/java/cp-base-new/jackson-dataformat-yaml-2.14.2.jar:/usr/share/java/cp-base-new/slf4j-api-1.7.36.jar:/usr/share/java/cp-base-new/paranamer-2.8.jar:/usr/share/java/cp-base-new/commons-beanutils-1.9.4.jar:/usr/share/java/cp-base-new/jmx_prometheus_javaagent-0.18.0.jar:/usr/share/java/cp-base-new/reload4j-1.2.25.jar:/usr/share/java/cp-base-new/jackson-core-2.14.2.jar:/usr/share/java/cp-base-new/disk-usage-agent-7.6.1.jar:/usr/share/java/cp-base-new/commons-digester-2.1.jar:/usr/share/java/cp-base-new/argparse4j-0.7.0.jar:/usr/share/java/cp-base-new/audience-annotations-0.12.0.jar:/usr/share/java/cp-base-new/kafka-raft-7.6.1-ccs.jar:/usr/share/java/cp-base-new/gson-2.9.0.jar:/usr/share/java/cp-base-new/snakeyaml-2.0.jar:/usr/share/java/cp-base-new/jopt-simple-5.0.4.jar:/usr/share/java/cp-base-new/error_prone_annotations-2.10.0.jar:/usr/share/java/cp-base-new/lz4-java-1.8.0.jar:/usr/share/java/cp-base-new/logredactor-metrics-1.0.12.jar:/usr/share/java/cp-base-new/checker-qual-3.19.0.jar:/usr/share/java/cp-base-new/kafka-metadata-7.6.1-ccs.jar:/usr/share/java/cp-base-new/pcollections-4.0.1.jar:/usr/share/java/cp-base-new/commons-logging-1.2.jar:/usr/share/java/cp-base-new/scala-reflect-2.13.10.jar:/usr/share/java/cp-base-new/scala-collection-compat_2.13-2.10.0.jar:/usr/share/java/cp-base-new/metrics-core-2.2.0.jar:/usr/share/java/cp-base-new/zookeeper-jute-3.8.3.jar:/usr/share/java/cp-base-new/jackson-dataformat-csv-2.14.2.jar:/usr/share/java/cp-base-new/commons-collections-3.2.2.jar:/usr/share/java/cp-base-new/caffeine-2.9.3.jar:/usr/share/java/cp-base-new/jolokia-jvm-1.7.1.jar:/usr/share/java/cp-base-new/kafka-tools-api-7.6.1-ccs.jar:/usr/share/java/cp-base-new/zookeeper-3.8.3.jar:/usr/share/java/cp-base-new/metrics-core-4.1.12.1.jar:/usr/share/java/cp-base-new/kafka_2.13-7.6.1-ccs.jar:/usr/share/java/cp-base-new/jackson-databind-2.14.2.jar:/usr/share/java/cp-base-new/snappy-java-1.1.10.5.jar:/usr/share/java/cp-base-new/jose4j-0.9.5.jar (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:os.version=4.15.0-192-generic (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:user.name=appuser (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:user.home=/home/appuser (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:user.dir=/home/appuser (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:os.memory.free=493MB (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:os.memory.max=8042MB (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,800] INFO Client environment:os.memory.total=504MB (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,803] INFO Initiating client connection, connectString=zookeeper:2181 sessionTimeout=40000 watcher=io.confluent.admin.utils.ZookeeperConnectionWatcher@b7f23d9 (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:47,807] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) kafka | [2024-04-24 09:37:47,813] INFO jute.maxbuffer value is 1048575 Bytes (org.apache.zookeeper.ClientCnxnSocket) kafka | [2024-04-24 09:37:47,823] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) kafka | [2024-04-24 09:37:47,847] INFO Opening socket connection to server zookeeper/172.17.0.5:2181. (org.apache.zookeeper.ClientCnxn) kafka | [2024-04-24 09:37:47,847] INFO SASL config status: Will not attempt to authenticate using SASL (unknown error) (org.apache.zookeeper.ClientCnxn) kafka | [2024-04-24 09:37:47,856] INFO Socket connection established, initiating session, client: /172.17.0.9:33838, server: zookeeper/172.17.0.5:2181 (org.apache.zookeeper.ClientCnxn) kafka | [2024-04-24 09:37:47,892] INFO Session establishment complete on server zookeeper/172.17.0.5:2181, session id = 0x1000005537a0000, negotiated timeout = 40000 (org.apache.zookeeper.ClientCnxn) kafka | [2024-04-24 09:37:48,014] INFO Session: 0x1000005537a0000 closed (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:48,014] INFO EventThread shut down for session: 0x1000005537a0000 (org.apache.zookeeper.ClientCnxn) kafka | Using log4j config /etc/kafka/log4j.properties kafka | ===> Launching ... kafka | ===> Launching kafka ... kafka | [2024-04-24 09:37:48,827] INFO Registered kafka:type=kafka.Log4jController MBean (kafka.utils.Log4jControllerRegistration$) kafka | [2024-04-24 09:37:49,196] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) kafka | [2024-04-24 09:37:49,265] INFO Registered signal handlers for TERM, INT, HUP (org.apache.kafka.common.utils.LoggingSignalHandler) kafka | [2024-04-24 09:37:49,266] INFO starting (kafka.server.KafkaServer) kafka | [2024-04-24 09:37:49,266] INFO Connecting to zookeeper on zookeeper:2181 (kafka.server.KafkaServer) kafka | [2024-04-24 09:37:49,280] INFO [ZooKeeperClient Kafka server] Initializing a new session to zookeeper:2181. (kafka.zookeeper.ZooKeeperClient) kafka | [2024-04-24 09:37:49,284] INFO Client environment:zookeeper.version=3.8.4-9316c2a7a97e1666d8f4593f34dd6fc36ecc436c, built on 2024-02-12 22:16 UTC (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:host.name=de58d82459ba (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:java.version=11.0.22 (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:java.vendor=Azul Systems, Inc. (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:java.home=/usr/lib/jvm/java-11-zulu-openjdk-ca (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:java.class.path=/usr/bin/../share/java/kafka/scala-logging_2.13-3.9.4.jar:/usr/bin/../share/java/kafka/jersey-common-2.39.1.jar:/usr/bin/../share/java/kafka/swagger-annotations-2.2.8.jar:/usr/bin/../share/java/kafka/connect-json-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-storage-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/commons-validator-1.7.jar:/usr/bin/../share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/bin/../share/java/kafka/aopalliance-repackaged-2.6.1.jar:/usr/bin/../share/java/kafka/netty-transport-4.1.100.Final.jar:/usr/bin/../share/java/kafka/rocksdbjni-7.9.2.jar:/usr/bin/../share/java/kafka/jetty-http-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka-log4j-appender-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/commons-io-2.11.0.jar:/usr/bin/../share/java/kafka/kafka-clients-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/javax.activation-api-1.2.0.jar:/usr/bin/../share/java/kafka/jetty-security-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/commons-cli-1.4.jar:/usr/bin/../share/java/kafka/jetty-server-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/slf4j-reload4j-1.7.36.jar:/usr/bin/../share/java/kafka/scala-reflect-2.13.11.jar:/usr/bin/../share/java/kafka/kafka-server-common-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-group-coordinator-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-buffer-4.1.100.Final.jar:/usr/bin/../share/java/kafka/connect-runtime-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/trogdor-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.ws.rs-api-2.1.6.jar:/usr/bin/../share/java/kafka/jakarta.annotation-api-1.3.5.jar:/usr/bin/../share/java/kafka/kafka-streams-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/scala-java8-compat_2.13-1.0.2.jar:/usr/bin/../share/java/kafka/javax.ws.rs-api-2.1.1.jar:/usr/bin/../share/java/kafka/zookeeper-jute-3.8.4.jar:/usr/bin/../share/java/kafka/netty-resolver-4.1.100.Final.jar:/usr/bin/../share/java/kafka/hk2-api-2.6.1.jar:/usr/bin/../share/java/kafka/kafka-storage-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/zstd-jni-1.5.5-1.jar:/usr/bin/../share/java/kafka/jackson-dataformat-csv-2.13.5.jar:/usr/bin/../share/java/kafka/kafka.jar:/usr/bin/../share/java/kafka/scala-library-2.13.11.jar:/usr/bin/../share/java/kafka/jakarta.inject-2.6.1.jar:/usr/bin/../share/java/kafka/jakarta.xml.bind-api-2.3.3.jar:/usr/bin/../share/java/kafka/jetty-continuation-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/connect-transforms-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jose4j-0.9.4.jar:/usr/bin/../share/java/kafka/hk2-locator-2.6.1.jar:/usr/bin/../share/java/kafka/reflections-0.10.2.jar:/usr/bin/../share/java/kafka/slf4j-api-1.7.36.jar:/usr/bin/../share/java/kafka/paranamer-2.8.jar:/usr/bin/../share/java/kafka/commons-beanutils-1.9.4.jar:/usr/bin/../share/java/kafka/jaxb-api-2.3.1.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-2.39.1.jar:/usr/bin/../share/java/kafka/hk2-utils-2.6.1.jar:/usr/bin/../share/java/kafka/jackson-module-scala_2.13-2.13.5.jar:/usr/bin/../share/java/kafka/reload4j-1.2.25.jar:/usr/bin/../share/java/kafka/connect-mirror-client-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-servlet-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jackson-core-2.13.5.jar:/usr/bin/../share/java/kafka/jersey-hk2-2.39.1.jar:/usr/bin/../share/java/kafka/jackson-databind-2.13.5.jar:/usr/bin/../share/java/kafka/jetty-servlets-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jersey-client-2.39.1.jar:/usr/bin/../share/java/kafka/osgi-resource-locator-1.0.3.jar:/usr/bin/../share/java/kafka/commons-digester-2.1.jar:/usr/bin/../share/java/kafka/netty-transport-native-epoll-4.1.100.Final.jar:/usr/bin/../share/java/kafka/argparse4j-0.7.0.jar:/usr/bin/../share/java/kafka/connect-mirror-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-datatype-jdk8-2.13.5.jar:/usr/bin/../share/java/kafka/audience-annotations-0.12.0.jar:/usr/bin/../share/java/kafka/jackson-module-jaxb-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/kafka-raft-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/javax.annotation-api-1.3.2.jar:/usr/bin/../share/java/kafka/maven-artifact-3.8.8.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-json-provider-2.13.5.jar:/usr/bin/../share/java/kafka/jakarta.validation-api-2.0.2.jar:/usr/bin/../share/java/kafka/zookeeper-3.8.4.jar:/usr/bin/../share/java/kafka/jersey-server-2.39.1.jar:/usr/bin/../share/java/kafka/commons-lang3-3.8.1.jar:/usr/bin/../share/java/kafka/jopt-simple-5.0.4.jar:/usr/bin/../share/java/kafka/error_prone_annotations-2.10.0.jar:/usr/bin/../share/java/kafka/lz4-java-1.8.0.jar:/usr/bin/../share/java/kafka/jetty-util-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka-tools-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.activation-api-1.2.2.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-core-2.39.1.jar:/usr/bin/../share/java/kafka/checker-qual-3.19.0.jar:/usr/bin/../share/java/kafka/kafka-metadata-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/pcollections-4.0.1.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-base-2.13.5.jar:/usr/bin/../share/java/kafka/commons-logging-1.2.jar:/usr/bin/../share/java/kafka/jsr305-3.0.2.jar:/usr/bin/../share/java/kafka/jetty-io-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/netty-codec-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-transport-native-unix-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/scala-collection-compat_2.13-2.10.0.jar:/usr/bin/../share/java/kafka/metrics-core-2.2.0.jar:/usr/bin/../share/java/kafka/kafka-streams-test-utils-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-streams-examples-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-handler-4.1.100.Final.jar:/usr/bin/../share/java/kafka/commons-collections-3.2.2.jar:/usr/bin/../share/java/kafka/javassist-3.29.2-GA.jar:/usr/bin/../share/java/kafka/caffeine-2.9.3.jar:/usr/bin/../share/java/kafka/plexus-utils-3.3.1.jar:/usr/bin/../share/java/kafka/kafka-tools-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/activation-1.1.1.jar:/usr/bin/../share/java/kafka/netty-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/kafka-streams-scala_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/metrics-core-4.1.12.1.jar:/usr/bin/../share/java/kafka/jline-3.25.1.jar:/usr/bin/../share/java/kafka/jetty-client-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-basic-auth-extension-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-util-ajax-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/snappy-java-1.1.10.5.jar:/usr/bin/../share/java/kafka/kafka-shell-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-transport-classes-epoll-4.1.100.Final.jar:/usr/bin/../share/java/confluent-telemetry/* (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) policy-db-migrator | Waiting for mariadb port 3306... policy-db-migrator | nc: connect to mariadb (172.17.0.4) port 3306 (tcp) failed: Connection refused policy-db-migrator | nc: connect to mariadb (172.17.0.4) port 3306 (tcp) failed: Connection refused policy-db-migrator | nc: connect to mariadb (172.17.0.4) port 3306 (tcp) failed: Connection refused policy-db-migrator | nc: connect to mariadb (172.17.0.4) port 3306 (tcp) failed: Connection refused policy-db-migrator | nc: connect to mariadb (172.17.0.4) port 3306 (tcp) failed: Connection refused policy-db-migrator | nc: connect to mariadb (172.17.0.4) port 3306 (tcp) failed: Connection refused policy-db-migrator | Connection to mariadb (172.17.0.4) 3306 port [tcp/mysql] succeeded! policy-db-migrator | 321 blocks policy-db-migrator | Preparing upgrade release version: 0800 policy-db-migrator | Preparing upgrade release version: 0900 policy-db-migrator | Preparing upgrade release version: 1000 policy-db-migrator | Preparing upgrade release version: 1100 policy-db-migrator | Preparing upgrade release version: 1200 policy-db-migrator | Preparing upgrade release version: 1300 policy-db-migrator | Done policy-db-migrator | name version policy-db-migrator | policyadmin 0 policy-db-migrator | policyadmin: upgrade available: 0 -> 1300 policy-db-migrator | upgrade: 0 -> 1300 policy-db-migrator | policy-db-migrator | > upgrade 0100-jpapdpgroup_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpgroup_properties (name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, PROPERTIES VARCHAR(255) DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) DEFAULT NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0110-jpapdpstatistics_enginestats.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpstatistics_enginestats (AVERAGEEXECUTIONTIME DOUBLE DEFAULT NULL, ENGINEID VARCHAR(255) DEFAULT NULL, ENGINETIMESTAMP BIGINT DEFAULT NULL, ENGINEWORKERSTATE INT DEFAULT NULL, EVENTCOUNT BIGINT DEFAULT NULL, LASTENTERTIME BIGINT DEFAULT NULL, LASTEXECUTIONTIME BIGINT DEFAULT NULL, LASTSTART BIGINT DEFAULT NULL, UPTIME BIGINT DEFAULT NULL, timeStamp datetime DEFAULT NULL, name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0120-jpapdpsubgroup_policies.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpsubgroup_policies (name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, parentLocalName VARCHAR(120) DEFAULT NULL, localName VARCHAR(120) DEFAULT NULL, parentKeyVersion VARCHAR(15) DEFAULT NULL, parentKeyName VARCHAR(120) DEFAULT NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0130-jpapdpsubgroup_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpsubgroup_properties (parentLocalName VARCHAR(120) DEFAULT NULL, localName VARCHAR(120) DEFAULT NULL, parentKeyVersion VARCHAR(15) DEFAULT NULL, parentKeyName VARCHAR(120) DEFAULT NULL, PROPERTIES VARCHAR(255) DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) DEFAULT NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0140-jpapdpsubgroup_supportedpolicytypes.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapdpsubgroup_supportedpolicytypes (name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, parentLocalName VARCHAR(120) DEFAULT NULL, localName VARCHAR(120) DEFAULT NULL, parentKeyVersion VARCHAR(15) DEFAULT NULL, parentKeyName VARCHAR(120) DEFAULT NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0150-jpatoscacapabilityassignment_attributes.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilityassignment_attributes (name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, ATTRIBUTES LONGTEXT DEFAULT NULL, ATTRIBUTES_KEY VARCHAR(255) DEFAULT NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0160-jpatoscacapabilityassignment_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilityassignment_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | grafana | logger=migrator t=2024-04-24T09:37:37.377763051Z level=info msg="Migration successfully executed" id="Add is_disabled column to user" duration=1.043101ms grafana | logger=migrator t=2024-04-24T09:37:37.381171654Z level=info msg="Executing migration" id="Add index user.login/user.email" grafana | logger=migrator t=2024-04-24T09:37:37.381895402Z level=info msg="Migration successfully executed" id="Add index user.login/user.email" duration=721.357µs grafana | logger=migrator t=2024-04-24T09:37:37.385317285Z level=info msg="Executing migration" id="Add is_service_account column to user" grafana | logger=migrator t=2024-04-24T09:37:37.386375067Z level=info msg="Migration successfully executed" id="Add is_service_account column to user" duration=1.056862ms grafana | logger=migrator t=2024-04-24T09:37:37.389580531Z level=info msg="Executing migration" id="Update is_service_account column to nullable" grafana | logger=migrator t=2024-04-24T09:37:37.396727349Z level=info msg="Migration successfully executed" id="Update is_service_account column to nullable" duration=7.146128ms grafana | logger=migrator t=2024-04-24T09:37:37.401474955Z level=info msg="Executing migration" id="Add uid column to user" grafana | logger=migrator t=2024-04-24T09:37:37.402566347Z level=info msg="Migration successfully executed" id="Add uid column to user" duration=1.089962ms grafana | logger=migrator t=2024-04-24T09:37:37.406138067Z level=info msg="Executing migration" id="Update uid column values for users" grafana | logger=migrator t=2024-04-24T09:37:37.406403217Z level=info msg="Migration successfully executed" id="Update uid column values for users" duration=267.22µs grafana | logger=migrator t=2024-04-24T09:37:37.409465416Z level=info msg="Executing migration" id="Add unique index user_uid" grafana | logger=migrator t=2024-04-24T09:37:37.410153173Z level=info msg="Migration successfully executed" id="Add unique index user_uid" duration=688.027µs grafana | logger=migrator t=2024-04-24T09:37:37.41468545Z level=info msg="Executing migration" id="update login field with orgid to allow for multiple service accounts with same name across orgs" grafana | logger=migrator t=2024-04-24T09:37:37.415075895Z level=info msg="Migration successfully executed" id="update login field with orgid to allow for multiple service accounts with same name across orgs" duration=391.615µs grafana | logger=migrator t=2024-04-24T09:37:37.418134564Z level=info msg="Executing migration" id="create temp user table v1-7" grafana | logger=migrator t=2024-04-24T09:37:37.418932495Z level=info msg="Migration successfully executed" id="create temp user table v1-7" duration=797.591µs grafana | logger=migrator t=2024-04-24T09:37:37.422718113Z level=info msg="Executing migration" id="create index IDX_temp_user_email - v1-7" grafana | logger=migrator t=2024-04-24T09:37:37.423489173Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_email - v1-7" duration=770.62µs grafana | logger=migrator t=2024-04-24T09:37:37.428194617Z level=info msg="Executing migration" id="create index IDX_temp_user_org_id - v1-7" grafana | logger=migrator t=2024-04-24T09:37:37.428942716Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_org_id - v1-7" duration=748.269µs grafana | logger=migrator t=2024-04-24T09:37:37.432229444Z level=info msg="Executing migration" id="create index IDX_temp_user_code - v1-7" grafana | logger=migrator t=2024-04-24T09:37:37.432986453Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_code - v1-7" duration=756.699µs grafana | logger=migrator t=2024-04-24T09:37:37.436071183Z level=info msg="Executing migration" id="create index IDX_temp_user_status - v1-7" grafana | logger=migrator t=2024-04-24T09:37:37.436818683Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_status - v1-7" duration=748.06µs grafana | logger=migrator t=2024-04-24T09:37:37.439864482Z level=info msg="Executing migration" id="Update temp_user table charset" grafana | logger=migrator t=2024-04-24T09:37:37.439884592Z level=info msg="Migration successfully executed" id="Update temp_user table charset" duration=22.771µs grafana | logger=migrator t=2024-04-24T09:37:37.444423289Z level=info msg="Executing migration" id="drop index IDX_temp_user_email - v1" grafana | logger=migrator t=2024-04-24T09:37:37.444995442Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_email - v1" duration=572.593µs grafana | logger=migrator t=2024-04-24T09:37:37.447681016Z level=info msg="Executing migration" id="drop index IDX_temp_user_org_id - v1" grafana | logger=migrator t=2024-04-24T09:37:37.448257149Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_org_id - v1" duration=576.113µs grafana | logger=migrator t=2024-04-24T09:37:37.451961963Z level=info msg="Executing migration" id="drop index IDX_temp_user_code - v1" grafana | logger=migrator t=2024-04-24T09:37:37.453164299Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_code - v1" duration=1.204406ms grafana | logger=migrator t=2024-04-24T09:37:37.458524298Z level=info msg="Executing migration" id="drop index IDX_temp_user_status - v1" grafana | logger=migrator t=2024-04-24T09:37:37.459261238Z level=info msg="Migration successfully executed" id="drop index IDX_temp_user_status - v1" duration=737.05µs grafana | logger=migrator t=2024-04-24T09:37:37.462488373Z level=info msg="Executing migration" id="Rename table temp_user to temp_user_tmp_qwerty - v1" grafana | logger=migrator t=2024-04-24T09:37:37.465838114Z level=info msg="Migration successfully executed" id="Rename table temp_user to temp_user_tmp_qwerty - v1" duration=3.348211ms grafana | logger=migrator t=2024-04-24T09:37:37.469030608Z level=info msg="Executing migration" id="create temp_user v2" grafana | logger=migrator t=2024-04-24T09:37:37.470009876Z level=info msg="Migration successfully executed" id="create temp_user v2" duration=978.758µs grafana | logger=migrator t=2024-04-24T09:37:37.475330973Z level=info msg="Executing migration" id="create index IDX_temp_user_email - v2" grafana | logger=migrator t=2024-04-24T09:37:37.476150016Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_email - v2" duration=818.923µs grafana | logger=migrator t=2024-04-24T09:37:37.480097229Z level=info msg="Executing migration" id="create index IDX_temp_user_org_id - v2" grafana | logger=migrator t=2024-04-24T09:37:37.480986734Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_org_id - v2" duration=889.225µs grafana | logger=migrator t=2024-04-24T09:37:37.484526282Z level=info msg="Executing migration" id="create index IDX_temp_user_code - v2" grafana | logger=migrator t=2024-04-24T09:37:37.485384116Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_code - v2" duration=857.963µs grafana | logger=migrator t=2024-04-24T09:37:37.491744893Z level=info msg="Executing migration" id="create index IDX_temp_user_status - v2" grafana | logger=migrator t=2024-04-24T09:37:37.493146258Z level=info msg="Migration successfully executed" id="create index IDX_temp_user_status - v2" duration=1.402755ms grafana | logger=migrator t=2024-04-24T09:37:37.497778648Z level=info msg="Executing migration" id="copy temp_user v1 to v2" grafana | logger=migrator t=2024-04-24T09:37:37.498599691Z level=info msg="Migration successfully executed" id="copy temp_user v1 to v2" duration=819.903µs grafana | logger=migrator t=2024-04-24T09:37:37.502403489Z level=info msg="Executing migration" id="drop temp_user_tmp_qwerty" policy-db-migrator | > upgrade 0170-jpatoscacapabilityassignment_occurrences.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilityassignment_occurrences (name VARCHAR(120) NULL, version VARCHAR(20) NULL, OCCURRENCES INT DEFAULT NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0180-jpatoscacapabilityassignment_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilityassignment_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGTEXT NULL, PROPERTIES_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0190-jpatoscacapabilitytype_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilitytype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0200-jpatoscacapabilitytype_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscacapabilitytype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0210-jpatoscadatatype_constraints.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscadatatype_constraints (name VARCHAR(120) NULL, version VARCHAR(20) NULL, CONSTRAINTS VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0220-jpatoscadatatype_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscadatatype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0230-jpatoscadatatype_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscadatatype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0240-jpatoscanodetemplate_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscanodetemplate_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0250-jpatoscanodetemplate_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscanodetemplate_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGTEXT NULL, PROPERTIES_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0260-jpatoscanodetype_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscanodetype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0270-jpatoscanodetype_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscanodetype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0280-jpatoscapolicy_metadata.sql kafka | [2024-04-24 09:37:49,284] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:os.version=4.15.0-192-generic (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:user.name=appuser (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:user.home=/home/appuser (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:user.dir=/home/appuser (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:os.memory.free=1008MB (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:os.memory.max=1024MB (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,284] INFO Client environment:os.memory.total=1024MB (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,286] INFO Initiating client connection, connectString=zookeeper:2181 sessionTimeout=18000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@66746f57 (org.apache.zookeeper.ZooKeeper) kafka | [2024-04-24 09:37:49,290] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket) kafka | [2024-04-24 09:37:49,297] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) kafka | [2024-04-24 09:37:49,299] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient) kafka | [2024-04-24 09:37:49,303] INFO Opening socket connection to server zookeeper/172.17.0.5:2181. (org.apache.zookeeper.ClientCnxn) kafka | [2024-04-24 09:37:49,310] INFO Socket connection established, initiating session, client: /172.17.0.9:33840, server: zookeeper/172.17.0.5:2181 (org.apache.zookeeper.ClientCnxn) kafka | [2024-04-24 09:37:49,323] INFO Session establishment complete on server zookeeper/172.17.0.5:2181, session id = 0x1000005537a0001, negotiated timeout = 18000 (org.apache.zookeeper.ClientCnxn) kafka | [2024-04-24 09:37:49,331] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient) kafka | [2024-04-24 09:37:49,685] INFO Cluster ID = 8oQQmVC5RD2DltlWw-KT5A (kafka.server.KafkaServer) kafka | [2024-04-24 09:37:49,690] WARN No meta.properties file under dir /var/lib/kafka/data/meta.properties (kafka.server.BrokerMetadataCheckpoint) kafka | [2024-04-24 09:37:49,752] INFO KafkaConfig values: kafka | advertised.listeners = PLAINTEXT://kafka:9092,PLAINTEXT_HOST://localhost:29092 kafka | alter.config.policy.class.name = null kafka | alter.log.dirs.replication.quota.window.num = 11 kafka | alter.log.dirs.replication.quota.window.size.seconds = 1 kafka | authorizer.class.name = kafka | auto.create.topics.enable = true kafka | auto.include.jmx.reporter = true kafka | auto.leader.rebalance.enable = true kafka | background.threads = 10 kafka | broker.heartbeat.interval.ms = 2000 kafka | broker.id = 1 kafka | broker.id.generation.enable = true kafka | broker.rack = null kafka | broker.session.timeout.ms = 9000 kafka | client.quota.callback.class = null kafka | compression.type = producer kafka | connection.failed.authentication.delay.ms = 100 kafka | connections.max.idle.ms = 600000 kafka | connections.max.reauth.ms = 0 kafka | control.plane.listener.name = null kafka | controlled.shutdown.enable = true kafka | controlled.shutdown.max.retries = 3 kafka | controlled.shutdown.retry.backoff.ms = 5000 policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicy_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0290-jpatoscapolicy_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicy_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGTEXT NULL, PROPERTIES_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0300-jpatoscapolicy_targets.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicy_targets (name VARCHAR(120) NULL, version VARCHAR(20) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0310-jpatoscapolicytype_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicytype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0320-jpatoscapolicytype_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicytype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0330-jpatoscapolicytype_targets.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicytype_targets (name VARCHAR(120) NULL, version VARCHAR(20) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0340-jpatoscapolicytype_triggers.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscapolicytype_triggers (name VARCHAR(120) NULL, version VARCHAR(20) NULL, TRIGGERS VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0350-jpatoscaproperty_constraints.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscaproperty_constraints (parentLocalName VARCHAR(120) NULL, localName VARCHAR(120) NULL, parentKeyVersion VARCHAR(15) NULL, parentKeyName VARCHAR(120) NULL, CONSTRAINTS VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0360-jpatoscaproperty_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscaproperty_metadata (parentLocalName VARCHAR(120) NULL, localName VARCHAR(120) NULL, parentKeyVersion VARCHAR(15) NULL, parentKeyName VARCHAR(120) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0370-jpatoscarelationshiptype_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarelationshiptype_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0380-jpatoscarelationshiptype_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarelationshiptype_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGBLOB DEFAULT NULL, PROPERTIES_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0390-jpatoscarequirement_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarequirement_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0400-jpatoscarequirement_occurrences.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarequirement_occurrences (name VARCHAR(120) NULL, version VARCHAR(20) NULL, OCCURRENCES INT DEFAULT NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0410-jpatoscarequirement_properties.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscarequirement_properties (name VARCHAR(120) NULL, version VARCHAR(20) NULL, PROPERTIES LONGTEXT NULL, PROPERTIES_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0420-jpatoscaservicetemplate_metadata.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscaservicetemplate_metadata (name VARCHAR(120) NULL, version VARCHAR(20) NULL, METADATA VARCHAR(255) NULL, METADATA_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0430-jpatoscatopologytemplate_inputs.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpatoscatopologytemplate_inputs (parentLocalName VARCHAR(120) NULL, localName VARCHAR(120) NULL, parentKeyVersion VARCHAR(15) NULL, parentKeyName VARCHAR(120) NULL, INPUTS LONGBLOB DEFAULT NULL, INPUTS_KEY VARCHAR(255) NULL) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0440-pdpgroup_pdpsubgroup.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpgroup_pdpsubgroup (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_PDPGROUP_PDPSUBGROUP (name, version, parentLocalName, localName, parentKeyVersion, parentKeyName)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0450-pdpgroup.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpgroup (`DESCRIPTION` VARCHAR(255) NULL, PDPGROUPSTATE INT DEFAULT NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_PDPGROUP (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0460-pdppolicystatus.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS pdppolicystatus (DEPLOY BOOLEAN DEFAULT 0, PDPGROUP VARCHAR(255) DEFAULT NULL, PDPTYPE VARCHAR(255) DEFAULT NULL, STATE INT DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, PRIMARY KEY PK_PDPPOLICYSTATUS (parentLocalName, localName, parentKeyVersion, parentKeyName)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0470-pdp.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS pdp (HEALTHY INT DEFAULT NULL, MESSAGE VARCHAR(255) DEFAULT NULL, PDPSTATE INT DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_PDP (parentLocalName, localName, parentKeyVersion, parentKeyName)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0480-pdpstatistics.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpstatistics (PDPGROUPNAME VARCHAR(120) NULL, PDPSUBGROUPNAME VARCHAR(120) NULL, POLICYDEPLOYCOUNT BIGINT DEFAULT NULL, POLICYDEPLOYFAILCOUNT BIGINT DEFAULT NULL, POLICYDEPLOYSUCCESSCOUNT BIGINT DEFAULT NULL, POLICYEXECUTEDCOUNT BIGINT DEFAULT NULL, POLICYEXECUTEDFAILCOUNT BIGINT DEFAULT NULL, POLICYEXECUTEDSUCCESSCOUNT BIGINT DEFAULT NULL, timeStamp datetime NOT NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_PDPSTATISTICS (timeStamp, name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0490-pdpsubgroup_pdp.sql policy-db-migrator | -------------- kafka | controller.listener.names = null kafka | controller.quorum.append.linger.ms = 25 kafka | controller.quorum.election.backoff.max.ms = 1000 kafka | controller.quorum.election.timeout.ms = 1000 kafka | controller.quorum.fetch.timeout.ms = 2000 kafka | controller.quorum.request.timeout.ms = 2000 kafka | controller.quorum.retry.backoff.ms = 20 kafka | controller.quorum.voters = [] kafka | controller.quota.window.num = 11 kafka | controller.quota.window.size.seconds = 1 kafka | controller.socket.timeout.ms = 30000 kafka | create.topic.policy.class.name = null kafka | default.replication.factor = 1 kafka | delegation.token.expiry.check.interval.ms = 3600000 kafka | delegation.token.expiry.time.ms = 86400000 kafka | delegation.token.master.key = null kafka | delegation.token.max.lifetime.ms = 604800000 kafka | delegation.token.secret.key = null kafka | delete.records.purgatory.purge.interval.requests = 1 kafka | delete.topic.enable = true kafka | early.start.listeners = null kafka | fetch.max.bytes = 57671680 kafka | fetch.purgatory.purge.interval.requests = 1000 kafka | group.consumer.assignors = [org.apache.kafka.coordinator.group.assignor.RangeAssignor] kafka | group.consumer.heartbeat.interval.ms = 5000 kafka | group.consumer.max.heartbeat.interval.ms = 15000 kafka | group.consumer.max.session.timeout.ms = 60000 kafka | group.consumer.max.size = 2147483647 kafka | group.consumer.min.heartbeat.interval.ms = 5000 kafka | group.consumer.min.session.timeout.ms = 45000 kafka | group.consumer.session.timeout.ms = 45000 kafka | group.coordinator.new.enable = false kafka | group.coordinator.threads = 1 kafka | group.initial.rebalance.delay.ms = 3000 kafka | group.max.session.timeout.ms = 1800000 kafka | group.max.size = 2147483647 kafka | group.min.session.timeout.ms = 6000 kafka | initial.broker.registration.timeout.ms = 60000 kafka | inter.broker.listener.name = PLAINTEXT kafka | inter.broker.protocol.version = 3.6-IV2 kafka | kafka.metrics.polling.interval.secs = 10 kafka | kafka.metrics.reporters = [] kafka | leader.imbalance.check.interval.seconds = 300 kafka | leader.imbalance.per.broker.percentage = 10 kafka | listener.security.protocol.map = PLAINTEXT:PLAINTEXT,PLAINTEXT_HOST:PLAINTEXT kafka | listeners = PLAINTEXT://0.0.0.0:9092,PLAINTEXT_HOST://0.0.0.0:29092 kafka | log.cleaner.backoff.ms = 15000 kafka | log.cleaner.dedupe.buffer.size = 134217728 kafka | log.cleaner.delete.retention.ms = 86400000 policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpsubgroup_pdp (pdpParentKeyName VARCHAR(120) NOT NULL, pdpParentKeyVersion VARCHAR(15) NOT NULL, pdpParentLocalName VARCHAR(120) NOT NULL, pdpLocalName VARCHAR(120) NOT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_PDPSUBGROUP_PDP (pdpParentKeyName, pdpParentKeyVersion, pdpParentLocalName, pdpLocalName, parentLocalName, localName, parentKeyVersion, parentKeyName)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0500-pdpsubgroup.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS pdpsubgroup (CURRENTINSTANCECOUNT INT DEFAULT NULL, DESIREDINSTANCECOUNT INT DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_PDPSUBGROUP (parentLocalName, localName, parentKeyVersion, parentKeyName)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0510-toscacapabilityassignment.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilityassignment (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, type_name VARCHAR(255) NULL, type_version VARCHAR(255) NULL, PRIMARY KEY PK_TOSCACAPABILITYASSIGNMENT(name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0520-toscacapabilityassignments.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilityassignments (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCACAPABILITYASSIGNMENTS (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0530-toscacapabilityassignments_toscacapabilityassignment.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilityassignments_toscacapabilityassignment (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCACAPABILITYASSIGNMENTS_TOSCACAPABILITYASSIGNMENT (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0540-toscacapabilitytype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilitytype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCACAPABILITYTYPE (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0550-toscacapabilitytypes.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilitytypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCACAPABILITYTYPES (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0560-toscacapabilitytypes_toscacapabilitytype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscacapabilitytypes_toscacapabilitytype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCACAPABILITYTYPES_TOSCACAPABILITYTYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0570-toscadatatype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscadatatype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCADATATYPE (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0580-toscadatatypes.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscadatatypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCADATATYPES (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0590-toscadatatypes_toscadatatype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscadatatypes_toscadatatype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCADATATYPES_TOSCADATATYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0600-toscanodetemplate.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetemplate (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, type_name VARCHAR(255) NULL, type_version VARCHAR(255) NULL, capabilitiesName VARCHAR(120) NULL, capabilitiesVersion VARCHAR(20) NULL, requirementsVersion VARCHAR(20) NULL, requirementsName VARCHAR(120) NULL, PRIMARY KEY PK_TOSCANODETEMPLATE (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0610-toscanodetemplates.sql policy-db-migrator | -------------- kafka | log.cleaner.enable = true kafka | log.cleaner.io.buffer.load.factor = 0.9 kafka | log.cleaner.io.buffer.size = 524288 kafka | log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308 kafka | log.cleaner.max.compaction.lag.ms = 9223372036854775807 kafka | log.cleaner.min.cleanable.ratio = 0.5 kafka | log.cleaner.min.compaction.lag.ms = 0 kafka | log.cleaner.threads = 1 kafka | log.cleanup.policy = [delete] kafka | log.dir = /tmp/kafka-logs kafka | log.dirs = /var/lib/kafka/data kafka | log.flush.interval.messages = 9223372036854775807 kafka | log.flush.interval.ms = null kafka | log.flush.offset.checkpoint.interval.ms = 60000 kafka | log.flush.scheduler.interval.ms = 9223372036854775807 kafka | log.flush.start.offset.checkpoint.interval.ms = 60000 kafka | log.index.interval.bytes = 4096 kafka | log.index.size.max.bytes = 10485760 kafka | log.local.retention.bytes = -2 kafka | log.local.retention.ms = -2 kafka | log.message.downconversion.enable = true kafka | log.message.format.version = 3.0-IV1 kafka | log.message.timestamp.after.max.ms = 9223372036854775807 kafka | log.message.timestamp.before.max.ms = 9223372036854775807 policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetemplates (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCANODETEMPLATES (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0620-toscanodetemplates_toscanodetemplate.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetemplates_toscanodetemplate (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCANODETEMPLATES_TOSCANODETEMPLATE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0630-toscanodetype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, requirementsVersion VARCHAR(20) NULL, requirementsName VARCHAR(120) NULL, PRIMARY KEY PK_TOSCANODETYPE (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0640-toscanodetypes.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCANODETYPES (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0650-toscanodetypes_toscanodetype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscanodetypes_toscanodetype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCANODETYPES_TOSCANODETYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0660-toscaparameter.sql policy-db-migrator | -------------- kafka | log.message.timestamp.difference.max.ms = 9223372036854775807 kafka | log.message.timestamp.type = CreateTime kafka | log.preallocate = false kafka | log.retention.bytes = -1 kafka | log.retention.check.interval.ms = 300000 kafka | log.retention.hours = 168 kafka | log.retention.minutes = null kafka | log.retention.ms = null kafka | log.roll.hours = 168 kafka | log.roll.jitter.hours = 0 kafka | log.roll.jitter.ms = null kafka | log.roll.ms = null kafka | log.segment.bytes = 1073741824 kafka | log.segment.delete.delay.ms = 60000 kafka | max.connection.creation.rate = 2147483647 kafka | max.connections = 2147483647 kafka | max.connections.per.ip = 2147483647 kafka | max.connections.per.ip.overrides = kafka | max.incremental.fetch.session.cache.slots = 1000 kafka | message.max.bytes = 1048588 kafka | metadata.log.dir = null kafka | metadata.log.max.record.bytes.between.snapshots = 20971520 kafka | metadata.log.max.snapshot.interval.ms = 3600000 kafka | metadata.log.segment.bytes = 1073741824 kafka | metadata.log.segment.min.bytes = 8388608 kafka | metadata.log.segment.ms = 604800000 kafka | metadata.max.idle.interval.ms = 500 kafka | metadata.max.retention.bytes = 104857600 kafka | metadata.max.retention.ms = 604800000 kafka | metric.reporters = [] kafka | metrics.num.samples = 2 kafka | metrics.recording.level = INFO kafka | metrics.sample.window.ms = 30000 kafka | min.insync.replicas = 1 kafka | node.id = 1 kafka | num.io.threads = 8 kafka | num.network.threads = 3 kafka | num.partitions = 1 kafka | num.recovery.threads.per.data.dir = 1 kafka | num.replica.alter.log.dirs.threads = null kafka | num.replica.fetchers = 1 kafka | offset.metadata.max.bytes = 4096 kafka | offsets.commit.required.acks = -1 kafka | offsets.commit.timeout.ms = 5000 kafka | offsets.load.buffer.size = 5242880 kafka | offsets.retention.check.interval.ms = 600000 kafka | offsets.retention.minutes = 10080 kafka | offsets.topic.compression.codec = 0 kafka | offsets.topic.num.partitions = 50 kafka | offsets.topic.replication.factor = 1 kafka | offsets.topic.segment.bytes = 104857600 kafka | password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding kafka | password.encoder.iterations = 4096 kafka | password.encoder.key.length = 128 kafka | password.encoder.keyfactory.algorithm = null kafka | password.encoder.old.secret = null kafka | password.encoder.secret = null kafka | principal.builder.class = class org.apache.kafka.common.security.authenticator.DefaultKafkaPrincipalBuilder kafka | process.roles = [] kafka | producer.id.expiration.check.interval.ms = 600000 kafka | producer.id.expiration.ms = 86400000 kafka | producer.purgatory.purge.interval.requests = 1000 kafka | queued.max.request.bytes = -1 kafka | queued.max.requests = 500 kafka | quota.window.num = 11 kafka | quota.window.size.seconds = 1 kafka | remote.log.index.file.cache.total.size.bytes = 1073741824 kafka | remote.log.manager.task.interval.ms = 30000 kafka | remote.log.manager.task.retry.backoff.max.ms = 30000 kafka | remote.log.manager.task.retry.backoff.ms = 500 kafka | remote.log.manager.task.retry.jitter = 0.2 kafka | remote.log.manager.thread.pool.size = 10 kafka | remote.log.metadata.custom.metadata.max.bytes = 128 kafka | remote.log.metadata.manager.class.name = org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemoteLogMetadataManager kafka | remote.log.metadata.manager.class.path = null kafka | remote.log.metadata.manager.impl.prefix = rlmm.config. kafka | remote.log.metadata.manager.listener.name = null kafka | remote.log.reader.max.pending.tasks = 100 kafka | remote.log.reader.threads = 10 kafka | remote.log.storage.manager.class.name = null kafka | remote.log.storage.manager.class.path = null kafka | remote.log.storage.manager.impl.prefix = rsm.config. kafka | remote.log.storage.system.enable = false kafka | replica.fetch.backoff.ms = 1000 kafka | replica.fetch.max.bytes = 1048576 kafka | replica.fetch.min.bytes = 1 kafka | replica.fetch.response.max.bytes = 10485760 kafka | replica.fetch.wait.max.ms = 500 kafka | replica.high.watermark.checkpoint.interval.ms = 5000 grafana | logger=migrator t=2024-04-24T09:37:37.504354195Z level=info msg="Migration successfully executed" id="drop temp_user_tmp_qwerty" duration=1.965387ms policy-db-migrator | CREATE TABLE IF NOT EXISTS toscaparameter (VALUE VARCHAR(255) NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCAPARAMETER (parentLocalName, localName, parentKeyVersion, parentKeyName)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0670-toscapolicies.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicies (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCAPOLICIES (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0680-toscapolicies_toscapolicy.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicies_toscapolicy (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCAPOLICIES_TOSCAPOLICY (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0690-toscapolicy.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicy (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, type_name VARCHAR(255) NULL, type_version VARCHAR(255) NULL, PRIMARY KEY PK_TOSCAPOLICY (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0700-toscapolicytype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicytype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCAPOLICYTYPE (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0710-toscapolicytypes.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicytypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCAPOLICYTYPES (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0720-toscapolicytypes_toscapolicytype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscapolicytypes_toscapolicytype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCAPOLICYTYPES_TOSCAPOLICYTYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0730-toscaproperty.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscaproperty (DEFAULTVALUE VARCHAR(255) DEFAULT NULL, `DESCRIPTION` VARCHAR(255) DEFAULT NULL, ENTRYSCHEMA LONGBLOB DEFAULT NULL, REQUIRED BOOLEAN DEFAULT 0, STATUS INT DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, name VARCHAR(120) DEFAULT NULL, version VARCHAR(20) DEFAULT NULL, PRIMARY KEY PK_TOSCAPROPERTY (parentLocalName, localName, parentKeyVersion, parentKeyName)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0740-toscarelationshiptype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarelationshiptype (`DESCRIPTION` VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCARELATIONSHIPTYPE (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0750-toscarelationshiptypes.sql policy-db-migrator | -------------- grafana | logger=migrator t=2024-04-24T09:37:37.510024216Z level=info msg="Executing migration" id="Set created for temp users that will otherwise prematurely expire" grafana | logger=migrator t=2024-04-24T09:37:37.510476014Z level=info msg="Migration successfully executed" id="Set created for temp users that will otherwise prematurely expire" duration=452.608µs grafana | logger=migrator t=2024-04-24T09:37:37.514057363Z level=info msg="Executing migration" id="create star table" grafana | logger=migrator t=2024-04-24T09:37:37.514789451Z level=info msg="Migration successfully executed" id="create star table" duration=731.878µs grafana | logger=migrator t=2024-04-24T09:37:37.518129042Z level=info msg="Executing migration" id="add unique index star.user_id_dashboard_id" grafana | logger=migrator t=2024-04-24T09:37:37.518895372Z level=info msg="Migration successfully executed" id="add unique index star.user_id_dashboard_id" duration=768.48µs grafana | logger=migrator t=2024-04-24T09:37:37.523949239Z level=info msg="Executing migration" id="create org table v1" grafana | logger=migrator t=2024-04-24T09:37:37.524780041Z level=info msg="Migration successfully executed" id="create org table v1" duration=830.362µs grafana | logger=migrator t=2024-04-24T09:37:37.567822849Z level=info msg="Executing migration" id="create index UQE_org_name - v1" grafana | logger=migrator t=2024-04-24T09:37:37.56989298Z level=info msg="Migration successfully executed" id="create index UQE_org_name - v1" duration=2.074301ms grafana | logger=migrator t=2024-04-24T09:37:37.573959399Z level=info msg="Executing migration" id="create org_user table v1" grafana | logger=migrator t=2024-04-24T09:37:37.575358453Z level=info msg="Migration successfully executed" id="create org_user table v1" duration=1.399354ms grafana | logger=migrator t=2024-04-24T09:37:37.578963273Z level=info msg="Executing migration" id="create index IDX_org_user_org_id - v1" grafana | logger=migrator t=2024-04-24T09:37:37.579691192Z level=info msg="Migration successfully executed" id="create index IDX_org_user_org_id - v1" duration=727.449µs grafana | logger=migrator t=2024-04-24T09:37:37.582704609Z level=info msg="Executing migration" id="create index UQE_org_user_org_id_user_id - v1" grafana | logger=migrator t=2024-04-24T09:37:37.583462869Z level=info msg="Migration successfully executed" id="create index UQE_org_user_org_id_user_id - v1" duration=757.17µs grafana | logger=migrator t=2024-04-24T09:37:37.586614991Z level=info msg="Executing migration" id="create index IDX_org_user_user_id - v1" grafana | logger=migrator t=2024-04-24T09:37:37.587404592Z level=info msg="Migration successfully executed" id="create index IDX_org_user_user_id - v1" duration=788.111µs grafana | logger=migrator t=2024-04-24T09:37:37.590874688Z level=info msg="Executing migration" id="Update org table charset" grafana | logger=migrator t=2024-04-24T09:37:37.590913749Z level=info msg="Migration successfully executed" id="Update org table charset" duration=40.111µs grafana | logger=migrator t=2024-04-24T09:37:37.594483418Z level=info msg="Executing migration" id="Update org_user table charset" grafana | logger=migrator t=2024-04-24T09:37:37.594549491Z level=info msg="Migration successfully executed" id="Update org_user table charset" duration=44.992µs grafana | logger=migrator t=2024-04-24T09:37:37.598226314Z level=info msg="Executing migration" id="Migrate all Read Only Viewers to Viewers" grafana | logger=migrator t=2024-04-24T09:37:37.598453033Z level=info msg="Migration successfully executed" id="Migrate all Read Only Viewers to Viewers" duration=226.949µs grafana | logger=migrator t=2024-04-24T09:37:37.601539053Z level=info msg="Executing migration" id="create dashboard table" grafana | logger=migrator t=2024-04-24T09:37:37.602251511Z level=info msg="Migration successfully executed" id="create dashboard table" duration=712.278µs grafana | logger=migrator t=2024-04-24T09:37:37.60529812Z level=info msg="Executing migration" id="add index dashboard.account_id" grafana | logger=migrator t=2024-04-24T09:37:37.607195323Z level=info msg="Migration successfully executed" id="add index dashboard.account_id" duration=1.895073ms grafana | logger=migrator t=2024-04-24T09:37:37.610879958Z level=info msg="Executing migration" id="add unique index dashboard_account_id_slug" grafana | logger=migrator t=2024-04-24T09:37:37.612451468Z level=info msg="Migration successfully executed" id="add unique index dashboard_account_id_slug" duration=1.570071ms grafana | logger=migrator t=2024-04-24T09:37:37.618297826Z level=info msg="Executing migration" id="create dashboard_tag table" grafana | logger=migrator t=2024-04-24T09:37:37.619217323Z level=info msg="Migration successfully executed" id="create dashboard_tag table" duration=920.766µs grafana | logger=migrator t=2024-04-24T09:37:37.622732239Z level=info msg="Executing migration" id="add unique index dashboard_tag.dasboard_id_term" grafana | logger=migrator t=2024-04-24T09:37:37.624191327Z level=info msg="Migration successfully executed" id="add unique index dashboard_tag.dasboard_id_term" duration=1.460228ms grafana | logger=migrator t=2024-04-24T09:37:37.62786693Z level=info msg="Executing migration" id="drop index UQE_dashboard_tag_dashboard_id_term - v1" grafana | logger=migrator t=2024-04-24T09:37:37.629062856Z level=info msg="Migration successfully executed" id="drop index UQE_dashboard_tag_dashboard_id_term - v1" duration=1.195647ms grafana | logger=migrator t=2024-04-24T09:37:37.634259919Z level=info msg="Executing migration" id="Rename table dashboard to dashboard_v1 - v1" grafana | logger=migrator t=2024-04-24T09:37:37.63969181Z level=info msg="Migration successfully executed" id="Rename table dashboard to dashboard_v1 - v1" duration=5.428542ms grafana | logger=migrator t=2024-04-24T09:37:37.643606853Z level=info msg="Executing migration" id="create dashboard v2" grafana | logger=migrator t=2024-04-24T09:37:37.644533469Z level=info msg="Migration successfully executed" id="create dashboard v2" duration=919.907µs grafana | logger=migrator t=2024-04-24T09:37:37.648007054Z level=info msg="Executing migration" id="create index IDX_dashboard_org_id - v2" grafana | logger=migrator t=2024-04-24T09:37:37.648910289Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_org_id - v2" duration=903.435µs grafana | logger=migrator t=2024-04-24T09:37:37.654158264Z level=info msg="Executing migration" id="create index UQE_dashboard_org_id_slug - v2" grafana | logger=migrator t=2024-04-24T09:37:37.655362752Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_org_id_slug - v2" duration=1.206747ms grafana | logger=migrator t=2024-04-24T09:37:37.660523512Z level=info msg="Executing migration" id="copy dashboard v1 to v2" grafana | logger=migrator t=2024-04-24T09:37:37.661066934Z level=info msg="Migration successfully executed" id="copy dashboard v1 to v2" duration=544.342µs grafana | logger=migrator t=2024-04-24T09:37:37.664521368Z level=info msg="Executing migration" id="drop table dashboard_v1" grafana | logger=migrator t=2024-04-24T09:37:37.665651652Z level=info msg="Migration successfully executed" id="drop table dashboard_v1" duration=1.129114ms grafana | logger=migrator t=2024-04-24T09:37:37.671779092Z level=info msg="Executing migration" id="alter dashboard.data to mediumtext v1" grafana | logger=migrator t=2024-04-24T09:37:37.672098724Z level=info msg="Migration successfully executed" id="alter dashboard.data to mediumtext v1" duration=319.833µs grafana | logger=migrator t=2024-04-24T09:37:37.675748976Z level=info msg="Executing migration" id="Add column updated_by in dashboard - v2" grafana | logger=migrator t=2024-04-24T09:37:37.677883109Z level=info msg="Migration successfully executed" id="Add column updated_by in dashboard - v2" duration=2.133673ms grafana | logger=migrator t=2024-04-24T09:37:37.681039292Z level=info msg="Executing migration" id="Add column created_by in dashboard - v2" policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarelationshiptypes (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCARELATIONSHIPTYPES (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0760-toscarelationshiptypes_toscarelationshiptype.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarelationshiptypes_toscarelationshiptype (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCARELATIONSHIPTYPES_TOSCARELATIONSHIPTYPE (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0770-toscarequirement.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarequirement (CAPABILITY VARCHAR(255) NULL, `DESCRIPTION` VARCHAR(255) NULL, NODE VARCHAR(255) NULL, RELATIONSHIP VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, type_name VARCHAR(255) NULL, type_version VARCHAR(255) NULL, PRIMARY KEY PK_TOSCAREQUIREMENT (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0780-toscarequirements.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarequirements (name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_TOSCAREQUIREMENTS (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0790-toscarequirements_toscarequirement.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscarequirements_toscarequirement (conceptContainerMapName VARCHAR(120) NOT NULL, concpetContainerMapVersion VARCHAR(20) NOT NULL, conceptContainerName VARCHAR(120) NOT NULL, conceptContainerVersion VARCHAR(20) NOT NULL, name VARCHAR(120) NULL, version VARCHAR(20) NULL, PRIMARY KEY PK_TOSCAREQUIREMENTS_TOSCAREQUIREMENT (conceptContainerMapName, concpetContainerMapVersion, conceptContainerName, conceptContainerVersion)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0800-toscaservicetemplate.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscaservicetemplate (`DESCRIPTION` VARCHAR(255) NULL, TOSCADEFINITIONSVERSION VARCHAR(255) NULL, derived_from_name VARCHAR(255) NULL, derived_from_version VARCHAR(255) NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, capabilityTypesVersion VARCHAR(20) NULL, capabilityTypesName VARCHAR(120) NULL, dataTypesName VARCHAR(120) NULL, dataTypesVersion VARCHAR(20) NULL, nodeTypesVersion VARCHAR(20) NULL, nodeTypesName VARCHAR(120) NULL, policyTypesName VARCHAR(120) NULL, policyTypesVersion VARCHAR(20) NULL, relationshipTypesVersion VARCHAR(20) NULL, relationshipTypesName VARCHAR(120) NULL, topologyTemplateLocalName VARCHAR(120) NULL, topologyTemplateParentKeyName VARCHAR(120) NULL, topologyTemplateParentKeyVersion VARCHAR(15) NULL, topologyTemplateParentLocalName VARCHAR(120) NULL, PRIMARY KEY PK_TOSCASERVICETEMPLATE (name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0810-toscatopologytemplate.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscatopologytemplate (`description` VARCHAR(255) NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, nodeTemplatessVersion VARCHAR(20) NULL, nodeTemplatesName VARCHAR(120) NULL, policyVersion VARCHAR(20) NULL, policyName VARCHAR(120) NULL, PRIMARY KEY PK_TOSCATOPOLOGYTEMPLATE (parentLocalName, localName, parentKeyVersion, parentKeyName)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | grafana | logger=migrator t=2024-04-24T09:37:37.682985128Z level=info msg="Migration successfully executed" id="Add column created_by in dashboard - v2" duration=1.944896ms grafana | logger=migrator t=2024-04-24T09:37:37.688513563Z level=info msg="Executing migration" id="Add column gnetId in dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.691569603Z level=info msg="Migration successfully executed" id="Add column gnetId in dashboard" duration=3.05567ms grafana | logger=migrator t=2024-04-24T09:37:37.695715454Z level=info msg="Executing migration" id="Add index for gnetId in dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.696875919Z level=info msg="Migration successfully executed" id="Add index for gnetId in dashboard" duration=1.163525ms grafana | logger=migrator t=2024-04-24T09:37:37.707852337Z level=info msg="Executing migration" id="Add column plugin_id in dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.709807763Z level=info msg="Migration successfully executed" id="Add column plugin_id in dashboard" duration=1.960326ms grafana | logger=migrator t=2024-04-24T09:37:37.714279818Z level=info msg="Executing migration" id="Add index for plugin_id in dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.715301578Z level=info msg="Migration successfully executed" id="Add index for plugin_id in dashboard" duration=1.021889ms grafana | logger=migrator t=2024-04-24T09:37:37.719539963Z level=info msg="Executing migration" id="Add index for dashboard_id in dashboard_tag" grafana | logger=migrator t=2024-04-24T09:37:37.720467809Z level=info msg="Migration successfully executed" id="Add index for dashboard_id in dashboard_tag" duration=927.176µs grafana | logger=migrator t=2024-04-24T09:37:37.724168053Z level=info msg="Executing migration" id="Update dashboard table charset" grafana | logger=migrator t=2024-04-24T09:37:37.724275437Z level=info msg="Migration successfully executed" id="Update dashboard table charset" duration=113.234µs grafana | logger=migrator t=2024-04-24T09:37:37.737206192Z level=info msg="Executing migration" id="Update dashboard_tag table charset" grafana | logger=migrator t=2024-04-24T09:37:37.737319756Z level=info msg="Migration successfully executed" id="Update dashboard_tag table charset" duration=120.765µs grafana | logger=migrator t=2024-04-24T09:37:37.741063792Z level=info msg="Executing migration" id="Add column folder_id in dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.743357291Z level=info msg="Migration successfully executed" id="Add column folder_id in dashboard" duration=2.28994ms grafana | logger=migrator t=2024-04-24T09:37:37.747104447Z level=info msg="Executing migration" id="Add column isFolder in dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.749099065Z level=info msg="Migration successfully executed" id="Add column isFolder in dashboard" duration=1.994718ms grafana | logger=migrator t=2024-04-24T09:37:37.754553848Z level=info msg="Executing migration" id="Add column has_acl in dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.756888709Z level=info msg="Migration successfully executed" id="Add column has_acl in dashboard" duration=2.335781ms grafana | logger=migrator t=2024-04-24T09:37:37.760433287Z level=info msg="Executing migration" id="Add column uid in dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.762625072Z level=info msg="Migration successfully executed" id="Add column uid in dashboard" duration=2.192946ms grafana | logger=migrator t=2024-04-24T09:37:37.766233823Z level=info msg="Executing migration" id="Update uid column values in dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.766618438Z level=info msg="Migration successfully executed" id="Update uid column values in dashboard" duration=385.884µs grafana | logger=migrator t=2024-04-24T09:37:37.770066853Z level=info msg="Executing migration" id="Add unique index dashboard_org_id_uid" grafana | logger=migrator t=2024-04-24T09:37:37.770969728Z level=info msg="Migration successfully executed" id="Add unique index dashboard_org_id_uid" duration=905.894µs grafana | logger=migrator t=2024-04-24T09:37:37.776308306Z level=info msg="Executing migration" id="Remove unique index org_id_slug" grafana | logger=migrator t=2024-04-24T09:37:37.777837415Z level=info msg="Migration successfully executed" id="Remove unique index org_id_slug" duration=1.531969ms grafana | logger=migrator t=2024-04-24T09:37:37.78127502Z level=info msg="Executing migration" id="Update dashboard title length" grafana | logger=migrator t=2024-04-24T09:37:37.781304701Z level=info msg="Migration successfully executed" id="Update dashboard title length" duration=30.982µs grafana | logger=migrator t=2024-04-24T09:37:37.78462526Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_title_folder_id" grafana | logger=migrator t=2024-04-24T09:37:37.785717622Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_title_folder_id" duration=1.092012ms grafana | logger=migrator t=2024-04-24T09:37:37.790902664Z level=info msg="Executing migration" id="create dashboard_provisioning" grafana | logger=migrator t=2024-04-24T09:37:37.791791969Z level=info msg="Migration successfully executed" id="create dashboard_provisioning" duration=888.995µs grafana | logger=migrator t=2024-04-24T09:37:37.795349118Z level=info msg="Executing migration" id="Rename table dashboard_provisioning to dashboard_provisioning_tmp_qwerty - v1" grafana | logger=migrator t=2024-04-24T09:37:37.800420475Z level=info msg="Migration successfully executed" id="Rename table dashboard_provisioning to dashboard_provisioning_tmp_qwerty - v1" duration=5.068777ms grafana | logger=migrator t=2024-04-24T09:37:37.804210333Z level=info msg="Executing migration" id="create dashboard_provisioning v2" grafana | logger=migrator t=2024-04-24T09:37:37.80566365Z level=info msg="Migration successfully executed" id="create dashboard_provisioning v2" duration=1.455666ms grafana | logger=migrator t=2024-04-24T09:37:37.810927325Z level=info msg="Executing migration" id="create index IDX_dashboard_provisioning_dashboard_id - v2" grafana | logger=migrator t=2024-04-24T09:37:37.811990577Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_provisioning_dashboard_id - v2" duration=1.060692ms grafana | logger=migrator t=2024-04-24T09:37:37.815722382Z level=info msg="Executing migration" id="create index IDX_dashboard_provisioning_dashboard_id_name - v2" grafana | logger=migrator t=2024-04-24T09:37:37.816553764Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_provisioning_dashboard_id_name - v2" duration=830.452µs grafana | logger=migrator t=2024-04-24T09:37:37.820816211Z level=info msg="Executing migration" id="copy dashboard_provisioning v1 to v2" grafana | logger=migrator t=2024-04-24T09:37:37.821449025Z level=info msg="Migration successfully executed" id="copy dashboard_provisioning v1 to v2" duration=632.004µs grafana | logger=migrator t=2024-04-24T09:37:37.827575784Z level=info msg="Executing migration" id="drop dashboard_provisioning_tmp_qwerty" grafana | logger=migrator t=2024-04-24T09:37:37.828672727Z level=info msg="Migration successfully executed" id="drop dashboard_provisioning_tmp_qwerty" duration=1.098383ms grafana | logger=migrator t=2024-04-24T09:37:37.832535877Z level=info msg="Executing migration" id="Add check_sum column" grafana | logger=migrator t=2024-04-24T09:37:37.835258864Z level=info msg="Migration successfully executed" id="Add check_sum column" duration=2.723946ms grafana | logger=migrator t=2024-04-24T09:37:37.838707108Z level=info msg="Executing migration" id="Add index for dashboard_title" grafana | logger=migrator t=2024-04-24T09:37:37.839720417Z level=info msg="Migration successfully executed" id="Add index for dashboard_title" duration=1.013769ms grafana | logger=migrator t=2024-04-24T09:37:37.844742903Z level=info msg="Executing migration" id="delete tags for deleted dashboards" grafana | logger=migrator t=2024-04-24T09:37:37.845054125Z level=info msg="Migration successfully executed" id="delete tags for deleted dashboards" duration=311.302µs grafana | logger=migrator t=2024-04-24T09:37:37.848678417Z level=info msg="Executing migration" id="delete stars for deleted dashboards" grafana | logger=migrator t=2024-04-24T09:37:37.848973328Z level=info msg="Migration successfully executed" id="delete stars for deleted dashboards" duration=295.311µs grafana | logger=migrator t=2024-04-24T09:37:37.852587399Z level=info msg="Executing migration" id="Add index for dashboard_is_folder" grafana | logger=migrator t=2024-04-24T09:37:37.853603238Z level=info msg="Migration successfully executed" id="Add index for dashboard_is_folder" duration=1.016129ms grafana | logger=migrator t=2024-04-24T09:37:37.85826477Z level=info msg="Executing migration" id="Add isPublic for dashboard" grafana | logger=migrator t=2024-04-24T09:37:37.860460446Z level=info msg="Migration successfully executed" id="Add isPublic for dashboard" duration=2.194986ms grafana | logger=migrator t=2024-04-24T09:37:37.864181271Z level=info msg="Executing migration" id="create data_source table" grafana | logger=migrator t=2024-04-24T09:37:37.865225482Z level=info msg="Migration successfully executed" id="create data_source table" duration=1.043991ms grafana | logger=migrator t=2024-04-24T09:37:37.868936586Z level=info msg="Executing migration" id="add index data_source.account_id" grafana | logger=migrator t=2024-04-24T09:37:37.870388322Z level=info msg="Migration successfully executed" id="add index data_source.account_id" duration=1.454846ms grafana | logger=migrator t=2024-04-24T09:37:37.875032574Z level=info msg="Executing migration" id="add unique index data_source.account_id_name" grafana | logger=migrator t=2024-04-24T09:37:37.876399798Z level=info msg="Migration successfully executed" id="add unique index data_source.account_id_name" duration=1.367243ms grafana | logger=migrator t=2024-04-24T09:37:37.880278008Z level=info msg="Executing migration" id="drop index IDX_data_source_account_id - v1" grafana | logger=migrator t=2024-04-24T09:37:37.881163113Z level=info msg="Migration successfully executed" id="drop index IDX_data_source_account_id - v1" duration=885.254µs grafana | logger=migrator t=2024-04-24T09:37:37.884622467Z level=info msg="Executing migration" id="drop index UQE_data_source_account_id_name - v1" grafana | logger=migrator t=2024-04-24T09:37:37.885685109Z level=info msg="Migration successfully executed" id="drop index UQE_data_source_account_id_name - v1" duration=1.061812ms grafana | logger=migrator t=2024-04-24T09:37:37.890531878Z level=info msg="Executing migration" id="Rename table data_source to data_source_v1 - v1" grafana | logger=migrator t=2024-04-24T09:37:37.89877492Z level=info msg="Migration successfully executed" id="Rename table data_source to data_source_v1 - v1" duration=8.237111ms grafana | logger=migrator t=2024-04-24T09:37:37.902883349Z level=info msg="Executing migration" id="create data_source table v2" grafana | logger=migrator t=2024-04-24T09:37:37.903873868Z level=info msg="Migration successfully executed" id="create data_source table v2" duration=985.678µs grafana | logger=migrator t=2024-04-24T09:37:37.907560642Z level=info msg="Executing migration" id="create index IDX_data_source_org_id - v2" grafana | logger=migrator t=2024-04-24T09:37:37.908421725Z level=info msg="Migration successfully executed" id="create index IDX_data_source_org_id - v2" duration=860.963µs grafana | logger=migrator t=2024-04-24T09:37:37.91315015Z level=info msg="Executing migration" id="create index UQE_data_source_org_id_name - v2" grafana | logger=migrator t=2024-04-24T09:37:37.914005793Z level=info msg="Migration successfully executed" id="create index UQE_data_source_org_id_name - v2" duration=855.123µs grafana | logger=migrator t=2024-04-24T09:37:37.917896804Z level=info msg="Executing migration" id="Drop old table data_source_v1 #2" grafana | logger=migrator t=2024-04-24T09:37:37.918456337Z level=info msg="Migration successfully executed" id="Drop old table data_source_v1 #2" duration=558.713µs grafana | logger=migrator t=2024-04-24T09:37:37.923656189Z level=info msg="Executing migration" id="Add column with_credentials" grafana | logger=migrator t=2024-04-24T09:37:37.92753288Z level=info msg="Migration successfully executed" id="Add column with_credentials" duration=3.876271ms grafana | logger=migrator t=2024-04-24T09:37:37.931312637Z level=info msg="Executing migration" id="Add secure json data column" grafana | logger=migrator t=2024-04-24T09:37:37.93369747Z level=info msg="Migration successfully executed" id="Add secure json data column" duration=2.385243ms grafana | logger=migrator t=2024-04-24T09:37:37.937305691Z level=info msg="Executing migration" id="Update data_source table charset" grafana | logger=migrator t=2024-04-24T09:37:37.937332502Z level=info msg="Migration successfully executed" id="Update data_source table charset" duration=27.261µs grafana | logger=migrator t=2024-04-24T09:37:37.940987645Z level=info msg="Executing migration" id="Update initial version to 1" grafana | logger=migrator t=2024-04-24T09:37:37.941206794Z level=info msg="Migration successfully executed" id="Update initial version to 1" duration=215.398µs grafana | logger=migrator t=2024-04-24T09:37:37.945572433Z level=info msg="Executing migration" id="Add read_only data column" grafana | logger=migrator t=2024-04-24T09:37:37.947957816Z level=info msg="Migration successfully executed" id="Add read_only data column" duration=2.380783ms grafana | logger=migrator t=2024-04-24T09:37:37.951495724Z level=info msg="Executing migration" id="Migrate logging ds to loki ds" grafana | logger=migrator t=2024-04-24T09:37:37.951700602Z level=info msg="Migration successfully executed" id="Migrate logging ds to loki ds" duration=204.988µs grafana | logger=migrator t=2024-04-24T09:37:37.955422848Z level=info msg="Executing migration" id="Update json_data with nulls" grafana | logger=migrator t=2024-04-24T09:37:37.955584714Z level=info msg="Migration successfully executed" id="Update json_data with nulls" duration=161.677µs grafana | logger=migrator t=2024-04-24T09:37:37.959129142Z level=info msg="Executing migration" id="Add uid column" grafana | logger=migrator t=2024-04-24T09:37:37.962868307Z level=info msg="Migration successfully executed" id="Add uid column" duration=3.734465ms grafana | logger=migrator t=2024-04-24T09:37:37.9672969Z level=info msg="Executing migration" id="Update uid value" grafana | logger=migrator t=2024-04-24T09:37:37.967508788Z level=info msg="Migration successfully executed" id="Update uid value" duration=208.978µs grafana | logger=migrator t=2024-04-24T09:37:37.971052356Z level=info msg="Executing migration" id="Add unique index datasource_org_id_uid" grafana | logger=migrator t=2024-04-24T09:37:37.971854578Z level=info msg="Migration successfully executed" id="Add unique index datasource_org_id_uid" duration=801.602µs grafana | logger=migrator t=2024-04-24T09:37:37.975555272Z level=info msg="Executing migration" id="add unique index datasource_org_id_is_default" grafana | logger=migrator t=2024-04-24T09:37:37.97679905Z level=info msg="Migration successfully executed" id="add unique index datasource_org_id_is_default" duration=1.242568ms grafana | logger=migrator t=2024-04-24T09:37:37.988286278Z level=info msg="Executing migration" id="create api_key table" grafana | logger=migrator t=2024-04-24T09:37:37.989696823Z level=info msg="Migration successfully executed" id="create api_key table" duration=1.409435ms grafana | logger=migrator t=2024-04-24T09:37:37.99420702Z level=info msg="Executing migration" id="add index api_key.account_id" grafana | logger=migrator t=2024-04-24T09:37:37.995560682Z level=info msg="Migration successfully executed" id="add index api_key.account_id" duration=1.353103ms grafana | logger=migrator t=2024-04-24T09:37:37.999479434Z level=info msg="Executing migration" id="add index api_key.key" grafana | logger=migrator t=2024-04-24T09:37:38.000682402Z level=info msg="Migration successfully executed" id="add index api_key.key" duration=1.202697ms grafana | logger=migrator t=2024-04-24T09:37:38.007317691Z level=info msg="Executing migration" id="add index api_key.account_id_name" grafana | logger=migrator t=2024-04-24T09:37:38.008654874Z level=info msg="Migration successfully executed" id="add index api_key.account_id_name" duration=1.337033ms grafana | logger=migrator t=2024-04-24T09:37:38.01239981Z level=info msg="Executing migration" id="drop index IDX_api_key_account_id - v1" grafana | logger=migrator t=2024-04-24T09:37:38.013199921Z level=info msg="Migration successfully executed" id="drop index IDX_api_key_account_id - v1" duration=801.891µs grafana | logger=migrator t=2024-04-24T09:37:38.016982619Z level=info msg="Executing migration" id="drop index UQE_api_key_key - v1" kafka | replica.lag.time.max.ms = 30000 kafka | replica.selector.class = null kafka | replica.socket.receive.buffer.bytes = 65536 kafka | replica.socket.timeout.ms = 30000 kafka | replication.quota.window.num = 11 kafka | replication.quota.window.size.seconds = 1 kafka | request.timeout.ms = 30000 kafka | reserved.broker.max.id = 1000 kafka | sasl.client.callback.handler.class = null kafka | sasl.enabled.mechanisms = [GSSAPI] kafka | sasl.jaas.config = null kafka | sasl.kerberos.kinit.cmd = /usr/bin/kinit kafka | sasl.kerberos.min.time.before.relogin = 60000 kafka | sasl.kerberos.principal.to.local.rules = [DEFAULT] kafka | sasl.kerberos.service.name = null kafka | sasl.kerberos.ticket.renew.jitter = 0.05 kafka | sasl.kerberos.ticket.renew.window.factor = 0.8 kafka | sasl.login.callback.handler.class = null kafka | sasl.login.class = null kafka | sasl.login.connect.timeout.ms = null kafka | sasl.login.read.timeout.ms = null kafka | sasl.login.refresh.buffer.seconds = 300 kafka | sasl.login.refresh.min.period.seconds = 60 kafka | sasl.login.refresh.window.factor = 0.8 kafka | sasl.login.refresh.window.jitter = 0.05 kafka | sasl.login.retry.backoff.max.ms = 10000 kafka | sasl.login.retry.backoff.ms = 100 kafka | sasl.mechanism.controller.protocol = GSSAPI kafka | sasl.mechanism.inter.broker.protocol = GSSAPI kafka | sasl.oauthbearer.clock.skew.seconds = 30 kafka | sasl.oauthbearer.expected.audience = null kafka | sasl.oauthbearer.expected.issuer = null kafka | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 kafka | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 kafka | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 kafka | sasl.oauthbearer.jwks.endpoint.url = null kafka | sasl.oauthbearer.scope.claim.name = scope kafka | sasl.oauthbearer.sub.claim.name = sub kafka | sasl.oauthbearer.token.endpoint.url = null kafka | sasl.server.callback.handler.class = null kafka | sasl.server.max.receive.size = 524288 kafka | security.inter.broker.protocol = PLAINTEXT kafka | security.providers = null kafka | server.max.startup.time.ms = 9223372036854775807 policy-db-migrator | > upgrade 0820-toscatrigger.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS toscatrigger (ACTION VARCHAR(255) NULL, toscaCondition LONGBLOB DEFAULT NULL, toscaConstraint LONGBLOB DEFAULT NULL, `DESCRIPTION` VARCHAR(255) NULL, EVALUATIONS INT DEFAULT NULL, EVENTTYPE VARCHAR(255) NULL, METHOD VARCHAR(255) NULL, `PERIOD` LONGBLOB DEFAULT NULL, SCHEDULE LONGBLOB DEFAULT NULL, TARGETFILTER LONGBLOB DEFAULT NULL, parentLocalName VARCHAR(120) NOT NULL, localName VARCHAR(120) NOT NULL, parentKeyVersion VARCHAR(15) NOT NULL, parentKeyName VARCHAR(120) NOT NULL, PRIMARY KEY PK_TOSCATRIGGER (parentLocalName, localName, parentKeyVersion, parentKeyName)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0830-FK_ToscaNodeTemplate_capabilitiesName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaNodeTemplate_capabilitiesName ON toscanodetemplate(capabilitiesName, capabilitiesVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0840-FK_ToscaNodeTemplate_requirementsName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaNodeTemplate_requirementsName ON toscanodetemplate(requirementsName, requirementsVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0850-FK_ToscaNodeType_requirementsName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaNodeType_requirementsName ON toscanodetype(requirementsName, requirementsVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0860-FK_ToscaServiceTemplate_capabilityTypesName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_capabilityTypesName ON toscaservicetemplate(capabilityTypesName, capabilityTypesVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0870-FK_ToscaServiceTemplate_dataTypesName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_dataTypesName ON toscaservicetemplate(dataTypesName, dataTypesVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0880-FK_ToscaServiceTemplate_nodeTypesName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_nodeTypesName ON toscaservicetemplate(nodeTypesName, nodeTypesVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0890-FK_ToscaServiceTemplate_policyTypesName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_policyTypesName ON toscaservicetemplate(policyTypesName, policyTypesVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0900-FK_ToscaServiceTemplate_relationshipTypesName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaServiceTemplate_relationshipTypesName ON toscaservicetemplate(relationshipTypesName, relationshipTypesVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0910-FK_ToscaTopologyTemplate_nodeTemplatesName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaTopologyTemplate_nodeTemplatesName ON toscatopologytemplate(nodeTemplatesName, nodeTemplatessVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0920-FK_ToscaTopologyTemplate_policyName.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX FK_ToscaTopologyTemplate_policyName ON toscatopologytemplate(policyName, policyVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0940-PdpPolicyStatus_PdpGroup.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX PdpPolicyStatus_PdpGroup ON pdppolicystatus(PDPGROUP) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0950-TscaServiceTemplatetopologyTemplateParentLocalName.sql policy-db-migrator | -------------- simulator | Policy simulator config file: /opt/app/policy/simulators/etc/mounted/simParameters.json simulator | overriding logback.xml simulator | 2024-04-24 09:37:42,918 INFO replacing 'HOST_NAME' with simulator in /opt/app/policy/simulators/etc/mounted/simParameters.json simulator | 2024-04-24 09:37:42,983 INFO org.onap.policy.models.simulators starting simulator | 2024-04-24 09:37:42,983 INFO org.onap.policy.models.simulators starting CDS gRPC Server Properties simulator | 2024-04-24 09:37:43,198 INFO org.onap.policy.models.simulators starting org.onap.policy.simulators.AaiSimulatorJaxRs_RESOURCE_LOCATION simulator | 2024-04-24 09:37:43,199 INFO org.onap.policy.models.simulators starting A&AI simulator simulator | 2024-04-24 09:37:43,325 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=A&AI simulator, host=0.0.0.0, port=6666, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@64a8c844{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@3f6db3fb{/,null,STOPPED}, connector=A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: WAITED-START simulator | 2024-04-24 09:37:43,339 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=A&AI simulator, host=0.0.0.0, port=6666, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@64a8c844{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@3f6db3fb{/,null,STOPPED}, connector=A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING simulator | 2024-04-24 09:37:43,342 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=A&AI simulator, host=0.0.0.0, port=6666, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@64a8c844{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@3f6db3fb{/,null,STOPPED}, connector=A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666}, jettyThread=Thread[A&AI simulator-6666,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING simulator | 2024-04-24 09:37:43,349 INFO jetty-11.0.20; built: 2024-01-29T21:04:22.394Z; git: 922f8dc188f7011e60d0361de585fd4ac4d63064; jvm 17.0.10+7-alpine-r0 simulator | 2024-04-24 09:37:43,414 INFO Session workerName=node0 simulator | 2024-04-24 09:37:44,063 INFO Using GSON for REST calls simulator | 2024-04-24 09:37:44,167 INFO Started o.e.j.s.ServletContextHandler@3f6db3fb{/,null,AVAILABLE} simulator | 2024-04-24 09:37:44,187 INFO Started A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666} simulator | 2024-04-24 09:37:44,214 INFO Started Server@64a8c844{STARTING}[11.0.20,sto=0] @1859ms simulator | 2024-04-24 09:37:44,214 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}}, swaggerId=null, toString()=JettyServletServer(name=A&AI simulator, host=0.0.0.0, port=6666, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@64a8c844{STARTED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@3f6db3fb{/,null,AVAILABLE}, connector=A&AI simulator@6f152006{HTTP/1.1, (http/1.1)}{0.0.0.0:6666}, jettyThread=Thread[A&AI simulator-6666,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-2e61d218==org.glassfish.jersey.servlet.ServletContainer@60d118f1{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}})]: pending time is 4127 ms. simulator | 2024-04-24 09:37:44,234 INFO org.onap.policy.models.simulators starting SDNC simulator simulator | 2024-04-24 09:37:44,237 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SDNC simulator, host=0.0.0.0, port=6668, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@70efb718{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@b70da4c{/,null,STOPPED}, connector=SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: WAITED-START simulator | 2024-04-24 09:37:44,237 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SDNC simulator, host=0.0.0.0, port=6668, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@70efb718{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@b70da4c{/,null,STOPPED}, connector=SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING simulator | 2024-04-24 09:37:44,239 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SDNC simulator, host=0.0.0.0, port=6668, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@70efb718{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@b70da4c{/,null,STOPPED}, connector=SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668}, jettyThread=Thread[SDNC simulator-6668,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING simulator | 2024-04-24 09:37:44,241 INFO jetty-11.0.20; built: 2024-01-29T21:04:22.394Z; git: 922f8dc188f7011e60d0361de585fd4ac4d63064; jvm 17.0.10+7-alpine-r0 simulator | 2024-04-24 09:37:44,255 INFO Session workerName=node0 simulator | 2024-04-24 09:37:44,398 INFO Using GSON for REST calls simulator | 2024-04-24 09:37:44,409 INFO Started o.e.j.s.ServletContextHandler@b70da4c{/,null,AVAILABLE} simulator | 2024-04-24 09:37:44,425 INFO Started SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668} simulator | 2024-04-24 09:37:44,426 INFO Started Server@70efb718{STARTING}[11.0.20,sto=0] @2071ms policy-db-migrator | CREATE INDEX TscaServiceTemplatetopologyTemplateParentLocalName ON toscaservicetemplate(topologyTemplateParentLocalName, topologyTemplateLocalName, topologyTemplateParentKeyVersion, topologyTemplateParentKeyName) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0960-FK_ToscaNodeTemplate_capabilitiesName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscanodetemplate ADD CONSTRAINT FK_ToscaNodeTemplate_capabilitiesName FOREIGN KEY (capabilitiesName, capabilitiesVersion) REFERENCES toscacapabilityassignments (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0970-FK_ToscaNodeTemplate_requirementsName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscanodetemplate ADD CONSTRAINT FK_ToscaNodeTemplate_requirementsName FOREIGN KEY (requirementsName, requirementsVersion) REFERENCES toscarequirements (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0980-FK_ToscaNodeType_requirementsName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscanodetype ADD CONSTRAINT FK_ToscaNodeType_requirementsName FOREIGN KEY (requirementsName, requirementsVersion) REFERENCES toscarequirements (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0990-FK_ToscaServiceTemplate_capabilityTypesName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_capabilityTypesName FOREIGN KEY (capabilityTypesName, capabilityTypesVersion) REFERENCES toscacapabilitytypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 1000-FK_ToscaServiceTemplate_dataTypesName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_dataTypesName FOREIGN KEY (dataTypesName, dataTypesVersion) REFERENCES toscadatatypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 1010-FK_ToscaServiceTemplate_nodeTypesName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_nodeTypesName FOREIGN KEY (nodeTypesName, nodeTypesVersion) REFERENCES toscanodetypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 1020-FK_ToscaServiceTemplate_policyTypesName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_policyTypesName FOREIGN KEY (policyTypesName, policyTypesVersion) REFERENCES toscapolicytypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 1030-FK_ToscaServiceTemplate_relationshipTypesName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT FK_ToscaServiceTemplate_relationshipTypesName FOREIGN KEY (relationshipTypesName, relationshipTypesVersion) REFERENCES toscarelationshiptypes (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 1040-FK_ToscaTopologyTemplate_nodeTemplatesName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscatopologytemplate ADD CONSTRAINT FK_ToscaTopologyTemplate_nodeTemplatesName FOREIGN KEY (nodeTemplatesName, nodeTemplatessVersion) REFERENCES toscanodetemplates (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 1050-FK_ToscaTopologyTemplate_policyName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscatopologytemplate ADD CONSTRAINT FK_ToscaTopologyTemplate_policyName FOREIGN KEY (policyName, policyVersion) REFERENCES toscapolicies (name, version) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | simulator | 2024-04-24 09:37:44,431 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}}, swaggerId=null, toString()=JettyServletServer(name=SDNC simulator, host=0.0.0.0, port=6668, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@70efb718{STARTED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@b70da4c{/,null,AVAILABLE}, connector=SDNC simulator@c5ee75e{HTTP/1.1, (http/1.1)}{0.0.0.0:6668}, jettyThread=Thread[SDNC simulator-6668,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-bf1ec20==org.glassfish.jersey.servlet.ServletContainer@636d8afb{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}})]: pending time is 4809 ms. simulator | 2024-04-24 09:37:44,432 INFO org.onap.policy.models.simulators starting SO simulator simulator | 2024-04-24 09:37:44,435 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SO simulator, host=0.0.0.0, port=6669, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@b7838a9{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@696f0212{/,null,STOPPED}, connector=SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: WAITED-START simulator | 2024-04-24 09:37:44,435 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SO simulator, host=0.0.0.0, port=6669, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@b7838a9{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@696f0212{/,null,STOPPED}, connector=SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING simulator | 2024-04-24 09:37:44,437 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=SO simulator, host=0.0.0.0, port=6669, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@b7838a9{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@696f0212{/,null,STOPPED}, connector=SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669}, jettyThread=Thread[SO simulator-6669,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING simulator | 2024-04-24 09:37:44,438 INFO jetty-11.0.20; built: 2024-01-29T21:04:22.394Z; git: 922f8dc188f7011e60d0361de585fd4ac4d63064; jvm 17.0.10+7-alpine-r0 simulator | 2024-04-24 09:37:44,446 INFO Session workerName=node0 simulator | 2024-04-24 09:37:44,522 INFO Using GSON for REST calls simulator | 2024-04-24 09:37:44,539 INFO Started o.e.j.s.ServletContextHandler@696f0212{/,null,AVAILABLE} simulator | 2024-04-24 09:37:44,543 INFO Started SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669} simulator | 2024-04-24 09:37:44,544 INFO Started Server@b7838a9{STARTING}[11.0.20,sto=0] @2190ms simulator | 2024-04-24 09:37:44,544 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}}, swaggerId=null, toString()=JettyServletServer(name=SO simulator, host=0.0.0.0, port=6669, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@b7838a9{STARTED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@696f0212{/,null,AVAILABLE}, connector=SO simulator@4e858e0a{HTTP/1.1, (http/1.1)}{0.0.0.0:6669}, jettyThread=Thread[SO simulator-6669,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-4e70a728==org.glassfish.jersey.servlet.ServletContainer@3238e994{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}})]: pending time is 4891 ms. simulator | 2024-04-24 09:37:44,545 INFO org.onap.policy.models.simulators starting VFC simulator simulator | 2024-04-24 09:37:44,548 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=VFC simulator, host=0.0.0.0, port=6670, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@f478a81{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@19553973{/,null,STOPPED}, connector=VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: WAITED-START simulator | 2024-04-24 09:37:44,549 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=VFC simulator, host=0.0.0.0, port=6670, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@f478a81{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@19553973{/,null,STOPPED}, connector=VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670}, jettyThread=null, servlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING simulator | 2024-04-24 09:37:44,550 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=null, toString()=JettyServletServer(name=VFC simulator, host=0.0.0.0, port=6670, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@f478a81{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@19553973{/,null,STOPPED}, connector=VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670}, jettyThread=Thread[VFC simulator-6670,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING simulator | 2024-04-24 09:37:44,552 INFO jetty-11.0.20; built: 2024-01-29T21:04:22.394Z; git: 922f8dc188f7011e60d0361de585fd4ac4d63064; jvm 17.0.10+7-alpine-r0 simulator | 2024-04-24 09:37:44,561 INFO Session workerName=node0 simulator | 2024-04-24 09:37:44,642 INFO Using GSON for REST calls simulator | 2024-04-24 09:37:44,661 INFO Started o.e.j.s.ServletContextHandler@19553973{/,null,AVAILABLE} simulator | 2024-04-24 09:37:44,664 INFO Started VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670} simulator | 2024-04-24 09:37:44,665 INFO Started Server@f478a81{STARTING}[11.0.20,sto=0] @2310ms simulator | 2024-04-24 09:37:44,665 INFO JettyJerseyServer [Jerseyservlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}}, swaggerId=null, toString()=JettyServletServer(name=VFC simulator, host=0.0.0.0, port=6670, sniHostCheck=false, user=null, password=null, contextPath=/, jettyServer=Server@f478a81{STARTED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@19553973{/,null,AVAILABLE}, connector=VFC simulator@1c3146bc{HTTP/1.1, (http/1.1)}{0.0.0.0:6670}, jettyThread=Thread[VFC simulator-6670,5,main], servlets={/*=org.glassfish.jersey.servlet.ServletContainer-47a86fbb==org.glassfish.jersey.servlet.ServletContainer@f8e3b478{jsp=null,order=0,inst=true,async=true,src=EMBEDDED:null,STARTED}})]: pending time is 4885 ms. simulator | 2024-04-24 09:37:44,666 INFO org.onap.policy.models.simulators started policy-db-migrator | > upgrade 1060-TscaServiceTemplatetopologyTemplateParentLocalName.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE toscaservicetemplate ADD CONSTRAINT TscaServiceTemplatetopologyTemplateParentLocalName FOREIGN KEY (topologyTemplateParentLocalName, topologyTemplateLocalName, topologyTemplateParentKeyVersion, topologyTemplateParentKeyName) REFERENCES toscatopologytemplate (parentLocalName, localName, parentKeyVersion, parentKeyName) ON UPDATE RESTRICT ON DELETE RESTRICT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0100-pdp.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE pdp ADD COLUMN LASTUPDATE datetime NOT NULL DEFAULT CURRENT_TIMESTAMP AFTER HEALTHY policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0110-idx_tsidx1.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX IDX_TSIDX1 ON pdpstatistics(timeStamp, name, version) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0120-pk_pdpstatistics.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE pdpstatistics DROP PRIMARY KEY policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0130-pdpstatistics.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE pdpstatistics ADD COLUMN POLICYUNDEPLOYCOUNT BIGINT DEFAULT NULL AFTER POLICYEXECUTEDSUCCESSCOUNT, ADD COLUMN POLICYUNDEPLOYFAILCOUNT BIGINT DEFAULT NULL, ADD COLUMN POLICYUNDEPLOYSUCCESSCOUNT BIGINT DEFAULT NULL, ADD COLUMN ID BIGINT NOT NULL policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0140-pk_pdpstatistics.sql policy-db-migrator | -------------- policy-db-migrator | UPDATE pdpstatistics as p JOIN (SELECT name, version, timeStamp, ROW_NUMBER() OVER (ORDER BY timeStamp ASC) AS row_num FROM pdpstatistics GROUP BY name, version, timeStamp) AS t ON (p.name=t.name AND p.version=t.version AND p.timeStamp = t.timeStamp) SET p.id=t.row_num policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE pdpstatistics ADD CONSTRAINT PK_PDPSTATISTICS PRIMARY KEY (ID, name, version) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0150-pdpstatistics.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE pdpstatistics MODIFY COLUMN timeStamp datetime(6) NULL policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0160-jpapdpstatistics_enginestats.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE jpapdpstatistics_enginestats ADD COLUMN ID BIGINT DEFAULT NULL AFTER UPTIME policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0170-jpapdpstatistics_enginestats.sql policy-db-migrator | -------------- policy-db-migrator | UPDATE jpapdpstatistics_enginestats a policy-db-migrator | JOIN pdpstatistics b policy-db-migrator | ON a.name = b.name AND a.version = b.version AND a.timeStamp = b.timeStamp policy-db-migrator | SET a.id = b.id policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0180-jpapdpstatistics_enginestats.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE jpapdpstatistics_enginestats DROP COLUMN timeStamp policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0190-jpapolicyaudit.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS jpapolicyaudit (ACTION INT DEFAULT NULL, PDPGROUP VARCHAR(255) NULL, PDPTYPE VARCHAR(255) NULL, TIMESTAMP datetime DEFAULT NULL, USER VARCHAR(255) NULL, ID BIGINT NOT NULL, name VARCHAR(120) NOT NULL, version VARCHAR(20) NOT NULL, PRIMARY KEY PK_JPAPOLICYAUDIT (ID, name, version)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0200-JpaPolicyAuditIndex_timestamp.sql policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX JpaPolicyAuditIndex_timestamp ON jpapolicyaudit(TIMESTAMP) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0210-sequence.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS sequence (SEQ_NAME VARCHAR(50) NOT NULL, SEQ_COUNT DECIMAL(38) DEFAULT NULL, PRIMARY KEY PK_SEQUENCE (SEQ_NAME)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0220-sequence.sql policy-db-migrator | -------------- policy-db-migrator | INSERT INTO sequence(SEQ_NAME, SEQ_COUNT) VALUES('SEQ_GEN', (SELECT IFNULL(max(id),0) FROM pdpstatistics)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0100-jpatoscapolicy_targets.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE jpatoscapolicy_targets ADD COLUMN toscaPolicyName VARCHAR(120) NOT NULL, ADD COLUMN toscaPolicyVersion VARCHAR(20) NOT NULL, ADD CONSTRAINT PK_JPATOSCAPOLICY_TARGETS PRIMARY KEY (toscaPolicyName, toscaPolicyVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0110-jpatoscapolicytype_targets.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE jpatoscapolicytype_targets ADD COLUMN toscaPolicyTypeName VARCHAR(120) NOT NULL, ADD COLUMN toscaPolicyTypeVersion VARCHAR(20) NOT NULL, ADD CONSTRAINT PK_JPATOSCAPOLICYTYPE_TARGETS PRIMARY KEY (toscaPolicyTypeName, toscaPolicyTypeVersion) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0120-toscatrigger.sql policy-db-migrator | -------------- policy-db-migrator | DROP TABLE IF EXISTS toscatrigger policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0130-jpatoscapolicytype_triggers.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE jpatoscapolicytype_triggers MODIFY COLUMN triggers LONGBLOB policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0140-toscaparameter.sql policy-db-migrator | -------------- policy-db-migrator | DROP TABLE IF EXISTS toscaparameter policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0150-toscaproperty.sql policy-db-migrator | -------------- policy-db-migrator | DROP TABLE IF EXISTS jpatoscaproperty_constraints policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-db-migrator | DROP TABLE IF EXISTS jpatoscaproperty_metadata policy-pap | Waiting for mariadb port 3306... policy-pap | mariadb (172.17.0.4:3306) open policy-pap | Waiting for kafka port 9092... policy-pap | kafka (172.17.0.9:9092) open policy-pap | Waiting for api port 6969... policy-pap | api (172.17.0.8:6969) open policy-pap | Policy pap config file: /opt/app/policy/pap/etc/papParameters.yaml policy-pap | PDP group configuration file: /opt/app/policy/pap/etc/mounted/groups.json policy-pap | policy-pap | . ____ _ __ _ _ policy-pap | /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \ policy-pap | ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \ policy-pap | \\/ ___)| |_)| | | | | || (_| | ) ) ) ) policy-pap | ' |____| .__|_| |_|_| |_\__, | / / / / policy-pap | =========|_|==============|___/=/_/_/_/ policy-pap | :: Spring Boot :: (v3.1.10) policy-pap | policy-pap | [2024-04-24T09:38:08.425+00:00|INFO|Version|background-preinit] HV000001: Hibernate Validator 8.0.1.Final policy-pap | [2024-04-24T09:38:08.495+00:00|INFO|PolicyPapApplication|main] Starting PolicyPapApplication using Java 17.0.10 with PID 33 (/app/pap.jar started by policy in /opt/app/policy/pap/bin) policy-pap | [2024-04-24T09:38:08.497+00:00|INFO|PolicyPapApplication|main] No active profile set, falling back to 1 default profile: "default" policy-pap | [2024-04-24T09:38:10.616+00:00|INFO|RepositoryConfigurationDelegate|main] Bootstrapping Spring Data JPA repositories in DEFAULT mode. policy-pap | [2024-04-24T09:38:10.740+00:00|INFO|RepositoryConfigurationDelegate|main] Finished Spring Data repository scanning in 113 ms. Found 7 JPA repository interfaces. policy-pap | [2024-04-24T09:38:11.249+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.exception.ServiceExceptionHandler policy-pap | [2024-04-24T09:38:11.250+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.exception.ServiceExceptionHandler policy-pap | [2024-04-24T09:38:11.948+00:00|INFO|TomcatWebServer|main] Tomcat initialized with port(s): 6969 (http) policy-pap | [2024-04-24T09:38:11.960+00:00|INFO|Http11NioProtocol|main] Initializing ProtocolHandler ["http-nio-6969"] policy-pap | [2024-04-24T09:38:11.962+00:00|INFO|StandardService|main] Starting service [Tomcat] policy-pap | [2024-04-24T09:38:11.962+00:00|INFO|StandardEngine|main] Starting Servlet engine: [Apache Tomcat/10.1.19] policy-pap | [2024-04-24T09:38:12.075+00:00|INFO|[/policy/pap/v1]|main] Initializing Spring embedded WebApplicationContext policy-pap | [2024-04-24T09:38:12.075+00:00|INFO|ServletWebServerApplicationContext|main] Root WebApplicationContext: initialization completed in 3498 ms policy-pap | [2024-04-24T09:38:12.521+00:00|INFO|LogHelper|main] HHH000204: Processing PersistenceUnitInfo [name: default] policy-pap | [2024-04-24T09:38:12.592+00:00|INFO|Version|main] HHH000412: Hibernate ORM core version 5.6.15.Final policy-pap | [2024-04-24T09:38:12.971+00:00|INFO|HikariDataSource|main] HikariPool-1 - Starting... policy-pap | [2024-04-24T09:38:13.087+00:00|INFO|HikariPool|main] HikariPool-1 - Added connection org.mariadb.jdbc.Connection@1e0895f5 policy-pap | [2024-04-24T09:38:13.089+00:00|INFO|HikariDataSource|main] HikariPool-1 - Start completed. policy-pap | [2024-04-24T09:38:13.119+00:00|INFO|Dialect|main] HHH000400: Using dialect: org.hibernate.dialect.MariaDB106Dialect policy-pap | [2024-04-24T09:38:14.705+00:00|INFO|JtaPlatformInitiator|main] HHH000490: Using JtaPlatform implementation: [org.hibernate.engine.transaction.jta.platform.internal.NoJtaPlatform] policy-pap | [2024-04-24T09:38:14.715+00:00|INFO|LocalContainerEntityManagerFactoryBean|main] Initialized JPA EntityManagerFactory for persistence unit 'default' policy-pap | [2024-04-24T09:38:15.252+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.repository.PdpGroupRepository policy-pap | [2024-04-24T09:38:15.633+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.repository.PolicyStatusRepository policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-apex-pdp | connections.max.idle.ms = 540000 policy-apex-pdp | default.api.timeout.ms = 60000 policy-apex-pdp | enable.auto.commit = true policy-apex-pdp | exclude.internal.topics = true policy-apex-pdp | fetch.max.bytes = 52428800 policy-apex-pdp | fetch.max.wait.ms = 500 policy-apex-pdp | fetch.min.bytes = 1 policy-apex-pdp | group.id = ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0 policy-apex-pdp | group.instance.id = null policy-apex-pdp | heartbeat.interval.ms = 3000 policy-apex-pdp | interceptor.classes = [] policy-apex-pdp | internal.leave.group.on.close = true policy-apex-pdp | internal.throw.on.fetch.stable.offset.unsupported = false policy-apex-pdp | isolation.level = read_uncommitted policy-apex-pdp | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-apex-pdp | max.partition.fetch.bytes = 1048576 policy-apex-pdp | max.poll.interval.ms = 300000 policy-apex-pdp | max.poll.records = 500 policy-apex-pdp | metadata.max.age.ms = 300000 policy-apex-pdp | metric.reporters = [] policy-apex-pdp | metrics.num.samples = 2 policy-apex-pdp | metrics.recording.level = INFO policy-apex-pdp | metrics.sample.window.ms = 30000 policy-apex-pdp | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-apex-pdp | receive.buffer.bytes = 65536 policy-apex-pdp | reconnect.backoff.max.ms = 1000 policy-apex-pdp | reconnect.backoff.ms = 50 policy-apex-pdp | request.timeout.ms = 30000 policy-apex-pdp | retry.backoff.ms = 100 policy-apex-pdp | sasl.client.callback.handler.class = null policy-apex-pdp | sasl.jaas.config = null policy-apex-pdp | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-apex-pdp | sasl.kerberos.min.time.before.relogin = 60000 policy-apex-pdp | sasl.kerberos.service.name = null policy-apex-pdp | sasl.kerberos.ticket.renew.jitter = 0.05 policy-apex-pdp | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-apex-pdp | sasl.login.callback.handler.class = null policy-apex-pdp | sasl.login.class = null policy-apex-pdp | sasl.login.connect.timeout.ms = null policy-apex-pdp | sasl.login.read.timeout.ms = null policy-apex-pdp | sasl.login.refresh.buffer.seconds = 300 policy-apex-pdp | sasl.login.refresh.min.period.seconds = 60 policy-apex-pdp | sasl.login.refresh.window.factor = 0.8 policy-apex-pdp | sasl.login.refresh.window.jitter = 0.05 policy-apex-pdp | sasl.login.retry.backoff.max.ms = 10000 policy-apex-pdp | sasl.login.retry.backoff.ms = 100 policy-apex-pdp | sasl.mechanism = GSSAPI policy-apex-pdp | sasl.oauthbearer.clock.skew.seconds = 30 policy-apex-pdp | sasl.oauthbearer.expected.audience = null policy-apex-pdp | sasl.oauthbearer.expected.issuer = null policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.url = null policy-apex-pdp | sasl.oauthbearer.scope.claim.name = scope policy-apex-pdp | sasl.oauthbearer.sub.claim.name = sub policy-apex-pdp | sasl.oauthbearer.token.endpoint.url = null policy-apex-pdp | security.protocol = PLAINTEXT policy-apex-pdp | security.providers = null policy-apex-pdp | send.buffer.bytes = 131072 policy-apex-pdp | session.timeout.ms = 45000 policy-apex-pdp | socket.connection.setup.timeout.max.ms = 30000 policy-apex-pdp | socket.connection.setup.timeout.ms = 10000 policy-apex-pdp | ssl.cipher.suites = null policy-apex-pdp | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-apex-pdp | ssl.endpoint.identification.algorithm = https policy-apex-pdp | ssl.engine.factory.class = null policy-apex-pdp | ssl.key.password = null policy-apex-pdp | ssl.keymanager.algorithm = SunX509 policy-apex-pdp | ssl.keystore.certificate.chain = null policy-apex-pdp | ssl.keystore.key = null policy-apex-pdp | ssl.keystore.location = null policy-apex-pdp | ssl.keystore.password = null policy-apex-pdp | ssl.keystore.type = JKS policy-apex-pdp | ssl.protocol = TLSv1.3 policy-apex-pdp | ssl.provider = null policy-apex-pdp | ssl.secure.random.implementation = null policy-apex-pdp | ssl.trustmanager.algorithm = PKIX policy-apex-pdp | ssl.truststore.certificates = null policy-apex-pdp | ssl.truststore.location = null policy-apex-pdp | ssl.truststore.password = null policy-apex-pdp | ssl.truststore.type = JKS policy-apex-pdp | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-apex-pdp | policy-apex-pdp | [2024-04-24T09:38:19.824+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 policy-apex-pdp | [2024-04-24T09:38:19.825+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 policy-apex-pdp | [2024-04-24T09:38:19.825+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1713951499824 policy-apex-pdp | [2024-04-24T09:38:19.825+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Subscribed to topic(s): policy-pdp-pap policy-apex-pdp | [2024-04-24T09:38:19.826+00:00|INFO|InlineBusTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=2a001ab9-b41c-4b40-b876-293a52215e8d, alive=false, publisher=null]]: starting policy-apex-pdp | [2024-04-24T09:38:19.838+00:00|INFO|ProducerConfig|main] ProducerConfig values: policy-apex-pdp | acks = -1 policy-apex-pdp | auto.include.jmx.reporter = true policy-apex-pdp | batch.size = 16384 policy-apex-pdp | bootstrap.servers = [kafka:9092] policy-apex-pdp | buffer.memory = 33554432 policy-apex-pdp | client.dns.lookup = use_all_dns_ips policy-apex-pdp | client.id = producer-1 policy-apex-pdp | compression.type = none policy-apex-pdp | connections.max.idle.ms = 540000 policy-apex-pdp | delivery.timeout.ms = 120000 policy-apex-pdp | enable.idempotence = true policy-apex-pdp | interceptor.classes = [] policy-apex-pdp | key.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-apex-pdp | linger.ms = 0 policy-apex-pdp | max.block.ms = 60000 policy-apex-pdp | max.in.flight.requests.per.connection = 5 policy-apex-pdp | max.request.size = 1048576 policy-db-migrator | DROP TABLE IF EXISTS toscaproperty policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0160-jpapolicyaudit_pk.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE jpapolicyaudit DROP PRIMARY KEY policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE jpapolicyaudit ADD CONSTRAINT PK_JPAPOLICYAUDIT PRIMARY KEY (ID) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0170-pdpstatistics_pk.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE pdpstatistics DROP PRIMARY KEY policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE pdpstatistics ADD CONSTRAINT PK_PDPSTATISTICS PRIMARY KEY (ID) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0180-jpatoscanodetemplate_metadata.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE jpatoscanodetemplate_metadata MODIFY COLUMN METADATA LONGTEXT policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0100-upgrade.sql policy-db-migrator | -------------- policy-db-migrator | select 'upgrade to 1100 completed' as msg policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | msg policy-db-migrator | upgrade to 1100 completed policy-db-migrator | policy-db-migrator | > upgrade 0100-jpapolicyaudit_renameuser.sql policy-db-migrator | -------------- policy-db-migrator | ALTER TABLE jpapolicyaudit RENAME COLUMN USER TO USERNAME policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0110-idx_tsidx1.sql policy-db-migrator | -------------- policy-db-migrator | DROP INDEX IDX_TSIDX1 ON pdpstatistics policy-apex-pdp | metadata.max.age.ms = 300000 policy-apex-pdp | metadata.max.idle.ms = 300000 policy-apex-pdp | metric.reporters = [] policy-apex-pdp | metrics.num.samples = 2 policy-apex-pdp | metrics.recording.level = INFO policy-apex-pdp | metrics.sample.window.ms = 30000 policy-apex-pdp | partitioner.adaptive.partitioning.enable = true policy-apex-pdp | partitioner.availability.timeout.ms = 0 policy-apex-pdp | partitioner.class = null policy-apex-pdp | partitioner.ignore.keys = false policy-apex-pdp | receive.buffer.bytes = 32768 policy-apex-pdp | reconnect.backoff.max.ms = 1000 policy-apex-pdp | reconnect.backoff.ms = 50 policy-apex-pdp | request.timeout.ms = 30000 policy-apex-pdp | retries = 2147483647 policy-apex-pdp | retry.backoff.ms = 100 policy-apex-pdp | sasl.client.callback.handler.class = null policy-apex-pdp | sasl.jaas.config = null policy-apex-pdp | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-apex-pdp | sasl.kerberos.min.time.before.relogin = 60000 policy-apex-pdp | sasl.kerberos.service.name = null policy-apex-pdp | sasl.kerberos.ticket.renew.jitter = 0.05 policy-apex-pdp | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-apex-pdp | sasl.login.callback.handler.class = null policy-apex-pdp | sasl.login.class = null policy-apex-pdp | sasl.login.connect.timeout.ms = null policy-apex-pdp | sasl.login.read.timeout.ms = null policy-apex-pdp | sasl.login.refresh.buffer.seconds = 300 policy-apex-pdp | sasl.login.refresh.min.period.seconds = 60 policy-apex-pdp | sasl.login.refresh.window.factor = 0.8 policy-apex-pdp | sasl.login.refresh.window.jitter = 0.05 policy-apex-pdp | sasl.login.retry.backoff.max.ms = 10000 policy-apex-pdp | sasl.login.retry.backoff.ms = 100 policy-apex-pdp | sasl.mechanism = GSSAPI policy-apex-pdp | sasl.oauthbearer.clock.skew.seconds = 30 policy-apex-pdp | sasl.oauthbearer.expected.audience = null policy-apex-pdp | sasl.oauthbearer.expected.issuer = null policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-apex-pdp | sasl.oauthbearer.jwks.endpoint.url = null policy-apex-pdp | sasl.oauthbearer.scope.claim.name = scope policy-apex-pdp | sasl.oauthbearer.sub.claim.name = sub policy-apex-pdp | sasl.oauthbearer.token.endpoint.url = null kafka | socket.connection.setup.timeout.max.ms = 30000 kafka | socket.connection.setup.timeout.ms = 10000 kafka | socket.listen.backlog.size = 50 kafka | socket.receive.buffer.bytes = 102400 kafka | socket.request.max.bytes = 104857600 kafka | socket.send.buffer.bytes = 102400 kafka | ssl.cipher.suites = [] kafka | ssl.client.auth = none kafka | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] kafka | ssl.endpoint.identification.algorithm = https kafka | ssl.engine.factory.class = null kafka | ssl.key.password = null kafka | ssl.keymanager.algorithm = SunX509 kafka | ssl.keystore.certificate.chain = null kafka | ssl.keystore.key = null kafka | ssl.keystore.location = null kafka | ssl.keystore.password = null kafka | ssl.keystore.type = JKS kafka | ssl.principal.mapping.rules = DEFAULT kafka | ssl.protocol = TLSv1.3 kafka | ssl.provider = null kafka | ssl.secure.random.implementation = null kafka | ssl.trustmanager.algorithm = PKIX kafka | ssl.truststore.certificates = null kafka | ssl.truststore.location = null kafka | ssl.truststore.password = null kafka | ssl.truststore.type = JKS kafka | transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000 kafka | transaction.max.timeout.ms = 900000 kafka | transaction.partition.verification.enable = true kafka | transaction.remove.expired.transaction.cleanup.interval.ms = 3600000 kafka | transaction.state.log.load.buffer.size = 5242880 kafka | transaction.state.log.min.isr = 2 kafka | transaction.state.log.num.partitions = 50 kafka | transaction.state.log.replication.factor = 3 kafka | transaction.state.log.segment.bytes = 104857600 kafka | transactional.id.expiration.ms = 604800000 kafka | unclean.leader.election.enable = false kafka | unstable.api.versions.enable = false kafka | zookeeper.clientCnxnSocket = null kafka | zookeeper.connect = zookeeper:2181 kafka | zookeeper.connection.timeout.ms = null kafka | zookeeper.max.in.flight.requests = 10 kafka | zookeeper.metadata.migration.enable = false kafka | zookeeper.metadata.migration.min.batch.size = 200 kafka | zookeeper.session.timeout.ms = 18000 kafka | zookeeper.set.acl = false kafka | zookeeper.ssl.cipher.suites = null kafka | zookeeper.ssl.client.enable = false kafka | zookeeper.ssl.crl.enable = false kafka | zookeeper.ssl.enabled.protocols = null kafka | zookeeper.ssl.endpoint.identification.algorithm = HTTPS kafka | zookeeper.ssl.keystore.location = null kafka | zookeeper.ssl.keystore.password = null kafka | zookeeper.ssl.keystore.type = null kafka | zookeeper.ssl.ocsp.enable = false kafka | zookeeper.ssl.protocol = TLSv1.2 kafka | zookeeper.ssl.truststore.location = null kafka | zookeeper.ssl.truststore.password = null kafka | zookeeper.ssl.truststore.type = null kafka | (kafka.server.KafkaConfig) kafka | [2024-04-24 09:37:49,791] INFO [ThrottledChannelReaper-Fetch]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) kafka | [2024-04-24 09:37:49,792] INFO [ThrottledChannelReaper-Produce]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) kafka | [2024-04-24 09:37:49,797] INFO [ThrottledChannelReaper-Request]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) kafka | [2024-04-24 09:37:49,799] INFO [ThrottledChannelReaper-ControllerMutation]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) kafka | [2024-04-24 09:37:49,836] INFO Loading logs from log dirs ArraySeq(/var/lib/kafka/data) (kafka.log.LogManager) kafka | [2024-04-24 09:37:49,841] INFO No logs found to be loaded in /var/lib/kafka/data (kafka.log.LogManager) kafka | [2024-04-24 09:37:49,851] INFO Loaded 0 logs in 15ms (kafka.log.LogManager) kafka | [2024-04-24 09:37:49,853] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager) kafka | [2024-04-24 09:37:49,854] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager) kafka | [2024-04-24 09:37:49,867] INFO Starting the log cleaner (kafka.log.LogCleaner) kafka | [2024-04-24 09:37:49,926] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner$CleanerThread) kafka | [2024-04-24 09:37:49,942] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) kafka | [2024-04-24 09:37:49,956] INFO Feature ZK node at path: /feature does not exist (kafka.server.FinalizedFeatureChangeListener) kafka | [2024-04-24 09:37:50,025] INFO [zk-broker-1-to-controller-forwarding-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) kafka | [2024-04-24 09:37:50,387] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) policy-apex-pdp | security.protocol = PLAINTEXT policy-apex-pdp | security.providers = null policy-apex-pdp | send.buffer.bytes = 131072 policy-apex-pdp | socket.connection.setup.timeout.max.ms = 30000 policy-apex-pdp | socket.connection.setup.timeout.ms = 10000 policy-apex-pdp | ssl.cipher.suites = null policy-apex-pdp | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-apex-pdp | ssl.endpoint.identification.algorithm = https policy-apex-pdp | ssl.engine.factory.class = null policy-apex-pdp | ssl.key.password = null policy-apex-pdp | ssl.keymanager.algorithm = SunX509 policy-apex-pdp | ssl.keystore.certificate.chain = null policy-apex-pdp | ssl.keystore.key = null policy-apex-pdp | ssl.keystore.location = null policy-apex-pdp | ssl.keystore.password = null policy-apex-pdp | ssl.keystore.type = JKS policy-apex-pdp | ssl.protocol = TLSv1.3 policy-apex-pdp | ssl.provider = null policy-apex-pdp | ssl.secure.random.implementation = null policy-apex-pdp | ssl.trustmanager.algorithm = PKIX policy-apex-pdp | ssl.truststore.certificates = null policy-apex-pdp | ssl.truststore.location = null policy-apex-pdp | ssl.truststore.password = null policy-apex-pdp | ssl.truststore.type = JKS policy-apex-pdp | transaction.timeout.ms = 60000 policy-apex-pdp | transactional.id = null policy-apex-pdp | value.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-apex-pdp | policy-apex-pdp | [2024-04-24T09:38:19.848+00:00|INFO|KafkaProducer|main] [Producer clientId=producer-1] Instantiated an idempotent producer. policy-apex-pdp | [2024-04-24T09:38:19.870+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 policy-apex-pdp | [2024-04-24T09:38:19.870+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 policy-apex-pdp | [2024-04-24T09:38:19.870+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1713951499869 policy-apex-pdp | [2024-04-24T09:38:19.870+00:00|INFO|InlineKafkaTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=2a001ab9-b41c-4b40-b876-293a52215e8d, alive=false, publisher=KafkaPublisherWrapper []]]: KAFKA SINK created policy-apex-pdp | [2024-04-24T09:38:19.870+00:00|INFO|ServiceManager|main] service manager starting set alive policy-apex-pdp | [2024-04-24T09:38:19.870+00:00|INFO|ServiceManager|main] service manager starting register pdp status context object policy-apex-pdp | [2024-04-24T09:38:19.877+00:00|INFO|ServiceManager|main] service manager starting topic sinks policy-apex-pdp | [2024-04-24T09:38:19.877+00:00|INFO|ServiceManager|main] service manager starting Pdp Status publisher policy-apex-pdp | [2024-04-24T09:38:19.878+00:00|INFO|ServiceManager|main] service manager starting Register pdp update listener policy-apex-pdp | [2024-04-24T09:38:19.878+00:00|INFO|ServiceManager|main] service manager starting Register pdp state change request dispatcher policy-apex-pdp | [2024-04-24T09:38:19.879+00:00|INFO|ServiceManager|main] service manager starting Message Dispatcher policy-apex-pdp | [2024-04-24T09:38:19.879+00:00|INFO|TopicBase|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0, consumerInstance=policy-apex-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: registering org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher@607fbe09 policy-apex-pdp | [2024-04-24T09:38:19.879+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0, consumerInstance=policy-apex-pdp, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: register: start not attempted policy-apex-pdp | [2024-04-24T09:38:19.879+00:00|INFO|ServiceManager|main] service manager starting Create REST server policy-apex-pdp | [2024-04-24T09:38:19.902+00:00|INFO|OrderedServiceImpl|Timer-0] ***** OrderedServiceImpl implementers: policy-apex-pdp | [] policy-apex-pdp | [2024-04-24T09:38:19.905+00:00|INFO|network|Timer-0] [OUT|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"38eb7d88-d4ed-47c7-87bc-621c6fee4177","timestampMs":1713951499889,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup"} policy-apex-pdp | [2024-04-24T09:38:20.095+00:00|INFO|ServiceManager|main] service manager starting Rest Server policy-apex-pdp | [2024-04-24T09:38:20.095+00:00|INFO|ServiceManager|main] service manager starting policy-apex-pdp | [2024-04-24T09:38:20.095+00:00|INFO|ServiceManager|main] service manager starting REST RestServerParameters policy-apex-pdp | [2024-04-24T09:38:20.095+00:00|INFO|JettyServletServer|main] JettyJerseyServer [Jerseyservlets={/metrics=io.prometheus.client.servlet.jakarta.exporter.MetricsServlet-21694e53==io.prometheus.client.servlet.jakarta.exporter.MetricsServlet@2326051b{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:null,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-46074492==org.glassfish.jersey.servlet.ServletContainer@705041b0{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=swagger-6969, toString()=JettyServletServer(name=RestServerParameters, host=0.0.0.0, port=6969, sniHostCheck=false, user=policyadmin, password=zb!XztG34, contextPath=/, jettyServer=Server@5aabbb29{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@72c927f1{/,null,STOPPED}, connector=RestServerParameters@53ab0286{HTTP/1.1, (http/1.1)}{0.0.0.0:6969}, jettyThread=null, servlets={/metrics=io.prometheus.client.servlet.jakarta.exporter.MetricsServlet-21694e53==io.prometheus.client.servlet.jakarta.exporter.MetricsServlet@2326051b{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:null,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-46074492==org.glassfish.jersey.servlet.ServletContainer@705041b0{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING policy-apex-pdp | [2024-04-24T09:38:20.106+00:00|INFO|ServiceManager|main] service manager started policy-apex-pdp | [2024-04-24T09:38:20.106+00:00|INFO|ServiceManager|main] service manager started policy-apex-pdp | [2024-04-24T09:38:20.107+00:00|INFO|ApexStarterMain|main] Started policy-apex-pdp service successfully. policy-apex-pdp | [2024-04-24T09:38:20.107+00:00|INFO|JettyServletServer|RestServerParameters-6969] JettyJerseyServer [Jerseyservlets={/metrics=io.prometheus.client.servlet.jakarta.exporter.MetricsServlet-21694e53==io.prometheus.client.servlet.jakarta.exporter.MetricsServlet@2326051b{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:null,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-46074492==org.glassfish.jersey.servlet.ServletContainer@705041b0{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}}, swaggerId=swagger-6969, toString()=JettyServletServer(name=RestServerParameters, host=0.0.0.0, port=6969, sniHostCheck=false, user=policyadmin, password=zb!XztG34, contextPath=/, jettyServer=Server@5aabbb29{STOPPED}[11.0.20,sto=0], context=o.e.j.s.ServletContextHandler@72c927f1{/,null,STOPPED}, connector=RestServerParameters@53ab0286{HTTP/1.1, (http/1.1)}{0.0.0.0:6969}, jettyThread=Thread[RestServerParameters-6969,5,main], servlets={/metrics=io.prometheus.client.servlet.jakarta.exporter.MetricsServlet-21694e53==io.prometheus.client.servlet.jakarta.exporter.MetricsServlet@2326051b{jsp=null,order=-1,inst=false,async=true,src=EMBEDDED:null,STOPPED}, /*=org.glassfish.jersey.servlet.ServletContainer-46074492==org.glassfish.jersey.servlet.ServletContainer@705041b0{jsp=null,order=0,inst=false,async=true,src=EMBEDDED:null,STOPPED}})]: STARTING policy-apex-pdp | [2024-04-24T09:38:20.248+00:00|INFO|Metadata|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] Cluster ID: 8oQQmVC5RD2DltlWw-KT5A policy-apex-pdp | [2024-04-24T09:38:20.248+00:00|INFO|Metadata|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Cluster ID: 8oQQmVC5RD2DltlWw-KT5A policy-apex-pdp | [2024-04-24T09:38:20.249+00:00|INFO|TransactionManager|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] ProducerId set to 2 with epoch 0 policy-apex-pdp | [2024-04-24T09:38:20.251+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Discovered group coordinator kafka:9092 (id: 2147483646 rack: null) policy-apex-pdp | [2024-04-24T09:38:20.257+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] (Re-)joining group policy-apex-pdp | [2024-04-24T09:38:20.272+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Request joining group due to: need to re-join with the given member-id: consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2-dd69be57-ac85-4787-b12c-453fd9504f7c policy-apex-pdp | [2024-04-24T09:38:20.272+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) policy-apex-pdp | [2024-04-24T09:38:20.272+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] (Re-)joining group policy-apex-pdp | [2024-04-24T09:38:20.774+00:00|INFO|YamlMessageBodyHandler|RestServerParameters-6969] Accepting YAML for REST calls policy-apex-pdp | [2024-04-24T09:38:20.774+00:00|INFO|GsonMessageBodyHandler|RestServerParameters-6969] Using GSON for REST calls policy-apex-pdp | [2024-04-24T09:38:23.280+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Successfully joined group with generation Generation{generationId=1, memberId='consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2-dd69be57-ac85-4787-b12c-453fd9504f7c', protocol='range'} policy-apex-pdp | [2024-04-24T09:38:23.287+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Finished assignment for group at generation 1: {consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2-dd69be57-ac85-4787-b12c-453fd9504f7c=Assignment(partitions=[policy-pdp-pap-0])} policy-apex-pdp | [2024-04-24T09:38:23.296+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Successfully synced group in generation Generation{generationId=1, memberId='consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2-dd69be57-ac85-4787-b12c-453fd9504f7c', protocol='range'} policy-apex-pdp | [2024-04-24T09:38:23.297+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Notifying assignor about the new Assignment(partitions=[policy-pdp-pap-0]) policy-apex-pdp | [2024-04-24T09:38:23.298+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Adding newly assigned partitions: policy-pdp-pap-0 policy-apex-pdp | [2024-04-24T09:38:23.307+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Found no committed offset for partition policy-pdp-pap-0 policy-apex-pdp | [2024-04-24T09:38:23.317+00:00|INFO|SubscriptionState|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2, groupId=ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0] Resetting offset for partition policy-pdp-pap-0 to position FetchPosition{offset=1, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:9092 (id: 1 rack: null)], epoch=0}}. policy-apex-pdp | [2024-04-24T09:38:39.878+00:00|INFO|network|Timer-0] [OUT|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"3e991c30-f7a7-4f78-960f-7e0ae2fecbef","timestampMs":1713951519878,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup"} policy-apex-pdp | [2024-04-24T09:38:39.908+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"3e991c30-f7a7-4f78-960f-7e0ae2fecbef","timestampMs":1713951519878,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup"} policy-apex-pdp | [2024-04-24T09:38:39.911+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-apex-pdp | [2024-04-24T09:38:40.049+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-apex-pdp | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"messageName":"PDP_UPDATE","requestId":"594a6c37-7874-45c6-b60d-69cb6ee9534e","timestampMs":1713951519988,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-apex-pdp | [2024-04-24T09:38:40.061+00:00|WARN|Registry|KAFKA-source-policy-pdp-pap] replacing previously registered: object:pdp/status/publisher policy-apex-pdp | [2024-04-24T09:38:40.062+00:00|INFO|network|Timer-1] [OUT|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"84be2407-9b5d-4edc-b9b9-3bb4f9743c8b","timestampMs":1713951520061,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup"} policy-apex-pdp | [2024-04-24T09:38:40.062+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"594a6c37-7874-45c6-b60d-69cb6ee9534e","responseStatus":"SUCCESS","responseMessage":"Pdp update successful."},"messageName":"PDP_STATUS","requestId":"77ab3928-ad7a-417a-b373-05e2442a45b1","timestampMs":1713951520062,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-apex-pdp | [2024-04-24T09:38:40.076+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"84be2407-9b5d-4edc-b9b9-3bb4f9743c8b","timestampMs":1713951520061,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup"} policy-apex-pdp | [2024-04-24T09:38:40.077+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-apex-pdp | [2024-04-24T09:38:40.077+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"594a6c37-7874-45c6-b60d-69cb6ee9534e","responseStatus":"SUCCESS","responseMessage":"Pdp update successful."},"messageName":"PDP_STATUS","requestId":"77ab3928-ad7a-417a-b373-05e2442a45b1","timestampMs":1713951520062,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-apex-pdp | [2024-04-24T09:38:40.078+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-apex-pdp | [2024-04-24T09:38:40.095+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-apex-pdp | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"0e043121-ee0e-429d-bc7d-b3a900f421a3","timestampMs":1713951519989,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-apex-pdp | [2024-04-24T09:38:40.098+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpStateChange","policies":[],"response":{"responseTo":"0e043121-ee0e-429d-bc7d-b3a900f421a3","responseStatus":"SUCCESS","responseMessage":"State changed to active. No policies found."},"messageName":"PDP_STATUS","requestId":"b4698cf3-0a26-43ef-845c-f0142a379e14","timestampMs":1713951520098,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-apex-pdp | [2024-04-24T09:38:40.105+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpStateChange","policies":[],"response":{"responseTo":"0e043121-ee0e-429d-bc7d-b3a900f421a3","responseStatus":"SUCCESS","responseMessage":"State changed to active. No policies found."},"messageName":"PDP_STATUS","requestId":"b4698cf3-0a26-43ef-845c-f0142a379e14","timestampMs":1713951520098,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-apex-pdp | [2024-04-24T09:38:40.106+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-apex-pdp | [2024-04-24T09:38:40.147+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-apex-pdp | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"0a475f2e-48b4-4ba1-aa36-217d754d9ba8","timestampMs":1713951520109,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-apex-pdp | [2024-04-24T09:38:40.149+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [OUT|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"0a475f2e-48b4-4ba1-aa36-217d754d9ba8","responseStatus":"SUCCESS","responseMessage":"Pdp already updated"},"messageName":"PDP_STATUS","requestId":"a21ae420-dcab-420b-8924-c4ec1683dfda","timestampMs":1713951520149,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-apex-pdp | [2024-04-24T09:38:40.157+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-apex-pdp | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"0a475f2e-48b4-4ba1-aa36-217d754d9ba8","responseStatus":"SUCCESS","responseMessage":"Pdp already updated"},"messageName":"PDP_STATUS","requestId":"a21ae420-dcab-420b-8924-c4ec1683dfda","timestampMs":1713951520149,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-apex-pdp | [2024-04-24T09:38:40.158+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATUS policy-apex-pdp | [2024-04-24T09:38:56.175+00:00|INFO|RequestLog|qtp1863100050-33] 172.17.0.2 - policyadmin [24/Apr/2024:09:38:56 +0000] "GET /metrics HTTP/1.1" 200 10646 "-" "Prometheus/2.51.2" policy-apex-pdp | [2024-04-24T09:39:56.083+00:00|INFO|RequestLog|qtp1863100050-28] 172.17.0.2 - policyadmin [24/Apr/2024:09:39:56 +0000] "GET /metrics HTTP/1.1" 200 10650 "-" "Prometheus/2.51.2" kafka | [2024-04-24 09:37:50,407] INFO [SocketServer listenerType=ZK_BROKER, nodeId=1] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer) kafka | [2024-04-24 09:37:50,408] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) kafka | [2024-04-24 09:37:50,414] INFO [SocketServer listenerType=ZK_BROKER, nodeId=1] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT_HOST) (kafka.network.SocketServer) grafana | logger=migrator t=2024-04-24T09:37:38.017860983Z level=info msg="Migration successfully executed" id="drop index UQE_api_key_key - v1" duration=878.514µs grafana | logger=migrator t=2024-04-24T09:37:38.022691212Z level=info msg="Executing migration" id="drop index UQE_api_key_account_id_name - v1" grafana | logger=migrator t=2024-04-24T09:37:38.023872768Z level=info msg="Migration successfully executed" id="drop index UQE_api_key_account_id_name - v1" duration=1.180676ms grafana | logger=migrator t=2024-04-24T09:37:38.02751332Z level=info msg="Executing migration" id="Rename table api_key to api_key_v1 - v1" grafana | logger=migrator t=2024-04-24T09:37:38.034449322Z level=info msg="Migration successfully executed" id="Rename table api_key to api_key_v1 - v1" duration=6.931712ms grafana | logger=migrator t=2024-04-24T09:37:38.039284891Z level=info msg="Executing migration" id="create api_key table v2" grafana | logger=migrator t=2024-04-24T09:37:38.039869654Z level=info msg="Migration successfully executed" id="create api_key table v2" duration=582.943µs grafana | logger=migrator t=2024-04-24T09:37:38.043407802Z level=info msg="Executing migration" id="create index IDX_api_key_org_id - v2" grafana | logger=migrator t=2024-04-24T09:37:38.045279416Z level=info msg="Migration successfully executed" id="create index IDX_api_key_org_id - v2" duration=1.868913ms grafana | logger=migrator t=2024-04-24T09:37:38.048881426Z level=info msg="Executing migration" id="create index UQE_api_key_key - v2" grafana | logger=migrator t=2024-04-24T09:37:38.049671717Z level=info msg="Migration successfully executed" id="create index UQE_api_key_key - v2" duration=790.291µs grafana | logger=migrator t=2024-04-24T09:37:38.052537649Z level=info msg="Executing migration" id="create index UQE_api_key_org_id_name - v2" grafana | logger=migrator t=2024-04-24T09:37:38.053350411Z level=info msg="Migration successfully executed" id="create index UQE_api_key_org_id_name - v2" duration=812.572µs grafana | logger=migrator t=2024-04-24T09:37:38.056336118Z level=info msg="Executing migration" id="copy api_key v1 to v2" grafana | logger=migrator t=2024-04-24T09:37:38.056734233Z level=info msg="Migration successfully executed" id="copy api_key v1 to v2" duration=398.015µs grafana | logger=migrator t=2024-04-24T09:37:38.059883386Z level=info msg="Executing migration" id="Drop old table api_key_v1" grafana | logger=migrator t=2024-04-24T09:37:38.060462029Z level=info msg="Migration successfully executed" id="Drop old table api_key_v1" duration=575.263µs grafana | logger=migrator t=2024-04-24T09:37:38.063466346Z level=info msg="Executing migration" id="Update api_key table charset" grafana | logger=migrator t=2024-04-24T09:37:38.063492147Z level=info msg="Migration successfully executed" id="Update api_key table charset" duration=26.021µs grafana | logger=migrator t=2024-04-24T09:37:38.066490184Z level=info msg="Executing migration" id="Add expires to api_key table" grafana | logger=migrator t=2024-04-24T09:37:38.070517702Z level=info msg="Migration successfully executed" id="Add expires to api_key table" duration=4.026988ms grafana | logger=migrator t=2024-04-24T09:37:38.074335091Z level=info msg="Executing migration" id="Add service account foreign key" grafana | logger=migrator t=2024-04-24T09:37:38.076874121Z level=info msg="Migration successfully executed" id="Add service account foreign key" duration=2.539789ms grafana | logger=migrator t=2024-04-24T09:37:38.080868537Z level=info msg="Executing migration" id="set service account foreign key to nil if 0" grafana | logger=migrator t=2024-04-24T09:37:38.081042764Z level=info msg="Migration successfully executed" id="set service account foreign key to nil if 0" duration=174.377µs grafana | logger=migrator t=2024-04-24T09:37:38.084323062Z level=info msg="Executing migration" id="Add last_used_at to api_key table" grafana | logger=migrator t=2024-04-24T09:37:38.087309399Z level=info msg="Migration successfully executed" id="Add last_used_at to api_key table" duration=2.986517ms grafana | logger=migrator t=2024-04-24T09:37:38.090329517Z level=info msg="Executing migration" id="Add is_revoked column to api_key table" grafana | logger=migrator t=2024-04-24T09:37:38.092926719Z level=info msg="Migration successfully executed" id="Add is_revoked column to api_key table" duration=2.596111ms grafana | logger=migrator t=2024-04-24T09:37:38.095731458Z level=info msg="Executing migration" id="create dashboard_snapshot table v4" grafana | logger=migrator t=2024-04-24T09:37:38.096610702Z level=info msg="Migration successfully executed" id="create dashboard_snapshot table v4" duration=879.394µs grafana | logger=migrator t=2024-04-24T09:37:38.100405551Z level=info msg="Executing migration" id="drop table dashboard_snapshot_v4 #1" grafana | logger=migrator t=2024-04-24T09:37:38.101163901Z level=info msg="Migration successfully executed" id="drop table dashboard_snapshot_v4 #1" duration=754.259µs grafana | logger=migrator t=2024-04-24T09:37:38.105882274Z level=info msg="Executing migration" id="create dashboard_snapshot table v5 #2" grafana | logger=migrator t=2024-04-24T09:37:38.106799981Z level=info msg="Migration successfully executed" id="create dashboard_snapshot table v5 #2" duration=918.537µs grafana | logger=migrator t=2024-04-24T09:37:38.109552969Z level=info msg="Executing migration" id="create index UQE_dashboard_snapshot_key - v5" grafana | logger=migrator t=2024-04-24T09:37:38.110339509Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_snapshot_key - v5" duration=783.12µs grafana | logger=migrator t=2024-04-24T09:37:38.113558515Z level=info msg="Executing migration" id="create index UQE_dashboard_snapshot_delete_key - v5" grafana | logger=migrator t=2024-04-24T09:37:38.114399948Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_snapshot_delete_key - v5" duration=840.483µs grafana | logger=migrator t=2024-04-24T09:37:38.117419476Z level=info msg="Executing migration" id="create index IDX_dashboard_snapshot_user_id - v5" grafana | logger=migrator t=2024-04-24T09:37:38.118199147Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_snapshot_user_id - v5" duration=779.43µs grafana | logger=migrator t=2024-04-24T09:37:38.121455044Z level=info msg="Executing migration" id="alter dashboard_snapshot to mediumtext v2" grafana | logger=migrator t=2024-04-24T09:37:38.121537977Z level=info msg="Migration successfully executed" id="alter dashboard_snapshot to mediumtext v2" duration=84.233µs grafana | logger=migrator t=2024-04-24T09:37:38.128049551Z level=info msg="Executing migration" id="Update dashboard_snapshot table charset" policy-pap | [2024-04-24T09:38:15.732+00:00|WARN|LocalVariableTableParameterNameDiscoverer|main] Using deprecated '-debug' fallback for parameter name resolution. Compile the affected code with '-parameters' instead or avoid its introspection: org.onap.policy.pap.main.repository.PolicyAuditRepository policy-pap | [2024-04-24T09:38:16.006+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-pap | allow.auto.create.topics = true policy-pap | auto.commit.interval.ms = 5000 policy-pap | auto.include.jmx.reporter = true policy-pap | auto.offset.reset = latest policy-pap | bootstrap.servers = [kafka:9092] policy-pap | check.crcs = true policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = consumer-0d71b024-87db-4bab-bad3-947ec269530c-1 policy-pap | client.rack = policy-pap | connections.max.idle.ms = 540000 policy-pap | default.api.timeout.ms = 60000 policy-pap | enable.auto.commit = true policy-pap | exclude.internal.topics = true policy-pap | fetch.max.bytes = 52428800 policy-pap | fetch.max.wait.ms = 500 policy-pap | fetch.min.bytes = 1 policy-pap | group.id = 0d71b024-87db-4bab-bad3-947ec269530c policy-pap | group.instance.id = null policy-pap | heartbeat.interval.ms = 3000 policy-pap | interceptor.classes = [] policy-pap | internal.leave.group.on.close = true policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false policy-pap | isolation.level = read_uncommitted policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | max.partition.fetch.bytes = 1048576 policy-pap | max.poll.interval.ms = 300000 policy-pap | max.poll.records = 500 policy-pap | metadata.max.age.ms = 300000 policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-pap | receive.buffer.bytes = 65536 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-db-migrator | CREATE INDEX IDXTSIDX1 ON pdpstatistics(timeStamp, name, version) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0120-audit_sequence.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS audit_sequence (SEQ_NAME VARCHAR(50) NOT NULL, SEQ_COUNT DECIMAL(38) DEFAULT NULL, PRIMARY KEY PK_SEQUENCE (SEQ_NAME)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-db-migrator | INSERT INTO audit_sequence(SEQ_NAME, SEQ_COUNT) VALUES('SEQ_GEN', (SELECT IFNULL(max(id),0) FROM jpapolicyaudit)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0130-statistics_sequence.sql policy-db-migrator | -------------- policy-db-migrator | CREATE TABLE IF NOT EXISTS statistics_sequence (SEQ_NAME VARCHAR(50) NOT NULL, SEQ_COUNT DECIMAL(38) DEFAULT NULL, PRIMARY KEY PK_SEQUENCE (SEQ_NAME)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-db-migrator | INSERT INTO statistics_sequence(SEQ_NAME, SEQ_COUNT) VALUES('SEQ_GEN', (SELECT IFNULL(max(id),0) FROM pdpstatistics)) policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-db-migrator | TRUNCATE TABLE sequence policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0100-pdpstatistics.sql policy-db-migrator | -------------- policy-db-migrator | DROP INDEX IDXTSIDX1 ON pdpstatistics policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | -------------- policy-db-migrator | DROP TABLE pdpstatistics policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0110-jpapdpstatistics_enginestats.sql policy-db-migrator | -------------- policy-db-migrator | DROP TABLE jpapdpstatistics_enginestats policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policy-db-migrator | > upgrade 0120-statistics_sequence.sql policy-db-migrator | -------------- policy-db-migrator | DROP TABLE statistics_sequence policy-db-migrator | -------------- policy-db-migrator | policy-db-migrator | policyadmin: OK: upgrade (1300) policy-db-migrator | name version policy-db-migrator | policyadmin 1300 policy-db-migrator | ID script operation from_version to_version tag success atTime policy-db-migrator | 1 0100-jpapdpgroup_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:45 policy-db-migrator | 2 0110-jpapdpstatistics_enginestats.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:45 policy-db-migrator | 3 0120-jpapdpsubgroup_policies.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:45 policy-db-migrator | 4 0130-jpapdpsubgroup_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:45 policy-db-migrator | 5 0140-jpapdpsubgroup_supportedpolicytypes.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:45 policy-db-migrator | 6 0150-jpatoscacapabilityassignment_attributes.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 7 0160-jpatoscacapabilityassignment_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 8 0170-jpatoscacapabilityassignment_occurrences.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 9 0180-jpatoscacapabilityassignment_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 10 0190-jpatoscacapabilitytype_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 11 0200-jpatoscacapabilitytype_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 12 0210-jpatoscadatatype_constraints.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 13 0220-jpatoscadatatype_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 14 0230-jpatoscadatatype_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 15 0240-jpatoscanodetemplate_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 16 0250-jpatoscanodetemplate_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 17 0260-jpatoscanodetype_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 18 0270-jpatoscanodetype_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | session.timeout.ms = 45000 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | policy-pap | [2024-04-24T09:38:16.177+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 policy-pap | [2024-04-24T09:38:16.177+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 policy-pap | [2024-04-24T09:38:16.177+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1713951496175 policy-pap | [2024-04-24T09:38:16.180+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-1, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Subscribed to topic(s): policy-pdp-pap policy-pap | [2024-04-24T09:38:16.181+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-pap | allow.auto.create.topics = true policy-pap | auto.commit.interval.ms = 5000 policy-pap | auto.include.jmx.reporter = true policy-pap | auto.offset.reset = latest policy-pap | bootstrap.servers = [kafka:9092] policy-pap | check.crcs = true policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = consumer-policy-pap-2 policy-pap | client.rack = policy-pap | connections.max.idle.ms = 540000 policy-pap | default.api.timeout.ms = 60000 policy-pap | enable.auto.commit = true policy-pap | exclude.internal.topics = true policy-pap | fetch.max.bytes = 52428800 policy-pap | fetch.max.wait.ms = 500 policy-pap | fetch.min.bytes = 1 policy-pap | group.id = policy-pap policy-pap | group.instance.id = null policy-pap | heartbeat.interval.ms = 3000 policy-pap | interceptor.classes = [] policy-pap | internal.leave.group.on.close = true policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false policy-pap | isolation.level = read_uncommitted policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | max.partition.fetch.bytes = 1048576 policy-pap | max.poll.interval.ms = 300000 policy-pap | max.poll.records = 500 policy-pap | metadata.max.age.ms = 300000 policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-pap | receive.buffer.bytes = 65536 policy-pap | reconnect.backoff.max.ms = 1000 policy-db-migrator | 19 0280-jpatoscapolicy_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 20 0290-jpatoscapolicy_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 21 0300-jpatoscapolicy_targets.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 22 0310-jpatoscapolicytype_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 23 0320-jpatoscapolicytype_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:46 policy-db-migrator | 24 0330-jpatoscapolicytype_targets.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 25 0340-jpatoscapolicytype_triggers.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 26 0350-jpatoscaproperty_constraints.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 27 0360-jpatoscaproperty_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 28 0370-jpatoscarelationshiptype_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 29 0380-jpatoscarelationshiptype_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 30 0390-jpatoscarequirement_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 31 0400-jpatoscarequirement_occurrences.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 32 0410-jpatoscarequirement_properties.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 33 0420-jpatoscaservicetemplate_metadata.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 34 0430-jpatoscatopologytemplate_inputs.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 35 0440-pdpgroup_pdpsubgroup.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 36 0450-pdpgroup.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 37 0460-pdppolicystatus.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 38 0470-pdp.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 39 0480-pdpstatistics.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 40 0490-pdpsubgroup_pdp.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 41 0500-pdpsubgroup.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 42 0510-toscacapabilityassignment.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 43 0520-toscacapabilityassignments.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 44 0530-toscacapabilityassignments_toscacapabilityassignment.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 45 0540-toscacapabilitytype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:47 policy-db-migrator | 46 0550-toscacapabilitytypes.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 47 0560-toscacapabilitytypes_toscacapabilitytype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 48 0570-toscadatatype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 49 0580-toscadatatypes.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 50 0590-toscadatatypes_toscadatatype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 51 0600-toscanodetemplate.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 52 0610-toscanodetemplates.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 53 0620-toscanodetemplates_toscanodetemplate.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 54 0630-toscanodetype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 55 0640-toscanodetypes.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 56 0650-toscanodetypes_toscanodetype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 57 0660-toscaparameter.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 58 0670-toscapolicies.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 59 0680-toscapolicies_toscapolicy.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 60 0690-toscapolicy.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 61 0700-toscapolicytype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 62 0710-toscapolicytypes.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 63 0720-toscapolicytypes_toscapolicytype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 64 0730-toscaproperty.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 65 0740-toscarelationshiptype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:48 policy-db-migrator | 66 0750-toscarelationshiptypes.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 67 0760-toscarelationshiptypes_toscarelationshiptype.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 68 0770-toscarequirement.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 69 0780-toscarequirements.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 70 0790-toscarequirements_toscarequirement.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 71 0800-toscaservicetemplate.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 72 0810-toscatopologytemplate.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 73 0820-toscatrigger.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 74 0830-FK_ToscaNodeTemplate_capabilitiesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 75 0840-FK_ToscaNodeTemplate_requirementsName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 76 0850-FK_ToscaNodeType_requirementsName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 77 0860-FK_ToscaServiceTemplate_capabilityTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 78 0870-FK_ToscaServiceTemplate_dataTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 79 0880-FK_ToscaServiceTemplate_nodeTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 80 0890-FK_ToscaServiceTemplate_policyTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 81 0900-FK_ToscaServiceTemplate_relationshipTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 82 0910-FK_ToscaTopologyTemplate_nodeTemplatesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 83 0920-FK_ToscaTopologyTemplate_policyName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 84 0940-PdpPolicyStatus_PdpGroup.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:49 policy-db-migrator | 85 0950-TscaServiceTemplatetopologyTemplateParentLocalName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 86 0960-FK_ToscaNodeTemplate_capabilitiesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 87 0970-FK_ToscaNodeTemplate_requirementsName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 88 0980-FK_ToscaNodeType_requirementsName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 89 0990-FK_ToscaServiceTemplate_capabilityTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 90 1000-FK_ToscaServiceTemplate_dataTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 91 1010-FK_ToscaServiceTemplate_nodeTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 92 1020-FK_ToscaServiceTemplate_policyTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 93 1030-FK_ToscaServiceTemplate_relationshipTypesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 94 1040-FK_ToscaTopologyTemplate_nodeTemplatesName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 95 1050-FK_ToscaTopologyTemplate_policyName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 96 1060-TscaServiceTemplatetopologyTemplateParentLocalName.sql upgrade 0 0800 2404240937450800u 1 2024-04-24 09:37:50 policy-db-migrator | 97 0100-pdp.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:50 policy-db-migrator | 98 0110-idx_tsidx1.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:50 prometheus | ts=2024-04-24T09:37:36.276Z caller=main.go:573 level=info msg="No time or size retention was set so using the default time retention" duration=15d prometheus | ts=2024-04-24T09:37:36.276Z caller=main.go:617 level=info msg="Starting Prometheus Server" mode=server version="(version=2.51.2, branch=HEAD, revision=b4c0ab52c3e9b940ab803581ddae9b3d9a452337)" prometheus | ts=2024-04-24T09:37:36.276Z caller=main.go:622 level=info build_context="(go=go1.22.2, platform=linux/amd64, user=root@b63f02a423d9, date=20240410-14:05:54, tags=netgo,builtinassets,stringlabels)" prometheus | ts=2024-04-24T09:37:36.276Z caller=main.go:623 level=info host_details="(Linux 4.15.0-192-generic #203-Ubuntu SMP Wed Aug 10 17:40:03 UTC 2022 x86_64 prometheus (none))" prometheus | ts=2024-04-24T09:37:36.276Z caller=main.go:624 level=info fd_limits="(soft=1048576, hard=1048576)" prometheus | ts=2024-04-24T09:37:36.276Z caller=main.go:625 level=info vm_limits="(soft=unlimited, hard=unlimited)" prometheus | ts=2024-04-24T09:37:36.279Z caller=web.go:568 level=info component=web msg="Start listening for connections" address=0.0.0.0:9090 prometheus | ts=2024-04-24T09:37:36.279Z caller=main.go:1129 level=info msg="Starting TSDB ..." prometheus | ts=2024-04-24T09:37:36.286Z caller=tls_config.go:313 level=info component=web msg="Listening on" address=[::]:9090 prometheus | ts=2024-04-24T09:37:36.286Z caller=tls_config.go:316 level=info component=web msg="TLS is disabled." http2=false address=[::]:9090 prometheus | ts=2024-04-24T09:37:36.290Z caller=head.go:616 level=info component=tsdb msg="Replaying on-disk memory mappable chunks if any" prometheus | ts=2024-04-24T09:37:36.290Z caller=head.go:698 level=info component=tsdb msg="On-disk memory mappable chunks replay completed" duration=3.85µs prometheus | ts=2024-04-24T09:37:36.290Z caller=head.go:706 level=info component=tsdb msg="Replaying WAL, this may take a while" prometheus | ts=2024-04-24T09:37:36.290Z caller=head.go:778 level=info component=tsdb msg="WAL segment loaded" segment=0 maxSegment=0 prometheus | ts=2024-04-24T09:37:36.290Z caller=head.go:815 level=info component=tsdb msg="WAL replay completed" checkpoint_replay_duration=144.035µs wal_replay_duration=483.719µs wbl_replay_duration=320ns total_replay_duration=672.395µs prometheus | ts=2024-04-24T09:37:36.293Z caller=main.go:1150 level=info fs_type=EXT4_SUPER_MAGIC prometheus | ts=2024-04-24T09:37:36.293Z caller=main.go:1153 level=info msg="TSDB started" prometheus | ts=2024-04-24T09:37:36.293Z caller=main.go:1335 level=info msg="Loading configuration file" filename=/etc/prometheus/prometheus.yml prometheus | ts=2024-04-24T09:37:36.294Z caller=main.go:1372 level=info msg="Completed loading of configuration file" filename=/etc/prometheus/prometheus.yml totalDuration=820.21µs db_storage=1.35µs remote_storage=2.52µs web_handler=450ns query_engine=1.06µs scrape=222.408µs scrape_sd=111.004µs notify=24.941µs notify_sd=10.331µs rules=2.27µs tracing=4.91µs prometheus | ts=2024-04-24T09:37:36.294Z caller=main.go:1114 level=info msg="Server is ready to receive web requests." prometheus | ts=2024-04-24T09:37:36.294Z caller=manager.go:163 level=info component="rule manager" msg="Starting rule manager..." policy-db-migrator | 99 0120-pk_pdpstatistics.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:50 policy-db-migrator | 100 0130-pdpstatistics.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:50 policy-db-migrator | 101 0140-pk_pdpstatistics.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:51 policy-db-migrator | 102 0150-pdpstatistics.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:51 policy-db-migrator | 103 0160-jpapdpstatistics_enginestats.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:51 policy-db-migrator | 104 0170-jpapdpstatistics_enginestats.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:51 policy-db-migrator | 105 0180-jpapdpstatistics_enginestats.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:51 policy-db-migrator | 106 0190-jpapolicyaudit.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:51 policy-db-migrator | 107 0200-JpaPolicyAuditIndex_timestamp.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:51 policy-db-migrator | 108 0210-sequence.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:51 policy-db-migrator | 109 0220-sequence.sql upgrade 0800 0900 2404240937450900u 1 2024-04-24 09:37:51 policy-db-migrator | 110 0100-jpatoscapolicy_targets.sql upgrade 0900 1000 2404240937451000u 1 2024-04-24 09:37:51 policy-db-migrator | 111 0110-jpatoscapolicytype_targets.sql upgrade 0900 1000 2404240937451000u 1 2024-04-24 09:37:51 policy-db-migrator | 112 0120-toscatrigger.sql upgrade 0900 1000 2404240937451000u 1 2024-04-24 09:37:51 policy-db-migrator | 113 0130-jpatoscapolicytype_triggers.sql upgrade 0900 1000 2404240937451000u 1 2024-04-24 09:37:51 policy-db-migrator | 114 0140-toscaparameter.sql upgrade 0900 1000 2404240937451000u 1 2024-04-24 09:37:51 policy-db-migrator | 115 0150-toscaproperty.sql upgrade 0900 1000 2404240937451000u 1 2024-04-24 09:37:51 policy-db-migrator | 116 0160-jpapolicyaudit_pk.sql upgrade 0900 1000 2404240937451000u 1 2024-04-24 09:37:51 policy-db-migrator | 117 0170-pdpstatistics_pk.sql upgrade 0900 1000 2404240937451000u 1 2024-04-24 09:37:52 policy-db-migrator | 118 0180-jpatoscanodetemplate_metadata.sql upgrade 0900 1000 2404240937451000u 1 2024-04-24 09:37:52 policy-db-migrator | 119 0100-upgrade.sql upgrade 1000 1100 2404240937451100u 1 2024-04-24 09:37:52 policy-db-migrator | 120 0100-jpapolicyaudit_renameuser.sql upgrade 1100 1200 2404240937451200u 1 2024-04-24 09:37:52 policy-db-migrator | 121 0110-idx_tsidx1.sql upgrade 1100 1200 2404240937451200u 1 2024-04-24 09:37:52 policy-db-migrator | 122 0120-audit_sequence.sql upgrade 1100 1200 2404240937451200u 1 2024-04-24 09:37:52 policy-db-migrator | 123 0130-statistics_sequence.sql upgrade 1100 1200 2404240937451200u 1 2024-04-24 09:37:52 policy-db-migrator | 124 0100-pdpstatistics.sql upgrade 1200 1300 2404240937451300u 1 2024-04-24 09:37:52 policy-db-migrator | 125 0110-jpapdpstatistics_enginestats.sql upgrade 1200 1300 2404240937451300u 1 2024-04-24 09:37:52 policy-db-migrator | 126 0120-statistics_sequence.sql upgrade 1200 1300 2404240937451300u 1 2024-04-24 09:37:52 policy-db-migrator | policyadmin: OK @ 1300 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | session.timeout.ms = 45000 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null grafana | logger=migrator t=2024-04-24T09:37:38.128165136Z level=info msg="Migration successfully executed" id="Update dashboard_snapshot table charset" duration=122.455µs kafka | [2024-04-24 09:37:50,419] INFO [zk-broker-1-to-controller-alter-partition-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) kafka | [2024-04-24 09:37:50,444] INFO [ExpirationReaper-1-Produce]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2024-04-24 09:37:50,446] INFO [ExpirationReaper-1-Fetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2024-04-24 09:37:50,447] INFO [ExpirationReaper-1-DeleteRecords]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2024-04-24 09:37:50,451] INFO [ExpirationReaper-1-ElectLeader]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2024-04-24 09:37:50,452] INFO [ExpirationReaper-1-RemoteFetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2024-04-24 09:37:50,468] INFO [LogDirFailureHandler]: Starting (kafka.server.ReplicaManager$LogDirFailureHandler) kafka | [2024-04-24 09:37:50,469] INFO [AddPartitionsToTxnSenderThread-1]: Starting (kafka.server.AddPartitionsToTxnManager) kafka | [2024-04-24 09:37:50,498] INFO Creating /brokers/ids/1 (is it secure? false) (kafka.zk.KafkaZkClient) kafka | [2024-04-24 09:37:50,528] INFO Stat of the created znode at /brokers/ids/1 is: 27,27,1713951470513,1713951470513,1,0,0,72057616913268737,258,0,27 kafka | (kafka.zk.KafkaZkClient) kafka | [2024-04-24 09:37:50,529] INFO Registered broker 1 at path /brokers/ids/1 with addresses: PLAINTEXT://kafka:9092,PLAINTEXT_HOST://localhost:29092, czxid (broker epoch): 27 (kafka.zk.KafkaZkClient) kafka | [2024-04-24 09:37:50,583] INFO [ControllerEventThread controllerId=1] Starting (kafka.controller.ControllerEventManager$ControllerEventThread) kafka | [2024-04-24 09:37:50,590] INFO [ExpirationReaper-1-topic]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2024-04-24 09:37:50,596] INFO [ExpirationReaper-1-Heartbeat]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2024-04-24 09:37:50,597] INFO [ExpirationReaper-1-Rebalance]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2024-04-24 09:37:50,608] INFO Successfully created /controller_epoch with initial epoch 0 (kafka.zk.KafkaZkClient) kafka | [2024-04-24 09:37:50,613] INFO [GroupCoordinator 1]: Starting up. (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:37:50,627] INFO [Controller id=1] 1 successfully elected as the controller. Epoch incremented to 1 and epoch zk version is now 1 (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,628] INFO [GroupCoordinator 1]: Startup complete. (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:37:50,638] INFO [Controller id=1] Creating FeatureZNode at path: /feature with contents: FeatureZNode(2,Enabled,Map()) (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,644] INFO Feature ZK node created at path: /feature (kafka.server.FinalizedFeatureChangeListener) kafka | [2024-04-24 09:37:50,645] INFO [TransactionCoordinator id=1] Starting up. (kafka.coordinator.transaction.TransactionCoordinator) kafka | [2024-04-24 09:37:50,651] INFO [TxnMarkerSenderThread-1]: Starting (kafka.coordinator.transaction.TransactionMarkerChannelManager) kafka | [2024-04-24 09:37:50,652] INFO [TransactionCoordinator id=1] Startup complete. (kafka.coordinator.transaction.TransactionCoordinator) kafka | [2024-04-24 09:37:50,675] INFO [MetadataCache brokerId=1] Updated cache from existing None to latest Features(version=3.6-IV2, finalizedFeatures={}, finalizedFeaturesEpoch=0). (kafka.server.metadata.ZkMetadataCache) kafka | [2024-04-24 09:37:50,675] INFO [Controller id=1] Registering handlers (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,686] INFO [ExpirationReaper-1-AlterAcls]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | [2024-04-24 09:37:50,686] INFO [Controller id=1] Deleting log dir event notifications (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,691] INFO [Controller id=1] Deleting isr change notifications (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,697] INFO [Controller id=1] Initializing controller context (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,712] INFO [/config/changes-event-process-thread]: Starting (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) kafka | [2024-04-24 09:37:50,714] INFO [Controller id=1] Initialized broker epochs cache: HashMap(1 -> 27) (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,724] DEBUG [Controller id=1] Register BrokerModifications handler for Set(1) (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,730] INFO [SocketServer listenerType=ZK_BROKER, nodeId=1] Enabling request processing. (kafka.network.SocketServer) kafka | [2024-04-24 09:37:50,731] DEBUG [Channel manager on controller 1]: Controller 1 trying to connect to broker 1 (kafka.controller.ControllerChannelManager) kafka | [2024-04-24 09:37:50,734] INFO Awaiting socket connections on 0.0.0.0:29092. (kafka.network.DataPlaneAcceptor) kafka | [2024-04-24 09:37:50,737] INFO Awaiting socket connections on 0.0.0.0:9092. (kafka.network.DataPlaneAcceptor) kafka | [2024-04-24 09:37:50,747] INFO [Controller id=1] Currently active brokers in the cluster: Set(1) (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,747] INFO [RequestSendThread controllerId=1] Starting (kafka.controller.RequestSendThread) kafka | [2024-04-24 09:37:50,747] INFO [Controller id=1] Currently shutting brokers in the cluster: HashSet() (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,749] INFO [Controller id=1] Current list of topics in the cluster: HashSet() (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,749] INFO [Controller id=1] Fetching topic deletions in progress (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,751] INFO Kafka version: 7.6.1-ccs (org.apache.kafka.common.utils.AppInfoParser) kafka | [2024-04-24 09:37:50,751] INFO Kafka commitId: 11e81ad2a49db00b1d2b8c731409cd09e563de67 (org.apache.kafka.common.utils.AppInfoParser) kafka | [2024-04-24 09:37:50,751] INFO Kafka startTimeMs: 1713951470745 (org.apache.kafka.common.utils.AppInfoParser) kafka | [2024-04-24 09:37:50,752] INFO [KafkaServer id=1] started (kafka.server.KafkaServer) kafka | [2024-04-24 09:37:50,755] INFO [Controller id=1] List of topics to be deleted: (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,756] INFO [Controller id=1] List of topics ineligible for deletion: (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,756] INFO [Controller id=1] Initializing topic deletion manager (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,757] INFO [Topic Deletion Manager 1] Initializing manager with initial deletions: Set(), initial ineligible deletions: HashSet() (kafka.controller.TopicDeletionManager) kafka | [2024-04-24 09:37:50,758] INFO [Controller id=1] Sending update metadata request (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,764] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet(1) for 0 partitions (state.change.logger) kafka | [2024-04-24 09:37:50,773] INFO [ReplicaStateMachine controllerId=1] Initializing replica state (kafka.controller.ZkReplicaStateMachine) kafka | [2024-04-24 09:37:50,774] INFO [ReplicaStateMachine controllerId=1] Triggering online replica state changes (kafka.controller.ZkReplicaStateMachine) kafka | [2024-04-24 09:37:50,777] INFO [ReplicaStateMachine controllerId=1] Triggering offline replica state changes (kafka.controller.ZkReplicaStateMachine) policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | policy-pap | [2024-04-24T09:38:16.187+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 policy-pap | [2024-04-24T09:38:16.187+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 policy-pap | [2024-04-24T09:38:16.187+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1713951496187 policy-pap | [2024-04-24T09:38:16.187+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-policy-pap-2, groupId=policy-pap] Subscribed to topic(s): policy-pdp-pap policy-pap | [2024-04-24T09:38:16.500+00:00|INFO|PapDatabaseInitializer|main] Created initial pdpGroup in DB - PdpGroups(groups=[PdpGroup(name=defaultGroup, description=The default group that registers all supported policy types and pdps., pdpGroupState=ACTIVE, properties=null, pdpSubgroups=[PdpSubGroup(pdpType=apex, supportedPolicyTypes=[onap.policies.controlloop.operational.common.Apex 1.0.0, onap.policies.native.Apex 1.0.0], policies=[], currentInstanceCount=0, desiredInstanceCount=1, properties=null, pdpInstances=null)])]) from /opt/app/policy/pap/etc/mounted/groups.json kafka | [2024-04-24 09:37:50,778] DEBUG [ReplicaStateMachine controllerId=1] Started replica state machine with initial state -> HashMap() (kafka.controller.ZkReplicaStateMachine) kafka | [2024-04-24 09:37:50,778] INFO [PartitionStateMachine controllerId=1] Initializing partition state (kafka.controller.ZkPartitionStateMachine) kafka | [2024-04-24 09:37:50,779] INFO [PartitionStateMachine controllerId=1] Triggering online partition state changes (kafka.controller.ZkPartitionStateMachine) kafka | [2024-04-24 09:37:50,781] DEBUG [PartitionStateMachine controllerId=1] Started partition state machine with initial state -> HashMap() (kafka.controller.ZkPartitionStateMachine) kafka | [2024-04-24 09:37:50,782] INFO [Controller id=1] Ready to serve as the new controller with epoch 1 (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,789] INFO [RequestSendThread controllerId=1] Controller 1 connected to kafka:9092 (id: 1 rack: null) for sending state change requests (kafka.controller.RequestSendThread) kafka | [2024-04-24 09:37:50,790] INFO [Controller id=1] Partitions undergoing preferred replica election: (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,790] INFO [Controller id=1] Partitions that completed preferred replica election: (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,791] INFO [Controller id=1] Skipping preferred replica election for partitions due to topic deletion: (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,792] INFO [Controller id=1] Resuming preferred replica election for partitions: (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,792] INFO [Controller id=1] Starting replica leader election (PREFERRED) for partitions triggered by ZkTriggered (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,807] INFO [Controller id=1] Starting the controller scheduler (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:50,843] INFO [zk-broker-1-to-controller-forwarding-channel-manager]: Recorded new controller, from now on will use node kafka:9092 (id: 1 rack: null) (kafka.server.BrokerToControllerRequestThread) kafka | [2024-04-24 09:37:50,846] TRACE [Controller id=1 epoch=1] Received response UpdateMetadataResponseData(errorCode=0) for request UPDATE_METADATA with correlation id 0 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) kafka | [2024-04-24 09:37:50,924] INFO [zk-broker-1-to-controller-alter-partition-channel-manager]: Recorded new controller, from now on will use node kafka:9092 (id: 1 rack: null) (kafka.server.BrokerToControllerRequestThread) kafka | [2024-04-24 09:37:55,809] INFO [Controller id=1] Processing automatic preferred replica leader election (kafka.controller.KafkaController) kafka | [2024-04-24 09:37:55,809] TRACE [Controller id=1] Checking need to trigger auto leader balancing (kafka.controller.KafkaController) kafka | [2024-04-24 09:38:18,401] INFO Creating topic policy-pdp-pap with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(1)) (kafka.zk.AdminZkClient) kafka | [2024-04-24 09:38:18,402] INFO Creating topic __consumer_offsets with configuration {compression.type=producer, cleanup.policy=compact, segment.bytes=104857600} and initial partition assignment HashMap(0 -> ArrayBuffer(1), 1 -> ArrayBuffer(1), 2 -> ArrayBuffer(1), 3 -> ArrayBuffer(1), 4 -> ArrayBuffer(1), 5 -> ArrayBuffer(1), 6 -> ArrayBuffer(1), 7 -> ArrayBuffer(1), 8 -> ArrayBuffer(1), 9 -> ArrayBuffer(1), 10 -> ArrayBuffer(1), 11 -> ArrayBuffer(1), 12 -> ArrayBuffer(1), 13 -> ArrayBuffer(1), 14 -> ArrayBuffer(1), 15 -> ArrayBuffer(1), 16 -> ArrayBuffer(1), 17 -> ArrayBuffer(1), 18 -> ArrayBuffer(1), 19 -> ArrayBuffer(1), 20 -> ArrayBuffer(1), 21 -> ArrayBuffer(1), 22 -> ArrayBuffer(1), 23 -> ArrayBuffer(1), 24 -> ArrayBuffer(1), 25 -> ArrayBuffer(1), 26 -> ArrayBuffer(1), 27 -> ArrayBuffer(1), 28 -> ArrayBuffer(1), 29 -> ArrayBuffer(1), 30 -> ArrayBuffer(1), 31 -> ArrayBuffer(1), 32 -> ArrayBuffer(1), 33 -> ArrayBuffer(1), 34 -> ArrayBuffer(1), 35 -> ArrayBuffer(1), 36 -> ArrayBuffer(1), 37 -> ArrayBuffer(1), 38 -> ArrayBuffer(1), 39 -> ArrayBuffer(1), 40 -> ArrayBuffer(1), 41 -> ArrayBuffer(1), 42 -> ArrayBuffer(1), 43 -> ArrayBuffer(1), 44 -> ArrayBuffer(1), 45 -> ArrayBuffer(1), 46 -> ArrayBuffer(1), 47 -> ArrayBuffer(1), 48 -> ArrayBuffer(1), 49 -> ArrayBuffer(1)) (kafka.zk.AdminZkClient) kafka | [2024-04-24 09:38:18,410] DEBUG [Controller id=1] There is no producerId block yet (Zk path version 0), creating the first block (kafka.controller.KafkaController) kafka | [2024-04-24 09:38:18,424] INFO [Controller id=1] Acquired new producerId block ProducerIdsBlock(assignedBrokerId=1, firstProducerId=0, size=1000) by writing to Zk with path version 1 (kafka.controller.KafkaController) grafana | logger=migrator t=2024-04-24T09:37:38.131348921Z level=info msg="Executing migration" id="Add column external_delete_url to dashboard_snapshots table" grafana | logger=migrator t=2024-04-24T09:37:38.134710822Z level=info msg="Migration successfully executed" id="Add column external_delete_url to dashboard_snapshots table" duration=3.362121ms grafana | logger=migrator t=2024-04-24T09:37:38.137560464Z level=info msg="Executing migration" id="Add encrypted dashboard json column" grafana | logger=migrator t=2024-04-24T09:37:38.140400865Z level=info msg="Migration successfully executed" id="Add encrypted dashboard json column" duration=2.840061ms grafana | logger=migrator t=2024-04-24T09:37:38.143636671Z level=info msg="Executing migration" id="Change dashboard_encrypted column to MEDIUMBLOB" grafana | logger=migrator t=2024-04-24T09:37:38.143736005Z level=info msg="Migration successfully executed" id="Change dashboard_encrypted column to MEDIUMBLOB" duration=98.564µs grafana | logger=migrator t=2024-04-24T09:37:38.146732392Z level=info msg="Executing migration" id="create quota table v1" grafana | logger=migrator t=2024-04-24T09:37:38.147561044Z level=info msg="Migration successfully executed" id="create quota table v1" duration=821.952µs grafana | logger=migrator t=2024-04-24T09:37:38.152197526Z level=info msg="Executing migration" id="create index UQE_quota_org_id_user_id_target - v1" grafana | logger=migrator t=2024-04-24T09:37:38.153661903Z level=info msg="Migration successfully executed" id="create index UQE_quota_org_id_user_id_target - v1" duration=1.462437ms grafana | logger=migrator t=2024-04-24T09:37:38.157549205Z level=info msg="Executing migration" id="Update quota table charset" grafana | logger=migrator t=2024-04-24T09:37:38.157590737Z level=info msg="Migration successfully executed" id="Update quota table charset" duration=40.992µs grafana | logger=migrator t=2024-04-24T09:37:38.161581602Z level=info msg="Executing migration" id="create plugin_setting table" grafana | logger=migrator t=2024-04-24T09:37:38.162410225Z level=info msg="Migration successfully executed" id="create plugin_setting table" duration=828.493µs grafana | logger=migrator t=2024-04-24T09:37:38.167116559Z level=info msg="Executing migration" id="create index UQE_plugin_setting_org_id_plugin_id - v1" grafana | logger=migrator t=2024-04-24T09:37:38.168513364Z level=info msg="Migration successfully executed" id="create index UQE_plugin_setting_org_id_plugin_id - v1" duration=1.393515ms grafana | logger=migrator t=2024-04-24T09:37:38.173226788Z level=info msg="Executing migration" id="Add column plugin_version to plugin_settings" grafana | logger=migrator t=2024-04-24T09:37:38.176383262Z level=info msg="Migration successfully executed" id="Add column plugin_version to plugin_settings" duration=3.158664ms grafana | logger=migrator t=2024-04-24T09:37:38.179946651Z level=info msg="Executing migration" id="Update plugin_setting table charset" grafana | logger=migrator t=2024-04-24T09:37:38.179969532Z level=info msg="Migration successfully executed" id="Update plugin_setting table charset" duration=23.511µs grafana | logger=migrator t=2024-04-24T09:37:38.184614794Z level=info msg="Executing migration" id="create session table" grafana | logger=migrator t=2024-04-24T09:37:38.185499798Z level=info msg="Migration successfully executed" id="create session table" duration=883.644µs grafana | logger=migrator t=2024-04-24T09:37:38.189111649Z level=info msg="Executing migration" id="Drop old table playlist table" grafana | logger=migrator t=2024-04-24T09:37:38.189198112Z level=info msg="Migration successfully executed" id="Drop old table playlist table" duration=86.513µs grafana | logger=migrator t=2024-04-24T09:37:38.192981521Z level=info msg="Executing migration" id="Drop old table playlist_item table" grafana | logger=migrator t=2024-04-24T09:37:38.193116056Z level=info msg="Migration successfully executed" id="Drop old table playlist_item table" duration=135.115µs grafana | logger=migrator t=2024-04-24T09:37:38.197019648Z level=info msg="Executing migration" id="create playlist table v2" grafana | logger=migrator t=2024-04-24T09:37:38.198136371Z level=info msg="Migration successfully executed" id="create playlist table v2" duration=1.112373ms grafana | logger=migrator t=2024-04-24T09:37:38.20218708Z level=info msg="Executing migration" id="create playlist item table v2" grafana | logger=migrator t=2024-04-24T09:37:38.203771702Z level=info msg="Migration successfully executed" id="create playlist item table v2" duration=1.589582ms grafana | logger=migrator t=2024-04-24T09:37:38.245026165Z level=info msg="Executing migration" id="Update playlist table charset" grafana | logger=migrator t=2024-04-24T09:37:38.245081997Z level=info msg="Migration successfully executed" id="Update playlist table charset" duration=59.852µs grafana | logger=migrator t=2024-04-24T09:37:38.248617085Z level=info msg="Executing migration" id="Update playlist_item table charset" grafana | logger=migrator t=2024-04-24T09:37:38.248660567Z level=info msg="Migration successfully executed" id="Update playlist_item table charset" duration=38.592µs grafana | logger=migrator t=2024-04-24T09:37:38.251661005Z level=info msg="Executing migration" id="Add playlist column created_at" grafana | logger=migrator t=2024-04-24T09:37:38.255863279Z level=info msg="Migration successfully executed" id="Add playlist column created_at" duration=4.203614ms grafana | logger=migrator t=2024-04-24T09:37:38.262077952Z level=info msg="Executing migration" id="Add playlist column updated_at" grafana | logger=migrator t=2024-04-24T09:37:38.265640342Z level=info msg="Migration successfully executed" id="Add playlist column updated_at" duration=3.562769ms grafana | logger=migrator t=2024-04-24T09:37:38.268753843Z level=info msg="Executing migration" id="drop preferences table v2" grafana | logger=migrator t=2024-04-24T09:37:38.268841027Z level=info msg="Migration successfully executed" id="drop preferences table v2" duration=87.084µs grafana | logger=migrator t=2024-04-24T09:37:38.2714743Z level=info msg="Executing migration" id="drop preferences table v3" grafana | logger=migrator t=2024-04-24T09:37:38.271579134Z level=info msg="Migration successfully executed" id="drop preferences table v3" duration=101.484µs grafana | logger=migrator t=2024-04-24T09:37:38.274702506Z level=info msg="Executing migration" id="create preferences table v3" grafana | logger=migrator t=2024-04-24T09:37:38.2755745Z level=info msg="Migration successfully executed" id="create preferences table v3" duration=870.633µs grafana | logger=migrator t=2024-04-24T09:37:38.280921419Z level=info msg="Executing migration" id="Update preferences table charset" grafana | logger=migrator t=2024-04-24T09:37:38.28096335Z level=info msg="Migration successfully executed" id="Update preferences table charset" duration=47.092µs grafana | logger=migrator t=2024-04-24T09:37:38.284455027Z level=info msg="Executing migration" id="Add column team_id in preferences" grafana | logger=migrator t=2024-04-24T09:37:38.289486894Z level=info msg="Migration successfully executed" id="Add column team_id in preferences" duration=5.031417ms grafana | logger=migrator t=2024-04-24T09:37:38.292895477Z level=info msg="Executing migration" id="Update team_id column values in preferences" grafana | logger=migrator t=2024-04-24T09:37:38.293045862Z level=info msg="Migration successfully executed" id="Update team_id column values in preferences" duration=150.635µs grafana | logger=migrator t=2024-04-24T09:37:38.299507666Z level=info msg="Executing migration" id="Add column week_start in preferences" grafana | logger=migrator t=2024-04-24T09:37:38.304964739Z level=info msg="Migration successfully executed" id="Add column week_start in preferences" duration=5.457393ms grafana | logger=migrator t=2024-04-24T09:37:38.308617442Z level=info msg="Executing migration" id="Add column preferences.json_data" policy-pap | [2024-04-24T09:38:16.664+00:00|WARN|JpaBaseConfiguration$JpaWebConfiguration|main] spring.jpa.open-in-view is enabled by default. Therefore, database queries may be performed during view rendering. Explicitly configure spring.jpa.open-in-view to disable this warning policy-pap | [2024-04-24T09:38:16.910+00:00|INFO|DefaultSecurityFilterChain|main] Will secure any request with [org.springframework.security.web.session.DisableEncodeUrlFilter@29dfc68f, org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter@320ba79f, org.springframework.security.web.context.SecurityContextHolderFilter@29ee8174, org.springframework.security.web.header.HeaderWriterFilter@79ee779c, org.springframework.security.web.authentication.logout.LogoutFilter@4c9d833, org.springframework.security.web.authentication.www.BasicAuthenticationFilter@2e7517aa, org.springframework.security.web.savedrequest.RequestCacheAwareFilter@6549ce71, org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter@2e030cc1, org.springframework.security.web.authentication.AnonymousAuthenticationFilter@333a2df2, org.springframework.security.web.access.ExceptionTranslationFilter@61c145cb, org.springframework.security.web.access.intercept.AuthorizationFilter@d02c00] policy-pap | [2024-04-24T09:38:17.704+00:00|INFO|EndpointLinksResolver|main] Exposing 3 endpoint(s) beneath base path '' policy-pap | [2024-04-24T09:38:17.809+00:00|INFO|Http11NioProtocol|main] Starting ProtocolHandler ["http-nio-6969"] policy-pap | [2024-04-24T09:38:17.844+00:00|INFO|TomcatWebServer|main] Tomcat started on port(s): 6969 (http) with context path '/policy/pap/v1' policy-pap | [2024-04-24T09:38:17.864+00:00|INFO|ServiceManager|main] Policy PAP starting policy-pap | [2024-04-24T09:38:17.864+00:00|INFO|ServiceManager|main] Policy PAP starting Meter Registry policy-pap | [2024-04-24T09:38:17.866+00:00|INFO|ServiceManager|main] Policy PAP starting PAP parameters policy-pap | [2024-04-24T09:38:17.866+00:00|INFO|ServiceManager|main] Policy PAP starting Pdp Heartbeat Listener policy-pap | [2024-04-24T09:38:17.866+00:00|INFO|ServiceManager|main] Policy PAP starting Response Request ID Dispatcher policy-pap | [2024-04-24T09:38:17.867+00:00|INFO|ServiceManager|main] Policy PAP starting Heartbeat Request ID Dispatcher policy-pap | [2024-04-24T09:38:17.867+00:00|INFO|ServiceManager|main] Policy PAP starting Response Message Dispatcher policy-pap | [2024-04-24T09:38:17.869+00:00|INFO|TopicBase|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=0d71b024-87db-4bab-bad3-947ec269530c, consumerInstance=policy-pap, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: registering org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher@5598dff2 policy-pap | [2024-04-24T09:38:17.882+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=0d71b024-87db-4bab-bad3-947ec269530c, consumerInstance=policy-pap, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-pap | [2024-04-24T09:38:17.883+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-pap | allow.auto.create.topics = true policy-pap | auto.commit.interval.ms = 5000 policy-pap | auto.include.jmx.reporter = true policy-pap | auto.offset.reset = latest policy-pap | bootstrap.servers = [kafka:9092] policy-pap | check.crcs = true policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = consumer-0d71b024-87db-4bab-bad3-947ec269530c-3 policy-pap | client.rack = policy-pap | connections.max.idle.ms = 540000 policy-pap | default.api.timeout.ms = 60000 policy-pap | enable.auto.commit = true policy-pap | exclude.internal.topics = true policy-pap | fetch.max.bytes = 52428800 policy-pap | fetch.max.wait.ms = 500 policy-pap | fetch.min.bytes = 1 policy-pap | group.id = 0d71b024-87db-4bab-bad3-947ec269530c policy-pap | group.instance.id = null policy-pap | heartbeat.interval.ms = 3000 policy-pap | interceptor.classes = [] policy-pap | internal.leave.group.on.close = true policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false policy-pap | isolation.level = read_uncommitted policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | max.partition.fetch.bytes = 1048576 grafana | logger=migrator t=2024-04-24T09:37:38.312060726Z level=info msg="Migration successfully executed" id="Add column preferences.json_data" duration=3.442564ms grafana | logger=migrator t=2024-04-24T09:37:38.315513611Z level=info msg="Executing migration" id="alter preferences.json_data to mediumtext v1" grafana | logger=migrator t=2024-04-24T09:37:38.315579893Z level=info msg="Migration successfully executed" id="alter preferences.json_data to mediumtext v1" duration=66.912µs grafana | logger=migrator t=2024-04-24T09:37:38.318910174Z level=info msg="Executing migration" id="Add preferences index org_id" grafana | logger=migrator t=2024-04-24T09:37:38.319851071Z level=info msg="Migration successfully executed" id="Add preferences index org_id" duration=931.906µs grafana | logger=migrator t=2024-04-24T09:37:38.325737851Z level=info msg="Executing migration" id="Add preferences index user_id" grafana | logger=migrator t=2024-04-24T09:37:38.327238699Z level=info msg="Migration successfully executed" id="Add preferences index user_id" duration=1.498488ms grafana | logger=migrator t=2024-04-24T09:37:38.331300908Z level=info msg="Executing migration" id="create alert table v1" grafana | logger=migrator t=2024-04-24T09:37:38.333076928Z level=info msg="Migration successfully executed" id="create alert table v1" duration=1.77593ms grafana | logger=migrator t=2024-04-24T09:37:38.338969299Z level=info msg="Executing migration" id="add index alert org_id & id " grafana | logger=migrator t=2024-04-24T09:37:38.339890904Z level=info msg="Migration successfully executed" id="add index alert org_id & id " duration=920.926µs grafana | logger=migrator t=2024-04-24T09:37:38.343221094Z level=info msg="Executing migration" id="add index alert state" grafana | logger=migrator t=2024-04-24T09:37:38.344068497Z level=info msg="Migration successfully executed" id="add index alert state" duration=849.273µs grafana | logger=migrator t=2024-04-24T09:37:38.347394368Z level=info msg="Executing migration" id="add index alert dashboard_id" grafana | logger=migrator t=2024-04-24T09:37:38.348251091Z level=info msg="Migration successfully executed" id="add index alert dashboard_id" duration=862.163µs grafana | logger=migrator t=2024-04-24T09:37:38.354190503Z level=info msg="Executing migration" id="Create alert_rule_tag table v1" grafana | logger=migrator t=2024-04-24T09:37:38.354851299Z level=info msg="Migration successfully executed" id="Create alert_rule_tag table v1" duration=660.466µs grafana | logger=migrator t=2024-04-24T09:37:38.358402228Z level=info msg="Executing migration" id="Add unique index alert_rule_tag.alert_id_tag_id" grafana | logger=migrator t=2024-04-24T09:37:38.359691578Z level=info msg="Migration successfully executed" id="Add unique index alert_rule_tag.alert_id_tag_id" duration=1.28797ms grafana | logger=migrator t=2024-04-24T09:37:38.363436614Z level=info msg="Executing migration" id="drop index UQE_alert_rule_tag_alert_id_tag_id - v1" grafana | logger=migrator t=2024-04-24T09:37:38.364775907Z level=info msg="Migration successfully executed" id="drop index UQE_alert_rule_tag_alert_id_tag_id - v1" duration=1.339273ms grafana | logger=migrator t=2024-04-24T09:37:38.370382397Z level=info msg="Executing migration" id="Rename table alert_rule_tag to alert_rule_tag_v1 - v1" grafana | logger=migrator t=2024-04-24T09:37:38.385815539Z level=info msg="Migration successfully executed" id="Rename table alert_rule_tag to alert_rule_tag_v1 - v1" duration=15.434323ms grafana | logger=migrator t=2024-04-24T09:37:38.389837607Z level=info msg="Executing migration" id="Create alert_rule_tag table v2" grafana | logger=migrator t=2024-04-24T09:37:38.391313204Z level=info msg="Migration successfully executed" id="Create alert_rule_tag table v2" duration=1.476477ms grafana | logger=migrator t=2024-04-24T09:37:38.403032523Z level=info msg="Executing migration" id="create index UQE_alert_rule_tag_alert_id_tag_id - Add unique index alert_rule_tag.alert_id_tag_id V2" grafana | logger=migrator t=2024-04-24T09:37:38.404533322Z level=info msg="Migration successfully executed" id="create index UQE_alert_rule_tag_alert_id_tag_id - Add unique index alert_rule_tag.alert_id_tag_id V2" duration=1.500548ms grafana | logger=migrator t=2024-04-24T09:37:38.408149573Z level=info msg="Executing migration" id="copy alert_rule_tag v1 to v2" grafana | logger=migrator t=2024-04-24T09:37:38.408785668Z level=info msg="Migration successfully executed" id="copy alert_rule_tag v1 to v2" duration=635.525µs grafana | logger=migrator t=2024-04-24T09:37:38.41420413Z level=info msg="Executing migration" id="drop table alert_rule_tag_v1" grafana | logger=migrator t=2024-04-24T09:37:38.415132336Z level=info msg="Migration successfully executed" id="drop table alert_rule_tag_v1" duration=927.356µs grafana | logger=migrator t=2024-04-24T09:37:38.420259807Z level=info msg="Executing migration" id="create alert_notification table v1" grafana | logger=migrator t=2024-04-24T09:37:38.421236105Z level=info msg="Migration successfully executed" id="create alert_notification table v1" duration=975.928µs grafana | logger=migrator t=2024-04-24T09:37:38.425119106Z level=info msg="Executing migration" id="Add column is_default" grafana | logger=migrator t=2024-04-24T09:37:38.433969093Z level=info msg="Migration successfully executed" id="Add column is_default" duration=8.837796ms grafana | logger=migrator t=2024-04-24T09:37:38.442004696Z level=info msg="Executing migration" id="Add column frequency" grafana | logger=migrator t=2024-04-24T09:37:38.447236251Z level=info msg="Migration successfully executed" id="Add column frequency" duration=5.233285ms grafana | logger=migrator t=2024-04-24T09:37:38.452625442Z level=info msg="Executing migration" id="Add column send_reminder" grafana | logger=migrator t=2024-04-24T09:37:38.456465503Z level=info msg="Migration successfully executed" id="Add column send_reminder" duration=3.83972ms grafana | logger=migrator t=2024-04-24T09:37:38.465601369Z level=info msg="Executing migration" id="Add column disable_resolve_message" grafana | logger=migrator t=2024-04-24T09:37:38.472312721Z level=info msg="Migration successfully executed" id="Add column disable_resolve_message" duration=6.708872ms grafana | logger=migrator t=2024-04-24T09:37:38.496846081Z level=info msg="Executing migration" id="add index alert_notification org_id & name" grafana | logger=migrator t=2024-04-24T09:37:38.498304118Z level=info msg="Migration successfully executed" id="add index alert_notification org_id & name" duration=1.461477ms grafana | logger=migrator t=2024-04-24T09:37:38.503272742Z level=info msg="Executing migration" id="Update alert table charset" grafana | logger=migrator t=2024-04-24T09:37:38.503302023Z level=info msg="Migration successfully executed" id="Update alert table charset" duration=27.271µs grafana | logger=migrator t=2024-04-24T09:37:38.506966437Z level=info msg="Executing migration" id="Update alert_notification table charset" grafana | logger=migrator t=2024-04-24T09:37:38.506990608Z level=info msg="Migration successfully executed" id="Update alert_notification table charset" duration=24.931µs grafana | logger=migrator t=2024-04-24T09:37:38.510664461Z level=info msg="Executing migration" id="create notification_journal table v1" grafana | logger=migrator t=2024-04-24T09:37:38.511628209Z level=info msg="Migration successfully executed" id="create notification_journal table v1" duration=964.318µs grafana | logger=migrator t=2024-04-24T09:37:38.517996048Z level=info msg="Executing migration" id="add index notification_journal org_id & alert_id & notifier_id" grafana | logger=migrator t=2024-04-24T09:37:38.519183434Z level=info msg="Migration successfully executed" id="add index notification_journal org_id & alert_id & notifier_id" duration=1.187696ms grafana | logger=migrator t=2024-04-24T09:37:38.525482Z level=info msg="Executing migration" id="drop alert_notification_journal" grafana | logger=migrator t=2024-04-24T09:37:38.52621662Z level=info msg="Migration successfully executed" id="drop alert_notification_journal" duration=735.87µs grafana | logger=migrator t=2024-04-24T09:37:38.532074379Z level=info msg="Executing migration" id="create alert_notification_state table v1" grafana | logger=migrator t=2024-04-24T09:37:38.533163511Z level=info msg="Migration successfully executed" id="create alert_notification_state table v1" duration=1.091993ms grafana | logger=migrator t=2024-04-24T09:37:38.538005361Z level=info msg="Executing migration" id="add index alert_notification_state org_id & alert_id & notifier_id" grafana | logger=migrator t=2024-04-24T09:37:38.539058741Z level=info msg="Migration successfully executed" id="add index alert_notification_state org_id & alert_id & notifier_id" duration=1.055611ms grafana | logger=migrator t=2024-04-24T09:37:38.545068196Z level=info msg="Executing migration" id="Add for to alert table" grafana | logger=migrator t=2024-04-24T09:37:38.548137756Z level=info msg="Migration successfully executed" id="Add for to alert table" duration=3.06461ms grafana | logger=migrator t=2024-04-24T09:37:38.555186382Z level=info msg="Executing migration" id="Add column uid in alert_notification" grafana | logger=migrator t=2024-04-24T09:37:38.560605154Z level=info msg="Migration successfully executed" id="Add column uid in alert_notification" duration=5.424292ms grafana | logger=migrator t=2024-04-24T09:37:38.565752225Z level=info msg="Executing migration" id="Update uid column values in alert_notification" grafana | logger=migrator t=2024-04-24T09:37:38.566178902Z level=info msg="Migration successfully executed" id="Update uid column values in alert_notification" duration=433.937µs grafana | logger=migrator t=2024-04-24T09:37:38.572426596Z level=info msg="Executing migration" id="Add unique index alert_notification_org_id_uid" grafana | logger=migrator t=2024-04-24T09:37:38.57455797Z level=info msg="Migration successfully executed" id="Add unique index alert_notification_org_id_uid" duration=2.130773ms grafana | logger=migrator t=2024-04-24T09:37:38.581671767Z level=info msg="Executing migration" id="Remove unique index org_id_name" grafana | logger=migrator t=2024-04-24T09:37:38.582528841Z level=info msg="Migration successfully executed" id="Remove unique index org_id_name" duration=856.734µs grafana | logger=migrator t=2024-04-24T09:37:38.585947975Z level=info msg="Executing migration" id="Add column secure_settings in alert_notification" grafana | logger=migrator t=2024-04-24T09:37:38.589983132Z level=info msg="Migration successfully executed" id="Add column secure_settings in alert_notification" duration=4.036657ms grafana | logger=migrator t=2024-04-24T09:37:38.594335603Z level=info msg="Executing migration" id="alter alert.settings to mediumtext" grafana | logger=migrator t=2024-04-24T09:37:38.594405746Z level=info msg="Migration successfully executed" id="alter alert.settings to mediumtext" duration=71.493µs grafana | logger=migrator t=2024-04-24T09:37:38.597713075Z level=info msg="Executing migration" id="Add non-unique index alert_notification_state_alert_id" grafana | logger=migrator t=2024-04-24T09:37:38.598555407Z level=info msg="Migration successfully executed" id="Add non-unique index alert_notification_state_alert_id" duration=842.062µs kafka | [2024-04-24 09:38:18,451] INFO [Controller id=1] New topics: [Set(policy-pdp-pap, __consumer_offsets)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(policy-pdp-pap,Some(lm-F3iiTSNeAYdJn1IimsA),Map(policy-pdp-pap-0 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=))), TopicIdReplicaAssignment(__consumer_offsets,Some(kkBxd83QQh-C7hymHduj-A),HashMap(__consumer_offsets-22 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-30 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-25 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-35 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-37 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-38 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-13 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-8 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-21 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-4 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-27 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-7 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-9 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-46 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-41 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-33 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-23 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-49 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-47 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-16 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-28 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-31 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-36 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-42 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-3 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-18 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-15 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-24 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-17 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-48 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-19 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-11 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-2 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-43 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-6 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-14 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-20 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-0 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-44 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-39 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-12 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-45 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-1 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-5 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-26 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-29 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-34 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-10 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-32 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=), __consumer_offsets-40 -> ReplicaAssignment(replicas=1, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController) kafka | [2024-04-24 09:38:18,453] INFO [Controller id=1] New partition creation callback for __consumer_offsets-22,__consumer_offsets-30,__consumer_offsets-25,__consumer_offsets-35,__consumer_offsets-38,__consumer_offsets-13,__consumer_offsets-8,__consumer_offsets-21,__consumer_offsets-4,__consumer_offsets-27,__consumer_offsets-7,__consumer_offsets-9,__consumer_offsets-46,__consumer_offsets-41,__consumer_offsets-33,__consumer_offsets-23,__consumer_offsets-49,__consumer_offsets-47,__consumer_offsets-16,__consumer_offsets-28,__consumer_offsets-31,__consumer_offsets-36,__consumer_offsets-42,__consumer_offsets-3,__consumer_offsets-18,__consumer_offsets-37,policy-pdp-pap-0,__consumer_offsets-15,__consumer_offsets-24,__consumer_offsets-17,__consumer_offsets-48,__consumer_offsets-19,__consumer_offsets-11,__consumer_offsets-2,__consumer_offsets-43,__consumer_offsets-6,__consumer_offsets-14,__consumer_offsets-20,__consumer_offsets-0,__consumer_offsets-44,__consumer_offsets-39,__consumer_offsets-12,__consumer_offsets-45,__consumer_offsets-1,__consumer_offsets-5,__consumer_offsets-26,__consumer_offsets-29,__consumer_offsets-34,__consumer_offsets-10,__consumer_offsets-32,__consumer_offsets-40 (kafka.controller.KafkaController) kafka | [2024-04-24 09:38:18,456] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-22 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,456] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-30 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,456] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-25 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,456] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-35 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,457] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-38 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,457] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-13 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,457] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-8 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,457] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-21 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,457] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-4 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,457] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-27 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,457] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-7 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,457] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-9 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,458] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-46 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,458] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-41 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,458] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-33 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,458] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-23 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,458] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-49 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,458] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-47 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,458] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-16 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,458] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-28 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,458] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-31 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-36 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-42 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-3 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-18 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-37 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition policy-pdp-pap-0 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-15 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-24 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-17 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-48 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,459] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-19 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-11 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-2 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-43 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-6 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-14 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-20 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-0 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-44 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-39 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-12 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-45 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-1 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-5 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-26 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-29 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-34 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-10 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-32 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,460] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-40 state from NonExistentPartition to NewPartition with assigned replicas 1 (state.change.logger) kafka | [2024-04-24 09:38:18,461] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) kafka | [2024-04-24 09:38:18,467] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-32 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,467] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-5 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,467] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-44 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,467] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-48 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,467] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-46 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,467] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-20 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,467] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition policy-pdp-pap-0 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,467] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-43 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,467] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-24 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-6 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-18 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-21 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-1 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-14 from NonExistentReplica to NewReplica (state.change.logger) policy-pap | max.poll.interval.ms = 300000 policy-pap | max.poll.records = 500 policy-pap | metadata.max.age.ms = 300000 policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-pap | receive.buffer.bytes = 65536 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | session.timeout.ms = 45000 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-34 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-16 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-29 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-11 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-0 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-22 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,468] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-47 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-36 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-28 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-42 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-9 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-37 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-13 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-30 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-35 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-39 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-12 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,469] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-27 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-45 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-19 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-49 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-40 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-41 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-38 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-8 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-7 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-33 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-25 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-31 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-23 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-10 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-2 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,470] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-17 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,471] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-4 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,471] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-15 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,471] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-26 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,471] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-3 from NonExistentReplica to NewReplica (state.change.logger) kafka | [2024-04-24 09:38:18,471] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) kafka | [2024-04-24 09:38:18,629] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-22 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,629] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-30 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 grafana | logger=migrator t=2024-04-24T09:37:38.602847746Z level=info msg="Executing migration" id="Add non-unique index alert_rule_tag_alert_id" grafana | logger=migrator t=2024-04-24T09:37:38.603883576Z level=info msg="Migration successfully executed" id="Add non-unique index alert_rule_tag_alert_id" duration=1.035ms grafana | logger=migrator t=2024-04-24T09:37:38.609288237Z level=info msg="Executing migration" id="Drop old annotation table v4" grafana | logger=migrator t=2024-04-24T09:37:38.609384141Z level=info msg="Migration successfully executed" id="Drop old annotation table v4" duration=96.634µs grafana | logger=migrator t=2024-04-24T09:37:38.6157544Z level=info msg="Executing migration" id="create annotation table v5" grafana | logger=migrator t=2024-04-24T09:37:38.617354473Z level=info msg="Migration successfully executed" id="create annotation table v5" duration=1.599202ms grafana | logger=migrator t=2024-04-24T09:37:38.624355607Z level=info msg="Executing migration" id="add index annotation 0 v3" grafana | logger=migrator t=2024-04-24T09:37:38.625667028Z level=info msg="Migration successfully executed" id="add index annotation 0 v3" duration=1.31062ms grafana | logger=migrator t=2024-04-24T09:37:38.628926815Z level=info msg="Executing migration" id="add index annotation 1 v3" grafana | logger=migrator t=2024-04-24T09:37:38.629795339Z level=info msg="Migration successfully executed" id="add index annotation 1 v3" duration=868.374µs grafana | logger=migrator t=2024-04-24T09:37:38.636342595Z level=info msg="Executing migration" id="add index annotation 2 v3" grafana | logger=migrator t=2024-04-24T09:37:38.63724688Z level=info msg="Migration successfully executed" id="add index annotation 2 v3" duration=904.465µs grafana | logger=migrator t=2024-04-24T09:37:38.642723294Z level=info msg="Executing migration" id="add index annotation 3 v3" grafana | logger=migrator t=2024-04-24T09:37:38.643722703Z level=info msg="Migration successfully executed" id="add index annotation 3 v3" duration=995.859µs grafana | logger=migrator t=2024-04-24T09:37:38.647115917Z level=info msg="Executing migration" id="add index annotation 4 v3" grafana | logger=migrator t=2024-04-24T09:37:38.648054673Z level=info msg="Migration successfully executed" id="add index annotation 4 v3" duration=942.137µs grafana | logger=migrator t=2024-04-24T09:37:38.654739254Z level=info msg="Executing migration" id="Update annotation table charset" grafana | logger=migrator t=2024-04-24T09:37:38.654801456Z level=info msg="Migration successfully executed" id="Update annotation table charset" duration=62.152µs grafana | logger=migrator t=2024-04-24T09:37:38.657879588Z level=info msg="Executing migration" id="Add column region_id to annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.66257335Z level=info msg="Migration successfully executed" id="Add column region_id to annotation table" duration=4.695043ms grafana | logger=migrator t=2024-04-24T09:37:38.665719143Z level=info msg="Executing migration" id="Drop category_id index" grafana | logger=migrator t=2024-04-24T09:37:38.666563967Z level=info msg="Migration successfully executed" id="Drop category_id index" duration=843.684µs grafana | logger=migrator t=2024-04-24T09:37:38.671960168Z level=info msg="Executing migration" id="Add column tags to annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.675894842Z level=info msg="Migration successfully executed" id="Add column tags to annotation table" duration=3.934113ms grafana | logger=migrator t=2024-04-24T09:37:38.681861005Z level=info msg="Executing migration" id="Create annotation_tag table v2" grafana | logger=migrator t=2024-04-24T09:37:38.68277962Z level=info msg="Migration successfully executed" id="Create annotation_tag table v2" duration=919.235µs grafana | logger=migrator t=2024-04-24T09:37:38.685867432Z level=info msg="Executing migration" id="Add unique index annotation_tag.annotation_id_tag_id" grafana | logger=migrator t=2024-04-24T09:37:38.686806058Z level=info msg="Migration successfully executed" id="Add unique index annotation_tag.annotation_id_tag_id" duration=937.916µs grafana | logger=migrator t=2024-04-24T09:37:38.69169904Z level=info msg="Executing migration" id="drop index UQE_annotation_tag_annotation_id_tag_id - v2" grafana | logger=migrator t=2024-04-24T09:37:38.692463509Z level=info msg="Migration successfully executed" id="drop index UQE_annotation_tag_annotation_id_tag_id - v2" duration=764.039µs grafana | logger=migrator t=2024-04-24T09:37:38.69553605Z level=info msg="Executing migration" id="Rename table annotation_tag to annotation_tag_v2 - v2" grafana | logger=migrator t=2024-04-24T09:37:38.7093528Z level=info msg="Migration successfully executed" id="Rename table annotation_tag to annotation_tag_v2 - v2" duration=13.815731ms grafana | logger=migrator t=2024-04-24T09:37:38.712854526Z level=info msg="Executing migration" id="Create annotation_tag table v3" grafana | logger=migrator t=2024-04-24T09:37:38.71345189Z level=info msg="Migration successfully executed" id="Create annotation_tag table v3" duration=598.144µs grafana | logger=migrator t=2024-04-24T09:37:38.719211105Z level=info msg="Executing migration" id="create index UQE_annotation_tag_annotation_id_tag_id - Add unique index annotation_tag.annotation_id_tag_id V3" grafana | logger=migrator t=2024-04-24T09:37:38.719857401Z level=info msg="Migration successfully executed" id="create index UQE_annotation_tag_annotation_id_tag_id - Add unique index annotation_tag.annotation_id_tag_id V3" duration=642.016µs grafana | logger=migrator t=2024-04-24T09:37:38.728687676Z level=info msg="Executing migration" id="copy annotation_tag v2 to v3" grafana | logger=migrator t=2024-04-24T09:37:38.729377023Z level=info msg="Migration successfully executed" id="copy annotation_tag v2 to v3" duration=690.066µs grafana | logger=migrator t=2024-04-24T09:37:38.732875639Z level=info msg="Executing migration" id="drop table annotation_tag_v2" grafana | logger=migrator t=2024-04-24T09:37:38.733549036Z level=info msg="Migration successfully executed" id="drop table annotation_tag_v2" duration=670.187µs grafana | logger=migrator t=2024-04-24T09:37:38.739332002Z level=info msg="Executing migration" id="Update alert annotations and set TEXT to empty" grafana | logger=migrator t=2024-04-24T09:37:38.739463437Z level=info msg="Migration successfully executed" id="Update alert annotations and set TEXT to empty" duration=131.885µs grafana | logger=migrator t=2024-04-24T09:37:38.74259449Z level=info msg="Executing migration" id="Add created time to annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.745616208Z level=info msg="Migration successfully executed" id="Add created time to annotation table" duration=3.021978ms grafana | logger=migrator t=2024-04-24T09:37:38.75104538Z level=info msg="Executing migration" id="Add updated time to annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.755724392Z level=info msg="Migration successfully executed" id="Add updated time to annotation table" duration=4.682082ms grafana | logger=migrator t=2024-04-24T09:37:38.761559081Z level=info msg="Executing migration" id="Add index for created in annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.762647844Z level=info msg="Migration successfully executed" id="Add index for created in annotation table" duration=1.090332ms grafana | logger=migrator t=2024-04-24T09:37:38.767490053Z level=info msg="Executing migration" id="Add index for updated in annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.768235812Z level=info msg="Migration successfully executed" id="Add index for updated in annotation table" duration=742.639µs grafana | logger=migrator t=2024-04-24T09:37:38.771798561Z level=info msg="Executing migration" id="Convert existing annotations from seconds to milliseconds" grafana | logger=migrator t=2024-04-24T09:37:38.771992399Z level=info msg="Migration successfully executed" id="Convert existing annotations from seconds to milliseconds" duration=194.377µs grafana | logger=migrator t=2024-04-24T09:37:38.777020565Z level=info msg="Executing migration" id="Add epoch_end column" grafana | logger=migrator t=2024-04-24T09:37:38.783136544Z level=info msg="Migration successfully executed" id="Add epoch_end column" duration=6.116239ms grafana | logger=migrator t=2024-04-24T09:37:38.789477142Z level=info msg="Executing migration" id="Add index for epoch_end" grafana | logger=migrator t=2024-04-24T09:37:38.790521573Z level=info msg="Migration successfully executed" id="Add index for epoch_end" duration=1.046751ms grafana | logger=migrator t=2024-04-24T09:37:38.796270348Z level=info msg="Executing migration" id="Make epoch_end the same as epoch" grafana | logger=migrator t=2024-04-24T09:37:38.796448225Z level=info msg="Migration successfully executed" id="Make epoch_end the same as epoch" duration=178.657µs grafana | logger=migrator t=2024-04-24T09:37:38.79964887Z level=info msg="Executing migration" id="Move region to single row" grafana | logger=migrator t=2024-04-24T09:37:38.800050716Z level=info msg="Migration successfully executed" id="Move region to single row" duration=401.695µs grafana | logger=migrator t=2024-04-24T09:37:38.803112645Z level=info msg="Executing migration" id="Remove index org_id_epoch from annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.80424446Z level=info msg="Migration successfully executed" id="Remove index org_id_epoch from annotation table" duration=1.130775ms grafana | logger=migrator t=2024-04-24T09:37:38.82163787Z level=info msg="Executing migration" id="Remove index org_id_dashboard_id_panel_id_epoch from annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.823440741Z level=info msg="Migration successfully executed" id="Remove index org_id_dashboard_id_panel_id_epoch from annotation table" duration=1.811162ms grafana | logger=migrator t=2024-04-24T09:37:38.832030206Z level=info msg="Executing migration" id="Add index for org_id_dashboard_id_epoch_end_epoch on annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.833082067Z level=info msg="Migration successfully executed" id="Add index for org_id_dashboard_id_epoch_end_epoch on annotation table" duration=1.051231ms grafana | logger=migrator t=2024-04-24T09:37:38.839023959Z level=info msg="Executing migration" id="Add index for org_id_epoch_end_epoch on annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.840417354Z level=info msg="Migration successfully executed" id="Add index for org_id_epoch_end_epoch on annotation table" duration=1.393545ms grafana | logger=migrator t=2024-04-24T09:37:38.845561086Z level=info msg="Executing migration" id="Remove index org_id_epoch_epoch_end from annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.846833935Z level=info msg="Migration successfully executed" id="Remove index org_id_epoch_epoch_end from annotation table" duration=1.27731ms grafana | logger=migrator t=2024-04-24T09:37:38.85617079Z level=info msg="Executing migration" id="Add index for alert_id on annotation table" grafana | logger=migrator t=2024-04-24T09:37:38.857654438Z level=info msg="Migration successfully executed" id="Add index for alert_id on annotation table" duration=1.482788ms grafana | logger=migrator t=2024-04-24T09:37:38.862524289Z level=info msg="Executing migration" id="Increase tags column to length 4096" grafana | logger=migrator t=2024-04-24T09:37:38.862634623Z level=info msg="Migration successfully executed" id="Increase tags column to length 4096" duration=111.204µs grafana | logger=migrator t=2024-04-24T09:37:38.866174021Z level=info msg="Executing migration" id="create test_data table" grafana | logger=migrator t=2024-04-24T09:37:38.867076246Z level=info msg="Migration successfully executed" id="create test_data table" duration=903.695µs grafana | logger=migrator t=2024-04-24T09:37:38.874197485Z level=info msg="Executing migration" id="create dashboard_version table v1" grafana | logger=migrator t=2024-04-24T09:37:38.875483635Z level=info msg="Migration successfully executed" id="create dashboard_version table v1" duration=1.28577ms grafana | logger=migrator t=2024-04-24T09:37:38.881894935Z level=info msg="Executing migration" id="add index dashboard_version.dashboard_id" grafana | logger=migrator t=2024-04-24T09:37:38.883404635Z level=info msg="Migration successfully executed" id="add index dashboard_version.dashboard_id" duration=1.50518ms grafana | logger=migrator t=2024-04-24T09:37:38.888416681Z level=info msg="Executing migration" id="add unique index dashboard_version.dashboard_id and dashboard_version.version" grafana | logger=migrator t=2024-04-24T09:37:38.889824876Z level=info msg="Migration successfully executed" id="add unique index dashboard_version.dashboard_id and dashboard_version.version" duration=1.408314ms grafana | logger=migrator t=2024-04-24T09:37:38.89454586Z level=info msg="Executing migration" id="Set dashboard version to 1 where 0" grafana | logger=migrator t=2024-04-24T09:37:38.894731728Z level=info msg="Migration successfully executed" id="Set dashboard version to 1 where 0" duration=185.568µs grafana | logger=migrator t=2024-04-24T09:37:38.899501674Z level=info msg="Executing migration" id="save existing dashboard data in dashboard_version table v1" grafana | logger=migrator t=2024-04-24T09:37:38.899862558Z level=info msg="Migration successfully executed" id="save existing dashboard data in dashboard_version table v1" duration=360.904µs grafana | logger=migrator t=2024-04-24T09:37:38.906091172Z level=info msg="Executing migration" id="alter dashboard_version.data to mediumtext v1" grafana | logger=migrator t=2024-04-24T09:37:38.906165225Z level=info msg="Migration successfully executed" id="alter dashboard_version.data to mediumtext v1" duration=74.593µs grafana | logger=migrator t=2024-04-24T09:37:38.911066386Z level=info msg="Executing migration" id="create team table" grafana | logger=migrator t=2024-04-24T09:37:38.912851207Z level=info msg="Migration successfully executed" id="create team table" duration=1.77979ms grafana | logger=migrator t=2024-04-24T09:37:38.919286168Z level=info msg="Executing migration" id="add index team.org_id" grafana | logger=migrator t=2024-04-24T09:37:38.920415562Z level=info msg="Migration successfully executed" id="add index team.org_id" duration=1.128644ms grafana | logger=migrator t=2024-04-24T09:37:38.925054193Z level=info msg="Executing migration" id="add unique index team_org_id_name" grafana | logger=migrator t=2024-04-24T09:37:38.925922787Z level=info msg="Migration successfully executed" id="add unique index team_org_id_name" duration=867.624µs grafana | logger=migrator t=2024-04-24T09:37:38.929942865Z level=info msg="Executing migration" id="Add column uid in team" grafana | logger=migrator t=2024-04-24T09:37:38.934530354Z level=info msg="Migration successfully executed" id="Add column uid in team" duration=4.583728ms grafana | logger=migrator t=2024-04-24T09:37:38.939690535Z level=info msg="Executing migration" id="Update uid column values in team" grafana | logger=migrator t=2024-04-24T09:37:38.939869342Z level=info msg="Migration successfully executed" id="Update uid column values in team" duration=178.387µs grafana | logger=migrator t=2024-04-24T09:37:38.944048386Z level=info msg="Executing migration" id="Add unique index team_org_id_uid" grafana | logger=migrator t=2024-04-24T09:37:38.945025754Z level=info msg="Migration successfully executed" id="Add unique index team_org_id_uid" duration=976.648µs grafana | logger=migrator t=2024-04-24T09:37:38.949301002Z level=info msg="Executing migration" id="create team member table" grafana | logger=migrator t=2024-04-24T09:37:38.950496148Z level=info msg="Migration successfully executed" id="create team member table" duration=1.193656ms grafana | logger=migrator t=2024-04-24T09:37:38.956904428Z level=info msg="Executing migration" id="add index team_member.org_id" grafana | logger=migrator t=2024-04-24T09:37:38.958256162Z level=info msg="Migration successfully executed" id="add index team_member.org_id" duration=1.354684ms grafana | logger=migrator t=2024-04-24T09:37:38.962907133Z level=info msg="Executing migration" id="add unique index team_member_org_id_team_id_user_id" grafana | logger=migrator t=2024-04-24T09:37:38.964911962Z level=info msg="Migration successfully executed" id="add unique index team_member_org_id_team_id_user_id" duration=2.004238ms grafana | logger=migrator t=2024-04-24T09:37:38.970832903Z level=info msg="Executing migration" id="add index team_member.team_id" grafana | logger=migrator t=2024-04-24T09:37:38.972826411Z level=info msg="Migration successfully executed" id="add index team_member.team_id" duration=1.992978ms grafana | logger=migrator t=2024-04-24T09:37:38.978761463Z level=info msg="Executing migration" id="Add column email to team table" grafana | logger=migrator t=2024-04-24T09:37:38.984007138Z level=info msg="Migration successfully executed" id="Add column email to team table" duration=5.244935ms grafana | logger=migrator t=2024-04-24T09:37:38.987691382Z level=info msg="Executing migration" id="Add column external to team_member table" grafana | logger=migrator t=2024-04-24T09:37:38.993383455Z level=info msg="Migration successfully executed" id="Add column external to team_member table" duration=5.691152ms grafana | logger=migrator t=2024-04-24T09:37:38.998308427Z level=info msg="Executing migration" id="Add column permission to team_member table" grafana | logger=migrator t=2024-04-24T09:37:39.002844835Z level=info msg="Migration successfully executed" id="Add column permission to team_member table" duration=4.531828ms grafana | logger=migrator t=2024-04-24T09:37:39.007037476Z level=info msg="Executing migration" id="create dashboard acl table" grafana | logger=migrator t=2024-04-24T09:37:39.007980874Z level=info msg="Migration successfully executed" id="create dashboard acl table" duration=943.428µs grafana | logger=migrator t=2024-04-24T09:37:39.013430622Z level=info msg="Executing migration" id="add index dashboard_acl_dashboard_id" grafana | logger=migrator t=2024-04-24T09:37:39.014336419Z level=info msg="Migration successfully executed" id="add index dashboard_acl_dashboard_id" duration=905.947µs grafana | logger=migrator t=2024-04-24T09:37:39.020060738Z level=info msg="Executing migration" id="add unique index dashboard_acl_dashboard_id_user_id" grafana | logger=migrator t=2024-04-24T09:37:39.022317939Z level=info msg="Migration successfully executed" id="add unique index dashboard_acl_dashboard_id_user_id" duration=2.256081ms grafana | logger=migrator t=2024-04-24T09:37:39.029386901Z level=info msg="Executing migration" id="add unique index dashboard_acl_dashboard_id_team_id" grafana | logger=migrator t=2024-04-24T09:37:39.030308038Z level=info msg="Migration successfully executed" id="add unique index dashboard_acl_dashboard_id_team_id" duration=920.297µs grafana | logger=migrator t=2024-04-24T09:37:39.034949204Z level=info msg="Executing migration" id="add index dashboard_acl_user_id" grafana | logger=migrator t=2024-04-24T09:37:39.036485696Z level=info msg="Migration successfully executed" id="add index dashboard_acl_user_id" duration=1.555953ms grafana | logger=migrator t=2024-04-24T09:37:39.04332466Z level=info msg="Executing migration" id="add index dashboard_acl_team_id" grafana | logger=migrator t=2024-04-24T09:37:39.044785079Z level=info msg="Migration successfully executed" id="add index dashboard_acl_team_id" duration=1.460449ms grafana | logger=migrator t=2024-04-24T09:37:39.050348452Z level=info msg="Executing migration" id="add index dashboard_acl_org_id_role" grafana | logger=migrator t=2024-04-24T09:37:39.051897503Z level=info msg="Migration successfully executed" id="add index dashboard_acl_org_id_role" duration=1.547742ms grafana | logger=migrator t=2024-04-24T09:37:39.057384653Z level=info msg="Executing migration" id="add index dashboard_permission" grafana | logger=migrator t=2024-04-24T09:37:39.05904407Z level=info msg="Migration successfully executed" id="add index dashboard_permission" duration=1.660087ms grafana | logger=migrator t=2024-04-24T09:37:39.063126823Z level=info msg="Executing migration" id="save default acl rules in dashboard_acl table" grafana | logger=migrator t=2024-04-24T09:37:39.063633733Z level=info msg="Migration successfully executed" id="save default acl rules in dashboard_acl table" duration=506.93µs policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | policy-pap | [2024-04-24T09:38:17.891+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 policy-pap | [2024-04-24T09:38:17.891+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 policy-pap | [2024-04-24T09:38:17.891+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1713951497891 policy-pap | [2024-04-24T09:38:17.891+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Subscribed to topic(s): policy-pdp-pap policy-pap | [2024-04-24T09:38:17.892+00:00|INFO|ServiceManager|main] Policy PAP starting Heartbeat Message Dispatcher policy-pap | [2024-04-24T09:38:17.892+00:00|INFO|TopicBase|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=policy-pap, consumerInstance=36ee11dc-f044-4dbb-89eb-a07ca9cc5bb6, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=0, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-heartbeat, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=0]]]]: registering org.onap.policy.common.endpoints.listeners.MessageTypeDispatcher@160e45c8 policy-pap | [2024-04-24T09:38:17.893+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=policy-pap, consumerInstance=36ee11dc-f044-4dbb-89eb-a07ca9cc5bb6, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=false, locked=false, uebThread=null, topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-heartbeat, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-pap | [2024-04-24T09:38:17.893+00:00|INFO|ConsumerConfig|main] ConsumerConfig values: policy-pap | allow.auto.create.topics = true policy-pap | auto.commit.interval.ms = 5000 policy-pap | auto.include.jmx.reporter = true policy-pap | auto.offset.reset = latest policy-pap | bootstrap.servers = [kafka:9092] policy-pap | check.crcs = true policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = consumer-policy-pap-4 policy-pap | client.rack = policy-pap | connections.max.idle.ms = 540000 policy-pap | default.api.timeout.ms = 60000 policy-pap | enable.auto.commit = true policy-pap | exclude.internal.topics = true policy-pap | fetch.max.bytes = 52428800 policy-pap | fetch.max.wait.ms = 500 policy-pap | fetch.min.bytes = 1 policy-pap | group.id = policy-pap policy-pap | group.instance.id = null policy-pap | heartbeat.interval.ms = 3000 policy-pap | interceptor.classes = [] policy-pap | internal.leave.group.on.close = true policy-pap | internal.throw.on.fetch.stable.offset.unsupported = false grafana | logger=migrator t=2024-04-24T09:37:39.070359213Z level=info msg="Executing migration" id="delete acl rules for deleted dashboards and folders" grafana | logger=migrator t=2024-04-24T09:37:39.070717537Z level=info msg="Migration successfully executed" id="delete acl rules for deleted dashboards and folders" duration=361.524µs grafana | logger=migrator t=2024-04-24T09:37:39.075404105Z level=info msg="Executing migration" id="create tag table" grafana | logger=migrator t=2024-04-24T09:37:39.076962947Z level=info msg="Migration successfully executed" id="create tag table" duration=1.558222ms grafana | logger=migrator t=2024-04-24T09:37:39.084425916Z level=info msg="Executing migration" id="add index tag.key_value" grafana | logger=migrator t=2024-04-24T09:37:39.085363523Z level=info msg="Migration successfully executed" id="add index tag.key_value" duration=936.617µs grafana | logger=migrator t=2024-04-24T09:37:39.093132865Z level=info msg="Executing migration" id="create login attempt table" grafana | logger=migrator t=2024-04-24T09:37:39.093855013Z level=info msg="Migration successfully executed" id="create login attempt table" duration=722.038µs grafana | logger=migrator t=2024-04-24T09:37:39.098606954Z level=info msg="Executing migration" id="add index login_attempt.username" grafana | logger=migrator t=2024-04-24T09:37:39.09951918Z level=info msg="Migration successfully executed" id="add index login_attempt.username" duration=907.736µs grafana | logger=migrator t=2024-04-24T09:37:39.104109075Z level=info msg="Executing migration" id="drop index IDX_login_attempt_username - v1" grafana | logger=migrator t=2024-04-24T09:37:39.105110064Z level=info msg="Migration successfully executed" id="drop index IDX_login_attempt_username - v1" duration=995.23µs grafana | logger=migrator t=2024-04-24T09:37:39.110983139Z level=info msg="Executing migration" id="Rename table login_attempt to login_attempt_tmp_qwerty - v1" grafana | logger=migrator t=2024-04-24T09:37:39.12697115Z level=info msg="Migration successfully executed" id="Rename table login_attempt to login_attempt_tmp_qwerty - v1" duration=16.077214ms grafana | logger=migrator t=2024-04-24T09:37:39.130813434Z level=info msg="Executing migration" id="create login_attempt v2" grafana | logger=migrator t=2024-04-24T09:37:39.131403858Z level=info msg="Migration successfully executed" id="create login_attempt v2" duration=589.984µs grafana | logger=migrator t=2024-04-24T09:37:39.1379538Z level=info msg="Executing migration" id="create index IDX_login_attempt_username - v2" grafana | logger=migrator t=2024-04-24T09:37:39.138854156Z level=info msg="Migration successfully executed" id="create index IDX_login_attempt_username - v2" duration=899.766µs grafana | logger=migrator t=2024-04-24T09:37:39.144078715Z level=info msg="Executing migration" id="copy login_attempt v1 to v2" grafana | logger=migrator t=2024-04-24T09:37:39.144416339Z level=info msg="Migration successfully executed" id="copy login_attempt v1 to v2" duration=335.753µs grafana | logger=migrator t=2024-04-24T09:37:39.151040504Z level=info msg="Executing migration" id="drop login_attempt_tmp_qwerty" grafana | logger=migrator t=2024-04-24T09:37:39.152240572Z level=info msg="Migration successfully executed" id="drop login_attempt_tmp_qwerty" duration=1.197118ms grafana | logger=migrator t=2024-04-24T09:37:39.155723461Z level=info msg="Executing migration" id="create user auth table" grafana | logger=migrator t=2024-04-24T09:37:39.15693463Z level=info msg="Migration successfully executed" id="create user auth table" duration=1.210729ms grafana | logger=migrator t=2024-04-24T09:37:39.162682061Z level=info msg="Executing migration" id="create index IDX_user_auth_auth_module_auth_id - v1" grafana | logger=migrator t=2024-04-24T09:37:39.164193171Z level=info msg="Migration successfully executed" id="create index IDX_user_auth_auth_module_auth_id - v1" duration=1.50908ms grafana | logger=migrator t=2024-04-24T09:37:39.169210992Z level=info msg="Executing migration" id="alter user_auth.auth_id to length 190" grafana | logger=migrator t=2024-04-24T09:37:39.169315406Z level=info msg="Migration successfully executed" id="alter user_auth.auth_id to length 190" duration=105.024µs grafana | logger=migrator t=2024-04-24T09:37:39.176420621Z level=info msg="Executing migration" id="Add OAuth access token to user_auth" grafana | logger=migrator t=2024-04-24T09:37:39.181556186Z level=info msg="Migration successfully executed" id="Add OAuth access token to user_auth" duration=5.134355ms grafana | logger=migrator t=2024-04-24T09:37:39.187227873Z level=info msg="Executing migration" id="Add OAuth refresh token to user_auth" grafana | logger=migrator t=2024-04-24T09:37:39.192690402Z level=info msg="Migration successfully executed" id="Add OAuth refresh token to user_auth" duration=5.463209ms grafana | logger=migrator t=2024-04-24T09:37:39.197706593Z level=info msg="Executing migration" id="Add OAuth token type to user_auth" grafana | logger=migrator t=2024-04-24T09:37:39.203034846Z level=info msg="Migration successfully executed" id="Add OAuth token type to user_auth" duration=5.327463ms grafana | logger=migrator t=2024-04-24T09:37:39.212910702Z level=info msg="Executing migration" id="Add OAuth expiry to user_auth" grafana | logger=migrator t=2024-04-24T09:37:39.218445624Z level=info msg="Migration successfully executed" id="Add OAuth expiry to user_auth" duration=5.531552ms grafana | logger=migrator t=2024-04-24T09:37:39.231965075Z level=info msg="Executing migration" id="Add index to user_id column in user_auth" grafana | logger=migrator t=2024-04-24T09:37:39.233557109Z level=info msg="Migration successfully executed" id="Add index to user_id column in user_auth" duration=1.591634ms grafana | logger=migrator t=2024-04-24T09:37:39.238687095Z level=info msg="Executing migration" id="Add OAuth ID token to user_auth" grafana | logger=migrator t=2024-04-24T09:37:39.245230666Z level=info msg="Migration successfully executed" id="Add OAuth ID token to user_auth" duration=6.544362ms grafana | logger=migrator t=2024-04-24T09:37:39.251013418Z level=info msg="Executing migration" id="create server_lock table" grafana | logger=migrator t=2024-04-24T09:37:39.251977257Z level=info msg="Migration successfully executed" id="create server_lock table" duration=962.289µs grafana | logger=migrator t=2024-04-24T09:37:39.256991487Z level=info msg="Executing migration" id="add index server_lock.operation_uid" grafana | logger=migrator t=2024-04-24T09:37:39.258008029Z level=info msg="Migration successfully executed" id="add index server_lock.operation_uid" duration=1.017701ms grafana | logger=migrator t=2024-04-24T09:37:39.261661045Z level=info msg="Executing migration" id="create user auth token table" grafana | logger=migrator t=2024-04-24T09:37:39.26255193Z level=info msg="Migration successfully executed" id="create user auth token table" duration=892.356µs grafana | logger=migrator t=2024-04-24T09:37:39.272155165Z level=info msg="Executing migration" id="add unique index user_auth_token.auth_token" grafana | logger=migrator t=2024-04-24T09:37:39.273164025Z level=info msg="Migration successfully executed" id="add unique index user_auth_token.auth_token" duration=1.00829ms grafana | logger=migrator t=2024-04-24T09:37:39.279148905Z level=info msg="Executing migration" id="add unique index user_auth_token.prev_auth_token" grafana | logger=migrator t=2024-04-24T09:37:39.280746089Z level=info msg="Migration successfully executed" id="add unique index user_auth_token.prev_auth_token" duration=1.564563ms grafana | logger=migrator t=2024-04-24T09:37:39.284742239Z level=info msg="Executing migration" id="add index user_auth_token.user_id" grafana | logger=migrator t=2024-04-24T09:37:39.286497819Z level=info msg="Migration successfully executed" id="add index user_auth_token.user_id" duration=1.780911ms policy-pap | isolation.level = read_uncommitted policy-pap | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | max.partition.fetch.bytes = 1048576 policy-pap | max.poll.interval.ms = 300000 policy-pap | max.poll.records = 500 policy-pap | metadata.max.age.ms = 300000 policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor] policy-pap | receive.buffer.bytes = 65536 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | session.timeout.ms = 45000 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS grafana | logger=migrator t=2024-04-24T09:37:39.291293722Z level=info msg="Executing migration" id="Add revoked_at to the user auth token" grafana | logger=migrator t=2024-04-24T09:37:39.297261521Z level=info msg="Migration successfully executed" id="Add revoked_at to the user auth token" duration=5.963718ms grafana | logger=migrator t=2024-04-24T09:37:39.303121445Z level=info msg="Executing migration" id="add index user_auth_token.revoked_at" grafana | logger=migrator t=2024-04-24T09:37:39.304793002Z level=info msg="Migration successfully executed" id="add index user_auth_token.revoked_at" duration=1.670767ms grafana | logger=migrator t=2024-04-24T09:37:39.310539222Z level=info msg="Executing migration" id="create cache_data table" grafana | logger=migrator t=2024-04-24T09:37:39.312121416Z level=info msg="Migration successfully executed" id="create cache_data table" duration=1.581613ms grafana | logger=migrator t=2024-04-24T09:37:39.317327524Z level=info msg="Executing migration" id="add unique index cache_data.cache_key" grafana | logger=migrator t=2024-04-24T09:37:39.318328884Z level=info msg="Migration successfully executed" id="add unique index cache_data.cache_key" duration=1.00084ms grafana | logger=migrator t=2024-04-24T09:37:39.324361065Z level=info msg="Executing migration" id="create short_url table v1" grafana | logger=migrator t=2024-04-24T09:37:39.325828545Z level=info msg="Migration successfully executed" id="create short_url table v1" duration=1.46679ms grafana | logger=migrator t=2024-04-24T09:37:39.329775562Z level=info msg="Executing migration" id="add index short_url.org_id-uid" grafana | logger=migrator t=2024-04-24T09:37:39.331028913Z level=info msg="Migration successfully executed" id="add index short_url.org_id-uid" duration=1.253571ms grafana | logger=migrator t=2024-04-24T09:37:39.336368037Z level=info msg="Executing migration" id="alter table short_url alter column created_by type to bigint" grafana | logger=migrator t=2024-04-24T09:37:39.336481342Z level=info msg="Migration successfully executed" id="alter table short_url alter column created_by type to bigint" duration=114.664µs grafana | logger=migrator t=2024-04-24T09:37:39.342966631Z level=info msg="Executing migration" id="delete alert_definition table" grafana | logger=migrator t=2024-04-24T09:37:39.343111577Z level=info msg="Migration successfully executed" id="delete alert_definition table" duration=145.926µs grafana | logger=migrator t=2024-04-24T09:37:39.350244583Z level=info msg="Executing migration" id="recreate alert_definition table" grafana | logger=migrator t=2024-04-24T09:37:39.351879858Z level=info msg="Migration successfully executed" id="recreate alert_definition table" duration=1.631645ms grafana | logger=migrator t=2024-04-24T09:37:39.356977432Z level=info msg="Executing migration" id="add index in alert_definition on org_id and title columns" grafana | logger=migrator t=2024-04-24T09:37:39.358000303Z level=info msg="Migration successfully executed" id="add index in alert_definition on org_id and title columns" duration=1.020851ms grafana | logger=migrator t=2024-04-24T09:37:39.365782325Z level=info msg="Executing migration" id="add index in alert_definition on org_id and uid columns" grafana | logger=migrator t=2024-04-24T09:37:39.366778405Z level=info msg="Migration successfully executed" id="add index in alert_definition on org_id and uid columns" duration=995.97µs grafana | logger=migrator t=2024-04-24T09:37:39.371238153Z level=info msg="Executing migration" id="alter alert_definition table data column to mediumtext in mysql" grafana | logger=migrator t=2024-04-24T09:37:39.371305736Z level=info msg="Migration successfully executed" id="alter alert_definition table data column to mediumtext in mysql" duration=72.203µs grafana | logger=migrator t=2024-04-24T09:37:39.378678642Z level=info msg="Executing migration" id="drop index in alert_definition on org_id and title columns" grafana | logger=migrator t=2024-04-24T09:37:39.380190712Z level=info msg="Migration successfully executed" id="drop index in alert_definition on org_id and title columns" duration=1.51339ms grafana | logger=migrator t=2024-04-24T09:37:39.385524775Z level=info msg="Executing migration" id="drop index in alert_definition on org_id and uid columns" grafana | logger=migrator t=2024-04-24T09:37:39.386401771Z level=info msg="Migration successfully executed" id="drop index in alert_definition on org_id and uid columns" duration=877.256µs grafana | logger=migrator t=2024-04-24T09:37:39.393655301Z level=info msg="Executing migration" id="add unique index in alert_definition on org_id and title columns" grafana | logger=migrator t=2024-04-24T09:37:39.395200953Z level=info msg="Migration successfully executed" id="add unique index in alert_definition on org_id and title columns" duration=1.545052ms grafana | logger=migrator t=2024-04-24T09:37:39.401288397Z level=info msg="Executing migration" id="add unique index in alert_definition on org_id and uid columns" grafana | logger=migrator t=2024-04-24T09:37:39.402492125Z level=info msg="Migration successfully executed" id="add unique index in alert_definition on org_id and uid columns" duration=1.204128ms grafana | logger=migrator t=2024-04-24T09:37:39.409025817Z level=info msg="Executing migration" id="Add column paused in alert_definition" grafana | logger=migrator t=2024-04-24T09:37:39.417000426Z level=info msg="Migration successfully executed" id="Add column paused in alert_definition" duration=7.971808ms grafana | logger=migrator t=2024-04-24T09:37:39.420772048Z level=info msg="Executing migration" id="drop alert_definition table" grafana | logger=migrator t=2024-04-24T09:37:39.421705985Z level=info msg="Migration successfully executed" id="drop alert_definition table" duration=933.437µs grafana | logger=migrator t=2024-04-24T09:37:39.42510289Z level=info msg="Executing migration" id="delete alert_definition_version table" grafana | logger=migrator t=2024-04-24T09:37:39.425186514Z level=info msg="Migration successfully executed" id="delete alert_definition_version table" duration=84.004µs grafana | logger=migrator t=2024-04-24T09:37:39.430863771Z level=info msg="Executing migration" id="recreate alert_definition_version table" grafana | logger=migrator t=2024-04-24T09:37:39.431903743Z level=info msg="Migration successfully executed" id="recreate alert_definition_version table" duration=1.038492ms grafana | logger=migrator t=2024-04-24T09:37:39.437284409Z level=info msg="Executing migration" id="add index in alert_definition_version table on alert_definition_id and version columns" grafana | logger=migrator t=2024-04-24T09:37:39.438835771Z level=info msg="Migration successfully executed" id="add index in alert_definition_version table on alert_definition_id and version columns" duration=1.551491ms grafana | logger=migrator t=2024-04-24T09:37:39.46180469Z level=info msg="Executing migration" id="add index in alert_definition_version table on alert_definition_uid and version columns" grafana | logger=migrator t=2024-04-24T09:37:39.464010769Z level=info msg="Migration successfully executed" id="add index in alert_definition_version table on alert_definition_uid and version columns" duration=2.208729ms grafana | logger=migrator t=2024-04-24T09:37:39.470804502Z level=info msg="Executing migration" id="alter alert_definition_version table data column to mediumtext in mysql" grafana | logger=migrator t=2024-04-24T09:37:39.470903035Z level=info msg="Migration successfully executed" id="alter alert_definition_version table data column to mediumtext in mysql" duration=98.744µs grafana | logger=migrator t=2024-04-24T09:37:39.476202088Z level=info msg="Executing migration" id="drop alert_definition_version table" grafana | logger=migrator t=2024-04-24T09:37:39.478613614Z level=info msg="Migration successfully executed" id="drop alert_definition_version table" duration=2.240639ms grafana | logger=migrator t=2024-04-24T09:37:39.485035051Z level=info msg="Executing migration" id="create alert_instance table" policy-pap | value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer policy-pap | policy-pap | [2024-04-24T09:38:17.898+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 policy-pap | [2024-04-24T09:38:17.899+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 policy-pap | [2024-04-24T09:38:17.899+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1713951497898 policy-pap | [2024-04-24T09:38:17.899+00:00|INFO|KafkaConsumer|main] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Subscribed to topic(s): policy-pdp-pap policy-pap | [2024-04-24T09:38:17.900+00:00|INFO|ServiceManager|main] Policy PAP starting topics policy-pap | [2024-04-24T09:38:17.900+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=policy-pap, consumerInstance=36ee11dc-f044-4dbb-89eb-a07ca9cc5bb6, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-heartbeat,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-heartbeat, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-pap | [2024-04-24T09:38:17.900+00:00|INFO|SingleThreadedBusTopicSource|main] SingleThreadedKafkaTopicSource [getTopicCommInfrastructure()=KAFKA, toString()=SingleThreadedBusTopicSource [consumerGroup=0d71b024-87db-4bab-bad3-947ec269530c, consumerInstance=policy-pap, fetchTimeout=15000, fetchLimit=-1, consumer=KafkaConsumerWrapper [fetchTimeout=15000], alive=true, locked=false, uebThread=Thread[KAFKA-source-policy-pdp-pap,5,main], topicListeners=1, toString()=BusTopicBase [apiKey=null, apiSecret=null, useHttps=false, allowSelfSignedCerts=false, toString()=TopicBase [servers=[kafka:9092], topic=policy-pdp-pap, effectiveTopic=policy-pdp-pap, #recentEvents=0, locked=false, #topicListeners=1]]]]: starting policy-pap | [2024-04-24T09:38:17.900+00:00|INFO|InlineBusTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=71d02006-bede-4d42-8d1e-9e27b286b49a, alive=false, publisher=null]]: starting policy-pap | [2024-04-24T09:38:17.918+00:00|INFO|ProducerConfig|main] ProducerConfig values: policy-pap | acks = -1 policy-pap | auto.include.jmx.reporter = true policy-pap | batch.size = 16384 policy-pap | bootstrap.servers = [kafka:9092] policy-pap | buffer.memory = 33554432 policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = producer-1 policy-pap | compression.type = none policy-pap | connections.max.idle.ms = 540000 policy-pap | delivery.timeout.ms = 120000 policy-pap | enable.idempotence = true policy-pap | interceptor.classes = [] policy-pap | key.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-pap | linger.ms = 0 policy-pap | max.block.ms = 60000 policy-pap | max.in.flight.requests.per.connection = 5 policy-pap | max.request.size = 1048576 policy-pap | metadata.max.age.ms = 300000 policy-pap | metadata.max.idle.ms = 300000 policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partitioner.adaptive.partitioning.enable = true policy-pap | partitioner.availability.timeout.ms = 0 policy-pap | partitioner.class = null policy-pap | partitioner.ignore.keys = false policy-pap | receive.buffer.bytes = 32768 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 policy-pap | request.timeout.ms = 30000 policy-pap | retries = 2147483647 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | transaction.timeout.ms = 60000 policy-pap | transactional.id = null policy-pap | value.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-pap | policy-pap | [2024-04-24T09:38:17.930+00:00|INFO|KafkaProducer|main] [Producer clientId=producer-1] Instantiated an idempotent producer. policy-pap | [2024-04-24T09:38:17.948+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 policy-pap | [2024-04-24T09:38:17.948+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 policy-pap | [2024-04-24T09:38:17.948+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1713951497948 policy-pap | [2024-04-24T09:38:17.949+00:00|INFO|InlineKafkaTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=71d02006-bede-4d42-8d1e-9e27b286b49a, alive=false, publisher=KafkaPublisherWrapper []]]: KAFKA SINK created policy-pap | [2024-04-24T09:38:17.949+00:00|INFO|InlineBusTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=39038e07-e58d-4d06-abfd-896ab52e354e, alive=false, publisher=null]]: starting policy-pap | [2024-04-24T09:38:17.950+00:00|INFO|ProducerConfig|main] ProducerConfig values: policy-pap | acks = -1 policy-pap | auto.include.jmx.reporter = true policy-pap | batch.size = 16384 policy-pap | bootstrap.servers = [kafka:9092] policy-pap | buffer.memory = 33554432 policy-pap | client.dns.lookup = use_all_dns_ips policy-pap | client.id = producer-2 policy-pap | compression.type = none policy-pap | connections.max.idle.ms = 540000 policy-pap | delivery.timeout.ms = 120000 policy-pap | enable.idempotence = true policy-pap | interceptor.classes = [] kafka | [2024-04-24 09:38:18,629] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-25 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,629] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-35 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,629] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-38 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,629] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-13 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-8 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-21 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-4 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-27 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-7 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-9 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-46 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-41 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-33 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,630] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-23 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,631] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-49 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,631] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-47 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,631] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-16 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,631] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-28 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,631] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-31 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,631] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-36 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-42 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-3 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-18 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-37 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition policy-pdp-pap-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-15 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-24 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-17 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-48 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-19 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-11 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-2 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-43 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-6 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,633] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-14 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-20 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-44 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-39 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-12 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-45 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-5 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-26 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-29 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,634] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-34 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,635] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-10 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,635] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-32 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,635] INFO [Controller id=1 epoch=1] Changed partition __consumer_offsets-40 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=1, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=1, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) kafka | [2024-04-24 09:38:18,638] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=13, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-13 (state.change.logger) kafka | [2024-04-24 09:38:18,639] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=46, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-46 (state.change.logger) kafka | [2024-04-24 09:38:18,639] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=9, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-9 (state.change.logger) kafka | [2024-04-24 09:38:18,639] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=42, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-42 (state.change.logger) kafka | [2024-04-24 09:38:18,639] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=21, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-21 (state.change.logger) kafka | [2024-04-24 09:38:18,639] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=17, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-17 (state.change.logger) kafka | [2024-04-24 09:38:18,639] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=30, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-30 (state.change.logger) kafka | [2024-04-24 09:38:18,639] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=26, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-26 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=5, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-5 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=38, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-38 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=1, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-1 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=34, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-34 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=16, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-16 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=45, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-45 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=12, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-12 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=41, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-41 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=24, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-24 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=20, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-20 (state.change.logger) kafka | [2024-04-24 09:38:18,640] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=49, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-49 (state.change.logger) policy-pap | key.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-pap | linger.ms = 0 policy-pap | max.block.ms = 60000 policy-pap | max.in.flight.requests.per.connection = 5 policy-pap | max.request.size = 1048576 policy-pap | metadata.max.age.ms = 300000 policy-pap | metadata.max.idle.ms = 300000 policy-pap | metric.reporters = [] policy-pap | metrics.num.samples = 2 policy-pap | metrics.recording.level = INFO policy-pap | metrics.sample.window.ms = 30000 policy-pap | partitioner.adaptive.partitioning.enable = true policy-pap | partitioner.availability.timeout.ms = 0 policy-pap | partitioner.class = null policy-pap | partitioner.ignore.keys = false policy-pap | receive.buffer.bytes = 32768 policy-pap | reconnect.backoff.max.ms = 1000 policy-pap | reconnect.backoff.ms = 50 grafana | logger=migrator t=2024-04-24T09:37:39.486559332Z level=info msg="Migration successfully executed" id="create alert_instance table" duration=1.523471ms grafana | logger=migrator t=2024-04-24T09:37:39.495515941Z level=info msg="Executing migration" id="add index in alert_instance table on def_org_id, def_uid and current_state columns" grafana | logger=migrator t=2024-04-24T09:37:39.49922145Z level=info msg="Migration successfully executed" id="add index in alert_instance table on def_org_id, def_uid and current_state columns" duration=3.702758ms grafana | logger=migrator t=2024-04-24T09:37:39.503986691Z level=info msg="Executing migration" id="add index in alert_instance table on def_org_id, current_state columns" grafana | logger=migrator t=2024-04-24T09:37:39.505158177Z level=info msg="Migration successfully executed" id="add index in alert_instance table on def_org_id, current_state columns" duration=1.170916ms grafana | logger=migrator t=2024-04-24T09:37:39.509492161Z level=info msg="Executing migration" id="add column current_state_end to alert_instance" grafana | logger=migrator t=2024-04-24T09:37:39.518299184Z level=info msg="Migration successfully executed" id="add column current_state_end to alert_instance" duration=8.807773ms grafana | logger=migrator t=2024-04-24T09:37:39.523795104Z level=info msg="Executing migration" id="remove index def_org_id, def_uid, current_state on alert_instance" grafana | logger=migrator t=2024-04-24T09:37:39.52469202Z level=info msg="Migration successfully executed" id="remove index def_org_id, def_uid, current_state on alert_instance" duration=896.736µs grafana | logger=migrator t=2024-04-24T09:37:39.537785524Z level=info msg="Executing migration" id="remove index def_org_id, current_state on alert_instance" grafana | logger=migrator t=2024-04-24T09:37:39.53917543Z level=info msg="Migration successfully executed" id="remove index def_org_id, current_state on alert_instance" duration=1.396195ms grafana | logger=migrator t=2024-04-24T09:37:39.547606817Z level=info msg="Executing migration" id="rename def_org_id to rule_org_id in alert_instance" grafana | logger=migrator t=2024-04-24T09:37:39.571192112Z level=info msg="Migration successfully executed" id="rename def_org_id to rule_org_id in alert_instance" duration=23.585455ms grafana | logger=migrator t=2024-04-24T09:37:39.578718853Z level=info msg="Executing migration" id="rename def_uid to rule_uid in alert_instance" grafana | logger=migrator t=2024-04-24T09:37:39.60560227Z level=info msg="Migration successfully executed" id="rename def_uid to rule_uid in alert_instance" duration=26.890937ms grafana | logger=migrator t=2024-04-24T09:37:39.61732873Z level=info msg="Executing migration" id="add index rule_org_id, rule_uid, current_state on alert_instance" grafana | logger=migrator t=2024-04-24T09:37:39.619274938Z level=info msg="Migration successfully executed" id="add index rule_org_id, rule_uid, current_state on alert_instance" duration=1.948328ms grafana | logger=migrator t=2024-04-24T09:37:39.623977916Z level=info msg="Executing migration" id="add index rule_org_id, current_state on alert_instance" grafana | logger=migrator t=2024-04-24T09:37:39.625586691Z level=info msg="Migration successfully executed" id="add index rule_org_id, current_state on alert_instance" duration=1.613055ms grafana | logger=migrator t=2024-04-24T09:37:39.635900104Z level=info msg="Executing migration" id="add current_reason column related to current_state" grafana | logger=migrator t=2024-04-24T09:37:39.643636493Z level=info msg="Migration successfully executed" id="add current_reason column related to current_state" duration=7.743049ms grafana | logger=migrator t=2024-04-24T09:37:39.654186246Z level=info msg="Executing migration" id="add result_fingerprint column to alert_instance" grafana | logger=migrator t=2024-04-24T09:37:39.660017359Z level=info msg="Migration successfully executed" id="add result_fingerprint column to alert_instance" duration=5.830163ms grafana | logger=migrator t=2024-04-24T09:37:39.664660346Z level=info msg="Executing migration" id="create alert_rule table" grafana | logger=migrator t=2024-04-24T09:37:39.665763939Z level=info msg="Migration successfully executed" id="create alert_rule table" duration=1.105683ms grafana | logger=migrator t=2024-04-24T09:37:39.671076212Z level=info msg="Executing migration" id="add index in alert_rule on org_id and title columns" grafana | logger=migrator t=2024-04-24T09:37:39.672661486Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id and title columns" duration=1.586664ms grafana | logger=migrator t=2024-04-24T09:37:39.678724559Z level=info msg="Executing migration" id="add index in alert_rule on org_id and uid columns" grafana | logger=migrator t=2024-04-24T09:37:39.680345143Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id and uid columns" duration=1.619584ms grafana | logger=migrator t=2024-04-24T09:37:39.765229844Z level=info msg="Executing migration" id="add index in alert_rule on org_id, namespace_uid, group_uid columns" grafana | logger=migrator t=2024-04-24T09:37:39.767362369Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, namespace_uid, group_uid columns" duration=2.131895ms grafana | logger=migrator t=2024-04-24T09:37:39.773955113Z level=info msg="Executing migration" id="alter alert_rule table data column to mediumtext in mysql" grafana | logger=migrator t=2024-04-24T09:37:39.774163931Z level=info msg="Migration successfully executed" id="alter alert_rule table data column to mediumtext in mysql" duration=222.298µs grafana | logger=migrator t=2024-04-24T09:37:39.781090858Z level=info msg="Executing migration" id="add column for to alert_rule" grafana | logger=migrator t=2024-04-24T09:37:39.787826309Z level=info msg="Migration successfully executed" id="add column for to alert_rule" duration=6.731451ms grafana | logger=migrator t=2024-04-24T09:37:39.792685754Z level=info msg="Executing migration" id="add column annotations to alert_rule" grafana | logger=migrator t=2024-04-24T09:37:39.798628132Z level=info msg="Migration successfully executed" id="add column annotations to alert_rule" duration=5.938997ms grafana | logger=migrator t=2024-04-24T09:37:39.804366111Z level=info msg="Executing migration" id="add column labels to alert_rule" grafana | logger=migrator t=2024-04-24T09:37:39.811001917Z level=info msg="Migration successfully executed" id="add column labels to alert_rule" duration=6.629655ms grafana | logger=migrator t=2024-04-24T09:37:39.820440395Z level=info msg="Executing migration" id="remove unique index from alert_rule on org_id, title columns" grafana | logger=migrator t=2024-04-24T09:37:39.821836651Z level=info msg="Migration successfully executed" id="remove unique index from alert_rule on org_id, title columns" duration=1.400026ms grafana | logger=migrator t=2024-04-24T09:37:39.828119962Z level=info msg="Executing migration" id="add index in alert_rule on org_id, namespase_uid and title columns" grafana | logger=migrator t=2024-04-24T09:37:39.829140404Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, namespase_uid and title columns" duration=1.021662ms grafana | logger=migrator t=2024-04-24T09:37:39.834247477Z level=info msg="Executing migration" id="add dashboard_uid column to alert_rule" grafana | logger=migrator t=2024-04-24T09:37:39.842417785Z level=info msg="Migration successfully executed" id="add dashboard_uid column to alert_rule" duration=8.169148ms grafana | logger=migrator t=2024-04-24T09:37:39.849127864Z level=info msg="Executing migration" id="add panel_id column to alert_rule" grafana | logger=migrator t=2024-04-24T09:37:39.855338423Z level=info msg="Migration successfully executed" id="add panel_id column to alert_rule" duration=6.211409ms kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-0 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=29, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-29 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=25, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-25 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=8, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-8 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=37, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-37 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=4, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-4 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=33, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-33 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=15, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-15 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=48, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-48 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=11, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-11 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=44, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-44 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=23, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-23 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=19, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-19 (state.change.logger) kafka | [2024-04-24 09:38:18,641] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=32, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-32 (state.change.logger) kafka | [2024-04-24 09:38:18,642] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=28, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-28 (state.change.logger) kafka | [2024-04-24 09:38:18,642] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=7, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-7 (state.change.logger) kafka | [2024-04-24 09:38:18,642] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=40, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-40 (state.change.logger) kafka | [2024-04-24 09:38:18,642] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=3, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-3 (state.change.logger) kafka | [2024-04-24 09:38:18,642] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=36, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-36 (state.change.logger) kafka | [2024-04-24 09:38:18,642] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=47, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-47 (state.change.logger) policy-pap | request.timeout.ms = 30000 policy-pap | retries = 2147483647 policy-pap | retry.backoff.ms = 100 policy-pap | sasl.client.callback.handler.class = null policy-pap | sasl.jaas.config = null policy-pap | sasl.kerberos.kinit.cmd = /usr/bin/kinit policy-pap | sasl.kerberos.min.time.before.relogin = 60000 policy-pap | sasl.kerberos.service.name = null policy-pap | sasl.kerberos.ticket.renew.jitter = 0.05 policy-pap | sasl.kerberos.ticket.renew.window.factor = 0.8 policy-pap | sasl.login.callback.handler.class = null policy-pap | sasl.login.class = null policy-pap | sasl.login.connect.timeout.ms = null policy-pap | sasl.login.read.timeout.ms = null policy-pap | sasl.login.refresh.buffer.seconds = 300 policy-pap | sasl.login.refresh.min.period.seconds = 60 policy-pap | sasl.login.refresh.window.factor = 0.8 policy-pap | sasl.login.refresh.window.jitter = 0.05 policy-pap | sasl.login.retry.backoff.max.ms = 10000 policy-pap | sasl.login.retry.backoff.ms = 100 policy-pap | sasl.mechanism = GSSAPI policy-pap | sasl.oauthbearer.clock.skew.seconds = 30 policy-pap | sasl.oauthbearer.expected.audience = null policy-pap | sasl.oauthbearer.expected.issuer = null policy-pap | sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 policy-pap | sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 policy-pap | sasl.oauthbearer.jwks.endpoint.url = null policy-pap | sasl.oauthbearer.scope.claim.name = scope policy-pap | sasl.oauthbearer.sub.claim.name = sub policy-pap | sasl.oauthbearer.token.endpoint.url = null policy-pap | security.protocol = PLAINTEXT policy-pap | security.providers = null policy-pap | send.buffer.bytes = 131072 policy-pap | socket.connection.setup.timeout.max.ms = 30000 policy-pap | socket.connection.setup.timeout.ms = 10000 policy-pap | ssl.cipher.suites = null policy-pap | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] policy-pap | ssl.endpoint.identification.algorithm = https policy-pap | ssl.engine.factory.class = null policy-pap | ssl.key.password = null policy-pap | ssl.keymanager.algorithm = SunX509 policy-pap | ssl.keystore.certificate.chain = null policy-pap | ssl.keystore.key = null policy-pap | ssl.keystore.location = null policy-pap | ssl.keystore.password = null policy-pap | ssl.keystore.type = JKS policy-pap | ssl.protocol = TLSv1.3 policy-pap | ssl.provider = null policy-pap | ssl.secure.random.implementation = null policy-pap | ssl.trustmanager.algorithm = PKIX policy-pap | ssl.truststore.certificates = null policy-pap | ssl.truststore.location = null policy-pap | ssl.truststore.password = null policy-pap | ssl.truststore.type = JKS policy-pap | transaction.timeout.ms = 60000 policy-pap | transactional.id = null policy-pap | value.serializer = class org.apache.kafka.common.serialization.StringSerializer policy-pap | policy-pap | [2024-04-24T09:38:17.951+00:00|INFO|KafkaProducer|main] [Producer clientId=producer-2] Instantiated an idempotent producer. policy-pap | [2024-04-24T09:38:17.953+00:00|INFO|AppInfoParser|main] Kafka version: 3.6.1 policy-pap | [2024-04-24T09:38:17.954+00:00|INFO|AppInfoParser|main] Kafka commitId: 5e3c2b738d253ff5 policy-pap | [2024-04-24T09:38:17.954+00:00|INFO|AppInfoParser|main] Kafka startTimeMs: 1713951497953 policy-pap | [2024-04-24T09:38:17.954+00:00|INFO|InlineKafkaTopicSink|main] InlineKafkaTopicSink [getTopicCommInfrastructure()=KAFKA, toString()=InlineBusTopicSink [partitionId=39038e07-e58d-4d06-abfd-896ab52e354e, alive=false, publisher=KafkaPublisherWrapper []]]: KAFKA SINK created policy-pap | [2024-04-24T09:38:17.955+00:00|INFO|ServiceManager|main] Policy PAP starting PAP Activator policy-pap | [2024-04-24T09:38:17.955+00:00|INFO|ServiceManager|main] Policy PAP starting PDP publisher policy-pap | [2024-04-24T09:38:17.958+00:00|INFO|ServiceManager|main] Policy PAP starting Policy Notification publisher policy-pap | [2024-04-24T09:38:17.959+00:00|INFO|ServiceManager|main] Policy PAP starting PDP update timers policy-pap | [2024-04-24T09:38:17.963+00:00|INFO|ServiceManager|main] Policy PAP starting PDP state-change timers policy-pap | [2024-04-24T09:38:17.963+00:00|INFO|TimerManager|Thread-9] timer manager update started policy-pap | [2024-04-24T09:38:17.967+00:00|INFO|ServiceManager|main] Policy PAP starting PDP modification lock policy-pap | [2024-04-24T09:38:17.967+00:00|INFO|ServiceManager|main] Policy PAP starting PDP modification requests policy-pap | [2024-04-24T09:38:17.968+00:00|INFO|TimerManager|Thread-10] timer manager state-change started policy-pap | [2024-04-24T09:38:17.968+00:00|INFO|ServiceManager|main] Policy PAP starting PDP expiration timer policy-pap | [2024-04-24T09:38:17.969+00:00|INFO|ServiceManager|main] Policy PAP started policy-pap | [2024-04-24T09:38:17.971+00:00|INFO|PolicyPapApplication|main] Started PolicyPapApplication in 10.308 seconds (process running for 10.979) policy-pap | [2024-04-24T09:38:18.411+00:00|INFO|Metadata|kafka-producer-network-thread | producer-2] [Producer clientId=producer-2] Cluster ID: 8oQQmVC5RD2DltlWw-KT5A policy-pap | [2024-04-24T09:38:18.412+00:00|INFO|Metadata|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] Cluster ID: 8oQQmVC5RD2DltlWw-KT5A policy-pap | [2024-04-24T09:38:18.414+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 2 : {policy-pdp-pap=UNKNOWN_TOPIC_OR_PARTITION} policy-pap | [2024-04-24T09:38:18.414+00:00|INFO|Metadata|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Cluster ID: 8oQQmVC5RD2DltlWw-KT5A policy-pap | [2024-04-24T09:38:18.440+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 2 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:18.441+00:00|INFO|Metadata|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Cluster ID: 8oQQmVC5RD2DltlWw-KT5A policy-pap | [2024-04-24T09:38:18.498+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 4 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:18.530+00:00|INFO|TransactionManager|kafka-producer-network-thread | producer-1] [Producer clientId=producer-1] ProducerId set to 0 with epoch 0 policy-pap | [2024-04-24T09:38:18.531+00:00|INFO|TransactionManager|kafka-producer-network-thread | producer-2] [Producer clientId=producer-2] ProducerId set to 1 with epoch 0 policy-pap | [2024-04-24T09:38:18.569+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 4 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:18.613+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 6 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:18.676+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 6 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:18.724+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 8 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} kafka | [2024-04-24 09:38:18,642] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=14, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-14 (state.change.logger) kafka | [2024-04-24 09:38:18,642] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=43, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-43 (state.change.logger) kafka | [2024-04-24 09:38:18,643] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=10, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-10 (state.change.logger) kafka | [2024-04-24 09:38:18,643] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=22, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-22 (state.change.logger) kafka | [2024-04-24 09:38:18,643] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=18, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-18 (state.change.logger) kafka | [2024-04-24 09:38:18,643] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=31, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-31 (state.change.logger) kafka | [2024-04-24 09:38:18,644] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=27, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-27 (state.change.logger) kafka | [2024-04-24 09:38:18,644] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=39, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-39 (state.change.logger) kafka | [2024-04-24 09:38:18,644] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=6, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-6 (state.change.logger) kafka | [2024-04-24 09:38:18,644] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=35, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-35 (state.change.logger) kafka | [2024-04-24 09:38:18,644] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='policy-pdp-pap', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition policy-pdp-pap-0 (state.change.logger) kafka | [2024-04-24 09:38:18,644] TRACE [Controller id=1 epoch=1] Sending become-leader LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=2, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) to broker 1 for partition __consumer_offsets-2 (state.change.logger) kafka | [2024-04-24 09:38:18,645] INFO [Controller id=1 epoch=1] Sending LeaderAndIsr request to broker 1 with 51 become-leader and 0 become-follower partitions (state.change.logger) kafka | [2024-04-24 09:38:18,648] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet(1) for 51 partitions (state.change.logger) kafka | [2024-04-24 09:38:18,656] INFO [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 for 51 partitions (state.change.logger) kafka | [2024-04-24 09:38:18,657] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-32 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='policy-pdp-pap', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=13, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=46, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=9, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=42, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=21, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=17, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=30, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=26, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=5, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=38, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=1, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=34, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=16, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=45, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=12, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,658] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=41, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=24, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=20, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=49, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=29, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=25, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=8, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=37, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=4, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=33, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=15, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=48, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=11, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=44, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=23, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=19, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=32, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=28, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=7, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=40, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=3, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=36, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=47, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=14, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=43, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=10, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=22, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,659] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=18, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,660] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=31, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,660] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=27, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) grafana | logger=migrator t=2024-04-24T09:37:39.860571822Z level=info msg="Executing migration" id="add index in alert_rule on org_id, dashboard_uid and panel_id columns" grafana | logger=migrator t=2024-04-24T09:37:39.861695187Z level=info msg="Migration successfully executed" id="add index in alert_rule on org_id, dashboard_uid and panel_id columns" duration=1.123445ms grafana | logger=migrator t=2024-04-24T09:37:39.867861844Z level=info msg="Executing migration" id="add rule_group_idx column to alert_rule" grafana | logger=migrator t=2024-04-24T09:37:39.873601954Z level=info msg="Migration successfully executed" id="add rule_group_idx column to alert_rule" duration=5.74005ms grafana | logger=migrator t=2024-04-24T09:37:39.879787982Z level=info msg="Executing migration" id="add is_paused column to alert_rule table" grafana | logger=migrator t=2024-04-24T09:37:39.889867775Z level=info msg="Migration successfully executed" id="add is_paused column to alert_rule table" duration=10.077713ms grafana | logger=migrator t=2024-04-24T09:37:39.895063374Z level=info msg="Executing migration" id="fix is_paused column for alert_rule table" grafana | logger=migrator t=2024-04-24T09:37:39.895161557Z level=info msg="Migration successfully executed" id="fix is_paused column for alert_rule table" duration=99.064µs grafana | logger=migrator t=2024-04-24T09:37:39.900162228Z level=info msg="Executing migration" id="create alert_rule_version table" grafana | logger=migrator t=2024-04-24T09:37:39.901522552Z level=info msg="Migration successfully executed" id="create alert_rule_version table" duration=1.359734ms grafana | logger=migrator t=2024-04-24T09:37:39.906928189Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_org_id, rule_uid and version columns" grafana | logger=migrator t=2024-04-24T09:37:39.908046014Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_org_id, rule_uid and version columns" duration=1.117866ms grafana | logger=migrator t=2024-04-24T09:37:39.91370396Z level=info msg="Executing migration" id="add index in alert_rule_version table on rule_org_id, rule_namespace_uid and rule_group columns" grafana | logger=migrator t=2024-04-24T09:37:39.915037713Z level=info msg="Migration successfully executed" id="add index in alert_rule_version table on rule_org_id, rule_namespace_uid and rule_group columns" duration=1.333773ms grafana | logger=migrator t=2024-04-24T09:37:39.919537324Z level=info msg="Executing migration" id="alter alert_rule_version table data column to mediumtext in mysql" grafana | logger=migrator t=2024-04-24T09:37:39.919608517Z level=info msg="Migration successfully executed" id="alter alert_rule_version table data column to mediumtext in mysql" duration=71.743µs grafana | logger=migrator t=2024-04-24T09:37:39.925036534Z level=info msg="Executing migration" id="add column for to alert_rule_version" grafana | logger=migrator t=2024-04-24T09:37:39.93167334Z level=info msg="Migration successfully executed" id="add column for to alert_rule_version" duration=6.639016ms grafana | logger=migrator t=2024-04-24T09:37:39.935748454Z level=info msg="Executing migration" id="add column annotations to alert_rule_version" grafana | logger=migrator t=2024-04-24T09:37:39.941973993Z level=info msg="Migration successfully executed" id="add column annotations to alert_rule_version" duration=6.220899ms grafana | logger=migrator t=2024-04-24T09:37:39.947775495Z level=info msg="Executing migration" id="add column labels to alert_rule_version" grafana | logger=migrator t=2024-04-24T09:37:39.95441164Z level=info msg="Migration successfully executed" id="add column labels to alert_rule_version" duration=6.636105ms grafana | logger=migrator t=2024-04-24T09:37:39.959160781Z level=info msg="Executing migration" id="add rule_group_idx column to alert_rule_version" grafana | logger=migrator t=2024-04-24T09:37:39.965073178Z level=info msg="Migration successfully executed" id="add rule_group_idx column to alert_rule_version" duration=5.911916ms grafana | logger=migrator t=2024-04-24T09:37:39.971272826Z level=info msg="Executing migration" id="add is_paused column to alert_rule_versions table" grafana | logger=migrator t=2024-04-24T09:37:39.977750516Z level=info msg="Migration successfully executed" id="add is_paused column to alert_rule_versions table" duration=6.47686ms grafana | logger=migrator t=2024-04-24T09:37:39.982080199Z level=info msg="Executing migration" id="fix is_paused column for alert_rule_version table" grafana | logger=migrator t=2024-04-24T09:37:39.982151282Z level=info msg="Migration successfully executed" id="fix is_paused column for alert_rule_version table" duration=72.162µs grafana | logger=migrator t=2024-04-24T09:37:39.988523717Z level=info msg="Executing migration" id=create_alert_configuration_table grafana | logger=migrator t=2024-04-24T09:37:39.989391672Z level=info msg="Migration successfully executed" id=create_alert_configuration_table duration=868.185µs grafana | logger=migrator t=2024-04-24T09:37:39.995447044Z level=info msg="Executing migration" id="Add column default in alert_configuration" grafana | logger=migrator t=2024-04-24T09:37:40.00185077Z level=info msg="Migration successfully executed" id="Add column default in alert_configuration" duration=6.403286ms grafana | logger=migrator t=2024-04-24T09:37:40.00929656Z level=info msg="Executing migration" id="alert alert_configuration alertmanager_configuration column from TEXT to MEDIUMTEXT if mysql" grafana | logger=migrator t=2024-04-24T09:37:40.009437445Z level=info msg="Migration successfully executed" id="alert alert_configuration alertmanager_configuration column from TEXT to MEDIUMTEXT if mysql" duration=144.926µs grafana | logger=migrator t=2024-04-24T09:37:40.01568913Z level=info msg="Executing migration" id="add column org_id in alert_configuration" grafana | logger=migrator t=2024-04-24T09:37:40.02287142Z level=info msg="Migration successfully executed" id="add column org_id in alert_configuration" duration=7.18267ms grafana | logger=migrator t=2024-04-24T09:37:40.030422232Z level=info msg="Executing migration" id="add index in alert_configuration table on org_id column" grafana | logger=migrator t=2024-04-24T09:37:40.032042111Z level=info msg="Migration successfully executed" id="add index in alert_configuration table on org_id column" duration=1.620489ms grafana | logger=migrator t=2024-04-24T09:37:40.038417061Z level=info msg="Executing migration" id="add configuration_hash column to alert_configuration" grafana | logger=migrator t=2024-04-24T09:37:40.045622092Z level=info msg="Migration successfully executed" id="add configuration_hash column to alert_configuration" duration=7.204721ms grafana | logger=migrator t=2024-04-24T09:37:40.051230784Z level=info msg="Executing migration" id=create_ngalert_configuration_table grafana | logger=migrator t=2024-04-24T09:37:40.051917489Z level=info msg="Migration successfully executed" id=create_ngalert_configuration_table duration=687.625µs grafana | logger=migrator t=2024-04-24T09:37:40.08850534Z level=info msg="Executing migration" id="add index in ngalert_configuration on org_id column" grafana | logger=migrator t=2024-04-24T09:37:40.090387459Z level=info msg="Migration successfully executed" id="add index in ngalert_configuration on org_id column" duration=1.886019ms grafana | logger=migrator t=2024-04-24T09:37:40.097858119Z level=info msg="Executing migration" id="add column send_alerts_to in ngalert_configuration" grafana | logger=migrator t=2024-04-24T09:37:40.105167852Z level=info msg="Migration successfully executed" id="add column send_alerts_to in ngalert_configuration" duration=7.310393ms grafana | logger=migrator t=2024-04-24T09:37:40.109648204Z level=info msg="Executing migration" id="create provenance_type table" kafka | [2024-04-24 09:38:18,660] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=39, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,660] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=6, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,660] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=35, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,660] TRACE [Broker id=1] Received LeaderAndIsr request LeaderAndIsrPartitionState(topicName='__consumer_offsets', partitionIndex=2, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], partitionEpoch=0, replicas=[1], addingReplicas=[], removingReplicas=[], isNew=true, leaderRecoveryState=0) correlation id 1 from controller 1 epoch 1 (state.change.logger) kafka | [2024-04-24 09:38:18,660] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-5 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,661] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-44 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,661] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-48 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,662] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-46 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,662] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-20 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,662] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition policy-pdp-pap-0 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,663] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-43 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,663] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-24 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,663] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-6 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,663] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-18 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,664] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-21 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,664] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-1 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,664] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-14 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,664] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-34 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,665] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-16 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,665] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-29 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,665] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-11 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,665] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-0 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,666] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-22 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,666] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-47 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,666] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-36 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,666] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-28 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,666] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-42 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,666] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-9 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,667] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-37 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,667] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-13 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,667] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-30 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,667] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-35 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,667] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-39 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,667] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-12 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,667] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-27 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,667] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-45 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,668] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-19 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,668] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-49 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,668] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-40 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,668] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-41 from NewReplica to OnlineReplica (state.change.logger) policy-pap | [2024-04-24T09:38:18.794+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 8 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:18.836+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 10 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:18.901+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 10 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:18.943+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 12 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.010+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 12 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.078+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 14 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.120+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 14 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.184+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 16 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.225+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 16 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.296+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 18 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.332+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 18 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.411+00:00|WARN|NetworkClient|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Error while fetching metadata with correlation id 20 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.444+00:00|WARN|NetworkClient|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Error while fetching metadata with correlation id 20 : {policy-pdp-pap=LEADER_NOT_AVAILABLE} policy-pap | [2024-04-24T09:38:19.527+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Discovered group coordinator kafka:9092 (id: 2147483646 rack: null) policy-pap | [2024-04-24T09:38:19.533+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] (Re-)joining group policy-pap | [2024-04-24T09:38:19.549+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Discovered group coordinator kafka:9092 (id: 2147483646 rack: null) policy-pap | [2024-04-24T09:38:19.551+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] (Re-)joining group policy-pap | [2024-04-24T09:38:19.561+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Request joining group due to: need to re-join with the given member-id: consumer-0d71b024-87db-4bab-bad3-947ec269530c-3-d1876d69-1880-43fc-9d10-af88199c63eb policy-pap | [2024-04-24T09:38:19.561+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Request joining group due to: need to re-join with the given member-id: consumer-policy-pap-4-5818d6ba-9a29-412c-a74d-37f074bc3da4 policy-pap | [2024-04-24T09:38:19.561+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) policy-pap | [2024-04-24T09:38:19.561+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] (Re-)joining group policy-pap | [2024-04-24T09:38:19.561+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Request joining group due to: rebalance failed due to 'The group member needs to have a valid member id before actually entering a consumer group.' (MemberIdRequiredException) policy-pap | [2024-04-24T09:38:19.562+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] (Re-)joining group policy-pap | [2024-04-24T09:38:22.588+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Successfully joined group with generation Generation{generationId=1, memberId='consumer-policy-pap-4-5818d6ba-9a29-412c-a74d-37f074bc3da4', protocol='range'} policy-pap | [2024-04-24T09:38:22.591+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Successfully joined group with generation Generation{generationId=1, memberId='consumer-0d71b024-87db-4bab-bad3-947ec269530c-3-d1876d69-1880-43fc-9d10-af88199c63eb', protocol='range'} policy-pap | [2024-04-24T09:38:22.599+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Finished assignment for group at generation 1: {consumer-policy-pap-4-5818d6ba-9a29-412c-a74d-37f074bc3da4=Assignment(partitions=[policy-pdp-pap-0])} policy-pap | [2024-04-24T09:38:22.599+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Finished assignment for group at generation 1: {consumer-0d71b024-87db-4bab-bad3-947ec269530c-3-d1876d69-1880-43fc-9d10-af88199c63eb=Assignment(partitions=[policy-pdp-pap-0])} policy-pap | [2024-04-24T09:38:22.633+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Successfully synced group in generation Generation{generationId=1, memberId='consumer-0d71b024-87db-4bab-bad3-947ec269530c-3-d1876d69-1880-43fc-9d10-af88199c63eb', protocol='range'} policy-pap | [2024-04-24T09:38:22.633+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Successfully synced group in generation Generation{generationId=1, memberId='consumer-policy-pap-4-5818d6ba-9a29-412c-a74d-37f074bc3da4', protocol='range'} policy-pap | [2024-04-24T09:38:22.634+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Notifying assignor about the new Assignment(partitions=[policy-pdp-pap-0]) policy-pap | [2024-04-24T09:38:22.634+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Notifying assignor about the new Assignment(partitions=[policy-pdp-pap-0]) policy-pap | [2024-04-24T09:38:22.638+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Adding newly assigned partitions: policy-pdp-pap-0 policy-pap | [2024-04-24T09:38:22.638+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Adding newly assigned partitions: policy-pdp-pap-0 policy-pap | [2024-04-24T09:38:22.660+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Found no committed offset for partition policy-pdp-pap-0 policy-pap | [2024-04-24T09:38:22.660+00:00|INFO|ConsumerCoordinator|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Found no committed offset for partition policy-pdp-pap-0 policy-pap | [2024-04-24T09:38:22.684+00:00|INFO|SubscriptionState|KAFKA-source-policy-heartbeat] [Consumer clientId=consumer-policy-pap-4, groupId=policy-pap] Resetting offset for partition policy-pdp-pap-0 to position FetchPosition{offset=1, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:9092 (id: 1 rack: null)], epoch=0}}. policy-pap | [2024-04-24T09:38:22.684+00:00|INFO|SubscriptionState|KAFKA-source-policy-pdp-pap] [Consumer clientId=consumer-0d71b024-87db-4bab-bad3-947ec269530c-3, groupId=0d71b024-87db-4bab-bad3-947ec269530c] Resetting offset for partition policy-pdp-pap-0 to position FetchPosition{offset=1, offsetEpoch=Optional.empty, currentLeader=LeaderAndEpoch{leader=Optional[kafka:9092 (id: 1 rack: null)], epoch=0}}. policy-pap | [2024-04-24T09:38:26.250+00:00|INFO|[/policy/pap/v1]|http-nio-6969-exec-3] Initializing Spring DispatcherServlet 'dispatcherServlet' policy-pap | [2024-04-24T09:38:26.250+00:00|INFO|DispatcherServlet|http-nio-6969-exec-3] Initializing Servlet 'dispatcherServlet' policy-pap | [2024-04-24T09:38:26.251+00:00|INFO|DispatcherServlet|http-nio-6969-exec-3] Completed initialization in 1 ms policy-pap | [2024-04-24T09:38:39.931+00:00|INFO|OrderedServiceImpl|KAFKA-source-policy-heartbeat] ***** OrderedServiceImpl implementers: policy-pap | [] policy-pap | [2024-04-24T09:38:39.932+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"3e991c30-f7a7-4f78-960f-7e0ae2fecbef","timestampMs":1713951519878,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup"} kafka | [2024-04-24 09:38:18,668] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-38 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,668] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-8 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,668] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-7 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,669] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-33 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,669] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-25 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,669] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-31 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,669] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-23 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,669] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-10 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,669] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-2 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,669] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-17 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,670] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-4 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,670] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-15 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,670] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-26 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,670] TRACE [Controller id=1 epoch=1] Changed state of replica 1 for partition __consumer_offsets-3 from NewReplica to OnlineReplica (state.change.logger) kafka | [2024-04-24 09:38:18,670] INFO [Controller id=1 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) kafka | [2024-04-24 09:38:18,695] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-3 (state.change.logger) kafka | [2024-04-24 09:38:18,695] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-18 (state.change.logger) kafka | [2024-04-24 09:38:18,695] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-41 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-10 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-33 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-48 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-19 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-34 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-4 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-11 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-26 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-49 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-39 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-9 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-24 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-31 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-46 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-1 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-16 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-2 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-25 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-40 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-47 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-17 (state.change.logger) grafana | logger=migrator t=2024-04-24T09:37:40.110464154Z level=info msg="Migration successfully executed" id="create provenance_type table" duration=815.92µs policy-pap | [2024-04-24T09:38:39.932+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"3e991c30-f7a7-4f78-960f-7e0ae2fecbef","timestampMs":1713951519878,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup"} policy-pap | [2024-04-24T09:38:39.941+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-pdp-pap] no listeners for autonomous message of type PdpStatus policy-pap | [2024-04-24T09:38:40.006+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate starting policy-pap | [2024-04-24T09:38:40.006+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate starting listener policy-pap | [2024-04-24T09:38:40.007+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate starting timer policy-pap | [2024-04-24T09:38:40.007+00:00|INFO|TimerManager|KAFKA-source-policy-heartbeat] update timer registered Timer [name=594a6c37-7874-45c6-b60d-69cb6ee9534e, expireMs=1713951550007] policy-pap | [2024-04-24T09:38:40.009+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate starting enqueue policy-pap | [2024-04-24T09:38:40.009+00:00|INFO|ServiceManager|KAFKA-source-policy-heartbeat] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate started policy-pap | [2024-04-24T09:38:40.010+00:00|INFO|TimerManager|Thread-9] update timer waiting 29998ms Timer [name=594a6c37-7874-45c6-b60d-69cb6ee9534e, expireMs=1713951550007] policy-pap | [2024-04-24T09:38:40.011+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"messageName":"PDP_UPDATE","requestId":"594a6c37-7874-45c6-b60d-69cb6ee9534e","timestampMs":1713951519988,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.052+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"messageName":"PDP_UPDATE","requestId":"594a6c37-7874-45c6-b60d-69cb6ee9534e","timestampMs":1713951519988,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.052+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"messageName":"PDP_UPDATE","requestId":"594a6c37-7874-45c6-b60d-69cb6ee9534e","timestampMs":1713951519988,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.052+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_UPDATE policy-pap | [2024-04-24T09:38:40.054+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_UPDATE policy-pap | [2024-04-24T09:38:40.075+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"84be2407-9b5d-4edc-b9b9-3bb4f9743c8b","timestampMs":1713951520061,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup"} policy-pap | [2024-04-24T09:38:40.075+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-pdp-pap] no listeners for autonomous message of type PdpStatus policy-pap | [2024-04-24T09:38:40.075+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"594a6c37-7874-45c6-b60d-69cb6ee9534e","responseStatus":"SUCCESS","responseMessage":"Pdp update successful."},"messageName":"PDP_STATUS","requestId":"77ab3928-ad7a-417a-b373-05e2442a45b1","timestampMs":1713951520062,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.076+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-32 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-37 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-7 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-22 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-29 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-44 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-14 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-23 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-38 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-8 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition policy-pdp-pap-0 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-45 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-15 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-30 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-0 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-35 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-5 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-20 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-27 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-42 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-12 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-21 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-36 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-6 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-43 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-13 (state.change.logger) kafka | [2024-04-24 09:38:18,696] TRACE [Broker id=1] Handling LeaderAndIsr request correlationId 1 from controller 1 epoch 1 starting the become-leader transition for partition __consumer_offsets-28 (state.change.logger) kafka | [2024-04-24 09:38:18,697] INFO [ReplicaFetcherManager on broker 1] Removed fetcher for partitions HashSet(__consumer_offsets-22, __consumer_offsets-30, __consumer_offsets-25, __consumer_offsets-35, __consumer_offsets-38, __consumer_offsets-13, __consumer_offsets-8, __consumer_offsets-21, __consumer_offsets-4, __consumer_offsets-27, __consumer_offsets-7, __consumer_offsets-9, __consumer_offsets-46, __consumer_offsets-41, __consumer_offsets-33, __consumer_offsets-23, __consumer_offsets-49, __consumer_offsets-47, __consumer_offsets-16, __consumer_offsets-28, __consumer_offsets-31, __consumer_offsets-36, __consumer_offsets-42, __consumer_offsets-3, __consumer_offsets-18, __consumer_offsets-37, policy-pdp-pap-0, __consumer_offsets-15, __consumer_offsets-24, __consumer_offsets-17, __consumer_offsets-48, __consumer_offsets-19, __consumer_offsets-11, __consumer_offsets-2, __consumer_offsets-43, __consumer_offsets-6, __consumer_offsets-14, __consumer_offsets-20, __consumer_offsets-0, __consumer_offsets-44, __consumer_offsets-39, __consumer_offsets-12, __consumer_offsets-45, __consumer_offsets-1, __consumer_offsets-5, __consumer_offsets-26, __consumer_offsets-29, __consumer_offsets-34, __consumer_offsets-10, __consumer_offsets-32, __consumer_offsets-40) (kafka.server.ReplicaFetcherManager) kafka | [2024-04-24 09:38:18,698] INFO [Broker id=1] Stopped fetchers as part of LeaderAndIsr request correlationId 1 from controller 1 epoch 1 as part of the become-leader transition for 51 partitions (state.change.logger) kafka | [2024-04-24 09:38:18,750] INFO [LogLoader partition=__consumer_offsets-3, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,767] INFO Created log for partition __consumer_offsets-3 in /var/lib/kafka/data/__consumer_offsets-3 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,771] INFO [Partition __consumer_offsets-3 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-3 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,775] INFO [Partition __consumer_offsets-3 broker=1] Log loaded for partition __consumer_offsets-3 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,777] INFO [Broker id=1] Leader __consumer_offsets-3 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,803] INFO [LogLoader partition=__consumer_offsets-18, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,805] INFO Created log for partition __consumer_offsets-18 in /var/lib/kafka/data/__consumer_offsets-18 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,805] INFO [Partition __consumer_offsets-18 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-18 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,805] INFO [Partition __consumer_offsets-18 broker=1] Log loaded for partition __consumer_offsets-18 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,805] INFO [Broker id=1] Leader __consumer_offsets-18 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,816] INFO [LogLoader partition=__consumer_offsets-41, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,818] INFO Created log for partition __consumer_offsets-41 in /var/lib/kafka/data/__consumer_offsets-41 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,818] INFO [Partition __consumer_offsets-41 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-41 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,818] INFO [Partition __consumer_offsets-41 broker=1] Log loaded for partition __consumer_offsets-41 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,818] INFO [Broker id=1] Leader __consumer_offsets-41 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,827] INFO [LogLoader partition=__consumer_offsets-10, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,827] INFO Created log for partition __consumer_offsets-10 in /var/lib/kafka/data/__consumer_offsets-10 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,828] INFO [Partition __consumer_offsets-10 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-10 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,828] INFO [Partition __consumer_offsets-10 broker=1] Log loaded for partition __consumer_offsets-10 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,828] INFO [Broker id=1] Leader __consumer_offsets-10 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,839] INFO [LogLoader partition=__consumer_offsets-33, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp Heartbeat","messageName":"PDP_STATUS","requestId":"84be2407-9b5d-4edc-b9b9-3bb4f9743c8b","timestampMs":1713951520061,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup"} policy-pap | [2024-04-24T09:38:40.076+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopping policy-pap | [2024-04-24T09:38:40.076+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopping enqueue policy-pap | [2024-04-24T09:38:40.076+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopping timer policy-pap | [2024-04-24T09:38:40.076+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer cancelled Timer [name=594a6c37-7874-45c6-b60d-69cb6ee9534e, expireMs=1713951550007] policy-pap | [2024-04-24T09:38:40.077+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopping listener policy-pap | [2024-04-24T09:38:40.077+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopped policy-pap | [2024-04-24T09:38:40.083+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate successful policy-pap | [2024-04-24T09:38:40.083+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 start publishing next request policy-pap | [2024-04-24T09:38:40.083+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange starting policy-pap | [2024-04-24T09:38:40.083+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange starting listener policy-pap | [2024-04-24T09:38:40.083+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange starting timer policy-pap | [2024-04-24T09:38:40.083+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] state-change timer registered Timer [name=0e043121-ee0e-429d-bc7d-b3a900f421a3, expireMs=1713951550083] policy-pap | [2024-04-24T09:38:40.083+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange starting enqueue policy-pap | [2024-04-24T09:38:40.083+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange started policy-pap | [2024-04-24T09:38:40.083+00:00|INFO|TimerManager|Thread-10] state-change timer waiting 30000ms Timer [name=0e043121-ee0e-429d-bc7d-b3a900f421a3, expireMs=1713951550083] policy-pap | [2024-04-24T09:38:40.084+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"0e043121-ee0e-429d-bc7d-b3a900f421a3","timestampMs":1713951519989,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.120+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"0e043121-ee0e-429d-bc7d-b3a900f421a3","timestampMs":1713951519989,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.121+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_STATE_CHANGE policy-pap | [2024-04-24T09:38:40.124+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpStateChange","policies":[],"response":{"responseTo":"0e043121-ee0e-429d-bc7d-b3a900f421a3","responseStatus":"SUCCESS","responseMessage":"State changed to active. No policies found."},"messageName":"PDP_STATUS","requestId":"b4698cf3-0a26-43ef-845c-f0142a379e14","timestampMs":1713951520098,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.136+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange stopping policy-pap | [2024-04-24T09:38:40.136+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange stopping enqueue policy-pap | [2024-04-24T09:38:40.136+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange stopping timer policy-pap | [2024-04-24T09:38:40.136+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"apex","state":"PASSIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"594a6c37-7874-45c6-b60d-69cb6ee9534e","responseStatus":"SUCCESS","responseMessage":"Pdp update successful."},"messageName":"PDP_STATUS","requestId":"77ab3928-ad7a-417a-b373-05e2442a45b1","timestampMs":1713951520062,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.136+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] state-change timer cancelled Timer [name=0e043121-ee0e-429d-bc7d-b3a900f421a3, expireMs=1713951550083] policy-pap | [2024-04-24T09:38:40.136+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange stopping listener policy-pap | [2024-04-24T09:38:40.136+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange stopped policy-pap | [2024-04-24T09:38:40.137+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpStateChange successful policy-pap | [2024-04-24T09:38:40.137+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 start publishing next request policy-pap | [2024-04-24T09:38:40.137+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate starting policy-pap | [2024-04-24T09:38:40.137+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate starting listener policy-pap | [2024-04-24T09:38:40.137+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate starting timer policy-pap | [2024-04-24T09:38:40.137+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer registered Timer [name=0a475f2e-48b4-4ba1-aa36-217d754d9ba8, expireMs=1713951550137] policy-pap | [2024-04-24T09:38:40.137+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate starting enqueue policy-pap | [2024-04-24T09:38:40.137+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate started policy-pap | [2024-04-24T09:38:40.137+00:00|INFO|network|Thread-7] [OUT|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"0a475f2e-48b4-4ba1-aa36-217d754d9ba8","timestampMs":1713951520109,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.138+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id 594a6c37-7874-45c6-b60d-69cb6ee9534e policy-pap | [2024-04-24T09:38:40.142+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","state":"ACTIVE","messageName":"PDP_STATE_CHANGE","requestId":"0e043121-ee0e-429d-bc7d-b3a900f421a3","timestampMs":1713951519989,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} kafka | [2024-04-24 09:38:18,843] INFO Created log for partition __consumer_offsets-33 in /var/lib/kafka/data/__consumer_offsets-33 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,843] INFO [Partition __consumer_offsets-33 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-33 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,843] INFO [Partition __consumer_offsets-33 broker=1] Log loaded for partition __consumer_offsets-33 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,844] INFO [Broker id=1] Leader __consumer_offsets-33 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,850] INFO [LogLoader partition=__consumer_offsets-48, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,851] INFO Created log for partition __consumer_offsets-48 in /var/lib/kafka/data/__consumer_offsets-48 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,851] INFO [Partition __consumer_offsets-48 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-48 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,851] INFO [Partition __consumer_offsets-48 broker=1] Log loaded for partition __consumer_offsets-48 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,851] INFO [Broker id=1] Leader __consumer_offsets-48 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,859] INFO [LogLoader partition=__consumer_offsets-19, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,859] INFO Created log for partition __consumer_offsets-19 in /var/lib/kafka/data/__consumer_offsets-19 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,859] INFO [Partition __consumer_offsets-19 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-19 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,860] INFO [Partition __consumer_offsets-19 broker=1] Log loaded for partition __consumer_offsets-19 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,861] INFO [Broker id=1] Leader __consumer_offsets-19 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,875] INFO [LogLoader partition=__consumer_offsets-34, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,876] INFO Created log for partition __consumer_offsets-34 in /var/lib/kafka/data/__consumer_offsets-34 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,876] INFO [Partition __consumer_offsets-34 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-34 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,877] INFO [Partition __consumer_offsets-34 broker=1] Log loaded for partition __consumer_offsets-34 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,877] INFO [Broker id=1] Leader __consumer_offsets-34 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,892] INFO [LogLoader partition=__consumer_offsets-4, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,894] INFO Created log for partition __consumer_offsets-4 in /var/lib/kafka/data/__consumer_offsets-4 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,895] INFO [Partition __consumer_offsets-4 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-4 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,895] INFO [Partition __consumer_offsets-4 broker=1] Log loaded for partition __consumer_offsets-4 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,896] INFO [Broker id=1] Leader __consumer_offsets-4 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,906] INFO [LogLoader partition=__consumer_offsets-11, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,907] INFO Created log for partition __consumer_offsets-11 in /var/lib/kafka/data/__consumer_offsets-11 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,907] INFO [Partition __consumer_offsets-11 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-11 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,907] INFO [Partition __consumer_offsets-11 broker=1] Log loaded for partition __consumer_offsets-11 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,908] INFO [Broker id=1] Leader __consumer_offsets-11 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,915] INFO [LogLoader partition=__consumer_offsets-26, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,916] INFO Created log for partition __consumer_offsets-26 in /var/lib/kafka/data/__consumer_offsets-26 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,916] INFO [Partition __consumer_offsets-26 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-26 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,916] INFO [Partition __consumer_offsets-26 broker=1] Log loaded for partition __consumer_offsets-26 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,916] INFO [Broker id=1] Leader __consumer_offsets-26 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,928] INFO [LogLoader partition=__consumer_offsets-49, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,929] INFO Created log for partition __consumer_offsets-49 in /var/lib/kafka/data/__consumer_offsets-49 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,929] INFO [Partition __consumer_offsets-49 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-49 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,929] INFO [Partition __consumer_offsets-49 broker=1] Log loaded for partition __consumer_offsets-49 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,930] INFO [Broker id=1] Leader __consumer_offsets-49 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,937] INFO [LogLoader partition=__consumer_offsets-39, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,938] INFO Created log for partition __consumer_offsets-39 in /var/lib/kafka/data/__consumer_offsets-39 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,938] INFO [Partition __consumer_offsets-39 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-39 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,938] INFO [Partition __consumer_offsets-39 broker=1] Log loaded for partition __consumer_offsets-39 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,939] INFO [Broker id=1] Leader __consumer_offsets-39 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,947] INFO [LogLoader partition=__consumer_offsets-9, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,948] INFO Created log for partition __consumer_offsets-9 in /var/lib/kafka/data/__consumer_offsets-9 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,948] INFO [Partition __consumer_offsets-9 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-9 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,948] INFO [Partition __consumer_offsets-9 broker=1] Log loaded for partition __consumer_offsets-9 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,949] INFO [Broker id=1] Leader __consumer_offsets-9 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,955] INFO [LogLoader partition=__consumer_offsets-24, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) policy-pap | [2024-04-24T09:38:40.142+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_STATE_CHANGE policy-pap | [2024-04-24T09:38:40.146+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpStateChange","policies":[],"response":{"responseTo":"0e043121-ee0e-429d-bc7d-b3a900f421a3","responseStatus":"SUCCESS","responseMessage":"State changed to active. No policies found."},"messageName":"PDP_STATUS","requestId":"b4698cf3-0a26-43ef-845c-f0142a379e14","timestampMs":1713951520098,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.147+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"0a475f2e-48b4-4ba1-aa36-217d754d9ba8","timestampMs":1713951520109,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.147+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-pdp-pap] discarding event of type PDP_UPDATE policy-pap | [2024-04-24T09:38:40.147+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id 0e043121-ee0e-429d-bc7d-b3a900f421a3 policy-pap | [2024-04-24T09:38:40.153+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"source":"pap-eaf62c87-1c53-4123-88dd-9318e81917a4","pdpHeartbeatIntervalMs":120000,"policiesToBeDeployed":[],"policiesToBeUndeployed":[],"messageName":"PDP_UPDATE","requestId":"0a475f2e-48b4-4ba1-aa36-217d754d9ba8","timestampMs":1713951520109,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.153+00:00|INFO|MessageTypeDispatcher|KAFKA-source-policy-heartbeat] discarding event of type PDP_UPDATE policy-pap | [2024-04-24T09:38:40.157+00:00|INFO|network|KAFKA-source-policy-heartbeat] [IN|KAFKA|policy-heartbeat] policy-pap | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"0a475f2e-48b4-4ba1-aa36-217d754d9ba8","responseStatus":"SUCCESS","responseMessage":"Pdp already updated"},"messageName":"PDP_STATUS","requestId":"a21ae420-dcab-420b-8924-c4ec1683dfda","timestampMs":1713951520149,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.158+00:00|INFO|RequestIdDispatcher|KAFKA-source-policy-heartbeat] no listener for request id 0a475f2e-48b4-4ba1-aa36-217d754d9ba8 policy-pap | [2024-04-24T09:38:40.159+00:00|INFO|network|KAFKA-source-policy-pdp-pap] [IN|KAFKA|policy-pdp-pap] policy-pap | {"pdpType":"apex","state":"ACTIVE","healthy":"HEALTHY","description":"Pdp status response message for PdpUpdate","policies":[],"response":{"responseTo":"0a475f2e-48b4-4ba1-aa36-217d754d9ba8","responseStatus":"SUCCESS","responseMessage":"Pdp already updated"},"messageName":"PDP_STATUS","requestId":"a21ae420-dcab-420b-8924-c4ec1683dfda","timestampMs":1713951520149,"name":"apex-bca25a69-b004-4f69-848a-7f2f4ff692c1","pdpGroup":"defaultGroup","pdpSubgroup":"apex"} policy-pap | [2024-04-24T09:38:40.160+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopping policy-pap | [2024-04-24T09:38:40.160+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopping enqueue policy-pap | [2024-04-24T09:38:40.160+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopping timer policy-pap | [2024-04-24T09:38:40.160+00:00|INFO|TimerManager|KAFKA-source-policy-pdp-pap] update timer cancelled Timer [name=0a475f2e-48b4-4ba1-aa36-217d754d9ba8, expireMs=1713951550137] policy-pap | [2024-04-24T09:38:40.160+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopping listener policy-pap | [2024-04-24T09:38:40.160+00:00|INFO|ServiceManager|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate stopped policy-pap | [2024-04-24T09:38:40.164+00:00|INFO|RequestImpl|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 PdpUpdate successful policy-pap | [2024-04-24T09:38:40.164+00:00|INFO|PdpRequests|KAFKA-source-policy-pdp-pap] apex-bca25a69-b004-4f69-848a-7f2f4ff692c1 has no more requests policy-pap | [2024-04-24T09:38:46.708+00:00|WARN|NonInjectionManager|pool-2-thread-1] Falling back to injection-less client. policy-pap | [2024-04-24T09:38:46.758+00:00|INFO|GsonMessageBodyHandler|pool-2-thread-1] Using GSON for REST calls policy-pap | [2024-04-24T09:38:46.766+00:00|INFO|GsonMessageBodyHandler|pool-2-thread-1] Using GSON for REST calls policy-pap | [2024-04-24T09:38:46.771+00:00|INFO|GsonMessageBodyHandler|pool-2-thread-1] Using GSON for REST calls policy-pap | [2024-04-24T09:38:47.176+00:00|INFO|SessionData|http-nio-6969-exec-6] unknown group testGroup policy-pap | [2024-04-24T09:38:47.698+00:00|INFO|SessionData|http-nio-6969-exec-6] create cached group testGroup policy-pap | [2024-04-24T09:38:47.699+00:00|INFO|SessionData|http-nio-6969-exec-6] creating DB group testGroup policy-pap | [2024-04-24T09:38:48.241+00:00|INFO|SessionData|http-nio-6969-exec-1] cache group testGroup policy-pap | [2024-04-24T09:38:48.428+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-1] Registering a deploy for policy onap.restart.tca 1.0.0 policy-pap | [2024-04-24T09:38:48.522+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-1] Registering a deploy for policy operational.apex.decisionMaker 1.0.0 policy-pap | [2024-04-24T09:38:48.522+00:00|INFO|SessionData|http-nio-6969-exec-1] update cached group testGroup policy-pap | [2024-04-24T09:38:48.523+00:00|INFO|SessionData|http-nio-6969-exec-1] updating DB group testGroup policy-pap | [2024-04-24T09:38:48.573+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-1] sending audit records to database: [PolicyAudit(auditId=null, pdpGroup=testGroup, pdpType=pdpTypeA, policy=onap.restart.tca 1.0.0, action=DEPLOYMENT, timestamp=2024-04-24T09:38:48Z, user=policyadmin), PolicyAudit(auditId=null, pdpGroup=testGroup, pdpType=pdpTypeC, policy=operational.apex.decisionMaker 1.0.0, action=DEPLOYMENT, timestamp=2024-04-24T09:38:48Z, user=policyadmin)] policy-pap | [2024-04-24T09:38:49.273+00:00|INFO|SessionData|http-nio-6969-exec-5] cache group testGroup policy-pap | [2024-04-24T09:38:49.275+00:00|INFO|PdpGroupDeleteProvider|http-nio-6969-exec-5] remove policy onap.restart.tca 1.0.0 from subgroup testGroup pdpTypeA count=0 policy-pap | [2024-04-24T09:38:49.276+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-5] Registering an undeploy for policy onap.restart.tca 1.0.0 policy-pap | [2024-04-24T09:38:49.276+00:00|INFO|SessionData|http-nio-6969-exec-5] update cached group testGroup policy-pap | [2024-04-24T09:38:49.276+00:00|INFO|SessionData|http-nio-6969-exec-5] updating DB group testGroup policy-pap | [2024-04-24T09:38:49.285+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-5] sending audit records to database: [PolicyAudit(auditId=null, pdpGroup=testGroup, pdpType=pdpTypeA, policy=onap.restart.tca 1.0.0, action=UNDEPLOYMENT, timestamp=2024-04-24T09:38:49Z, user=policyadmin)] policy-pap | [2024-04-24T09:38:49.615+00:00|INFO|SessionData|http-nio-6969-exec-7] cache group defaultGroup policy-pap | [2024-04-24T09:38:49.615+00:00|INFO|SessionData|http-nio-6969-exec-7] cache group testGroup policy-pap | [2024-04-24T09:38:49.615+00:00|INFO|PdpGroupDeleteProvider|http-nio-6969-exec-7] remove policy operational.apex.decisionMaker 1.0.0 from subgroup testGroup pdpTypeC count=0 policy-pap | [2024-04-24T09:38:49.616+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-7] Registering an undeploy for policy operational.apex.decisionMaker 1.0.0 policy-pap | [2024-04-24T09:38:49.616+00:00|INFO|SessionData|http-nio-6969-exec-7] update cached group testGroup policy-pap | [2024-04-24T09:38:49.616+00:00|INFO|SessionData|http-nio-6969-exec-7] updating DB group testGroup policy-pap | [2024-04-24T09:38:49.624+00:00|INFO|PolicyAuditManager|http-nio-6969-exec-7] sending audit records to database: [PolicyAudit(auditId=null, pdpGroup=testGroup, pdpType=pdpTypeC, policy=operational.apex.decisionMaker 1.0.0, action=UNDEPLOYMENT, timestamp=2024-04-24T09:38:49Z, user=policyadmin)] policy-pap | [2024-04-24T09:39:10.008+00:00|INFO|TimerManager|Thread-9] update timer discarded (expired) Timer [name=594a6c37-7874-45c6-b60d-69cb6ee9534e, expireMs=1713951550007] policy-pap | [2024-04-24T09:39:10.084+00:00|INFO|TimerManager|Thread-10] state-change timer discarded (expired) Timer [name=0e043121-ee0e-429d-bc7d-b3a900f421a3, expireMs=1713951550083] policy-pap | [2024-04-24T09:39:10.189+00:00|INFO|SessionData|http-nio-6969-exec-1] cache group testGroup policy-pap | [2024-04-24T09:39:10.191+00:00|INFO|SessionData|http-nio-6969-exec-1] deleting DB group testGroup kafka | [2024-04-24 09:38:18,956] INFO Created log for partition __consumer_offsets-24 in /var/lib/kafka/data/__consumer_offsets-24 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,956] INFO [Partition __consumer_offsets-24 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-24 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,956] INFO [Partition __consumer_offsets-24 broker=1] Log loaded for partition __consumer_offsets-24 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,956] INFO [Broker id=1] Leader __consumer_offsets-24 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,963] INFO [LogLoader partition=__consumer_offsets-31, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,963] INFO Created log for partition __consumer_offsets-31 in /var/lib/kafka/data/__consumer_offsets-31 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,963] INFO [Partition __consumer_offsets-31 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-31 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,964] INFO [Partition __consumer_offsets-31 broker=1] Log loaded for partition __consumer_offsets-31 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,964] INFO [Broker id=1] Leader __consumer_offsets-31 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,971] INFO [LogLoader partition=__consumer_offsets-46, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,972] INFO Created log for partition __consumer_offsets-46 in /var/lib/kafka/data/__consumer_offsets-46 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,972] INFO [Partition __consumer_offsets-46 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-46 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,972] INFO [Partition __consumer_offsets-46 broker=1] Log loaded for partition __consumer_offsets-46 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,972] INFO [Broker id=1] Leader __consumer_offsets-46 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,981] INFO [LogLoader partition=__consumer_offsets-1, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,982] INFO Created log for partition __consumer_offsets-1 in /var/lib/kafka/data/__consumer_offsets-1 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,982] INFO [Partition __consumer_offsets-1 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-1 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,983] INFO [Partition __consumer_offsets-1 broker=1] Log loaded for partition __consumer_offsets-1 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,983] INFO [Broker id=1] Leader __consumer_offsets-1 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,990] INFO [LogLoader partition=__consumer_offsets-16, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,991] INFO Created log for partition __consumer_offsets-16 in /var/lib/kafka/data/__consumer_offsets-16 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,991] INFO [Partition __consumer_offsets-16 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-16 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,991] INFO [Partition __consumer_offsets-16 broker=1] Log loaded for partition __consumer_offsets-16 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,991] INFO [Broker id=1] Leader __consumer_offsets-16 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:18,998] INFO [LogLoader partition=__consumer_offsets-2, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:18,999] INFO Created log for partition __consumer_offsets-2 in /var/lib/kafka/data/__consumer_offsets-2 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:18,999] INFO [Partition __consumer_offsets-2 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-2 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:18,999] INFO [Partition __consumer_offsets-2 broker=1] Log loaded for partition __consumer_offsets-2 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,000] INFO [Broker id=1] Leader __consumer_offsets-2 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,008] INFO [LogLoader partition=__consumer_offsets-25, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,008] INFO Created log for partition __consumer_offsets-25 in /var/lib/kafka/data/__consumer_offsets-25 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,008] INFO [Partition __consumer_offsets-25 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-25 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,009] INFO [Partition __consumer_offsets-25 broker=1] Log loaded for partition __consumer_offsets-25 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,009] INFO [Broker id=1] Leader __consumer_offsets-25 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,018] INFO [LogLoader partition=__consumer_offsets-40, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) grafana | logger=migrator t=2024-04-24T09:37:40.116938377Z level=info msg="Executing migration" id="add index to uniquify (record_key, record_type, org_id) columns" grafana | logger=migrator t=2024-04-24T09:37:40.118016947Z level=info msg="Migration successfully executed" id="add index to uniquify (record_key, record_type, org_id) columns" duration=1.07841ms grafana | logger=migrator t=2024-04-24T09:37:40.124385707Z level=info msg="Executing migration" id="create alert_image table" grafana | logger=migrator t=2024-04-24T09:37:40.125267378Z level=info msg="Migration successfully executed" id="create alert_image table" duration=881.582µs grafana | logger=migrator t=2024-04-24T09:37:40.132272622Z level=info msg="Executing migration" id="add unique index on token to alert_image table" grafana | logger=migrator t=2024-04-24T09:37:40.133398042Z level=info msg="Migration successfully executed" id="add unique index on token to alert_image table" duration=1.125681ms grafana | logger=migrator t=2024-04-24T09:37:40.13887489Z level=info msg="Executing migration" id="support longer URLs in alert_image table" grafana | logger=migrator t=2024-04-24T09:37:40.138973503Z level=info msg="Migration successfully executed" id="support longer URLs in alert_image table" duration=95.683µs grafana | logger=migrator t=2024-04-24T09:37:40.145002951Z level=info msg="Executing migration" id=create_alert_configuration_history_table grafana | logger=migrator t=2024-04-24T09:37:40.146609439Z level=info msg="Migration successfully executed" id=create_alert_configuration_history_table duration=1.607418ms grafana | logger=migrator t=2024-04-24T09:37:40.154866868Z level=info msg="Executing migration" id="drop non-unique orgID index on alert_configuration" grafana | logger=migrator t=2024-04-24T09:37:40.155991178Z level=info msg="Migration successfully executed" id="drop non-unique orgID index on alert_configuration" duration=1.124601ms grafana | logger=migrator t=2024-04-24T09:37:40.162256455Z level=info msg="Executing migration" id="drop unique orgID index on alert_configuration if exists" grafana | logger=migrator t=2024-04-24T09:37:40.162877617Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop unique orgID index on alert_configuration if exists" grafana | logger=migrator t=2024-04-24T09:37:40.16932982Z level=info msg="Executing migration" id="extract alertmanager configuration history to separate table" grafana | logger=migrator t=2024-04-24T09:37:40.170053066Z level=info msg="Migration successfully executed" id="extract alertmanager configuration history to separate table" duration=720.405µs grafana | logger=migrator t=2024-04-24T09:37:40.174188146Z level=info msg="Executing migration" id="add unique index on orgID to alert_configuration" grafana | logger=migrator t=2024-04-24T09:37:40.175811744Z level=info msg="Migration successfully executed" id="add unique index on orgID to alert_configuration" duration=1.623328ms grafana | logger=migrator t=2024-04-24T09:37:40.179689254Z level=info msg="Executing migration" id="add last_applied column to alert_configuration_history" grafana | logger=migrator t=2024-04-24T09:37:40.186917905Z level=info msg="Migration successfully executed" id="add last_applied column to alert_configuration_history" duration=7.227871ms grafana | logger=migrator t=2024-04-24T09:37:40.192081701Z level=info msg="Executing migration" id="create library_element table v1" grafana | logger=migrator t=2024-04-24T09:37:40.193305726Z level=info msg="Migration successfully executed" id="create library_element table v1" duration=1.224465ms grafana | logger=migrator t=2024-04-24T09:37:40.19728955Z level=info msg="Executing migration" id="add index library_element org_id-folder_id-name-kind" grafana | logger=migrator t=2024-04-24T09:37:40.19841979Z level=info msg="Migration successfully executed" id="add index library_element org_id-folder_id-name-kind" duration=1.134321ms grafana | logger=migrator t=2024-04-24T09:37:40.202719856Z level=info msg="Executing migration" id="create library_element_connection table v1" grafana | logger=migrator t=2024-04-24T09:37:40.203633419Z level=info msg="Migration successfully executed" id="create library_element_connection table v1" duration=913.502µs grafana | logger=migrator t=2024-04-24T09:37:40.208179413Z level=info msg="Executing migration" id="add index library_element_connection element_id-kind-connection_id" grafana | logger=migrator t=2024-04-24T09:37:40.209292403Z level=info msg="Migration successfully executed" id="add index library_element_connection element_id-kind-connection_id" duration=1.11265ms grafana | logger=migrator t=2024-04-24T09:37:40.214007554Z level=info msg="Executing migration" id="add unique index library_element org_id_uid" grafana | logger=migrator t=2024-04-24T09:37:40.215202126Z level=info msg="Migration successfully executed" id="add unique index library_element org_id_uid" duration=1.196123ms grafana | logger=migrator t=2024-04-24T09:37:40.218966573Z level=info msg="Executing migration" id="increase max description length to 2048" grafana | logger=migrator t=2024-04-24T09:37:40.218996324Z level=info msg="Migration successfully executed" id="increase max description length to 2048" duration=30.771µs grafana | logger=migrator t=2024-04-24T09:37:40.223402533Z level=info msg="Executing migration" id="alter library_element model to mediumtext" grafana | logger=migrator t=2024-04-24T09:37:40.223481455Z level=info msg="Migration successfully executed" id="alter library_element model to mediumtext" duration=79.733µs grafana | logger=migrator t=2024-04-24T09:37:40.280813107Z level=info msg="Executing migration" id="clone move dashboard alerts to unified alerting" grafana | logger=migrator t=2024-04-24T09:37:40.281629976Z level=info msg="Migration successfully executed" id="clone move dashboard alerts to unified alerting" duration=821.95µs grafana | logger=migrator t=2024-04-24T09:37:40.286262483Z level=info msg="Executing migration" id="create data_keys table" grafana | logger=migrator t=2024-04-24T09:37:40.288156151Z level=info msg="Migration successfully executed" id="create data_keys table" duration=1.892808ms grafana | logger=migrator t=2024-04-24T09:37:40.293062189Z level=info msg="Executing migration" id="create secrets table" grafana | logger=migrator t=2024-04-24T09:37:40.294274163Z level=info msg="Migration successfully executed" id="create secrets table" duration=1.214855ms grafana | logger=migrator t=2024-04-24T09:37:40.299754301Z level=info msg="Executing migration" id="rename data_keys name column to id" grafana | logger=migrator t=2024-04-24T09:37:40.33296766Z level=info msg="Migration successfully executed" id="rename data_keys name column to id" duration=33.205538ms grafana | logger=migrator t=2024-04-24T09:37:40.337412751Z level=info msg="Executing migration" id="add name column into data_keys" grafana | logger=migrator t=2024-04-24T09:37:40.343561723Z level=info msg="Migration successfully executed" id="add name column into data_keys" duration=6.144711ms grafana | logger=migrator t=2024-04-24T09:37:40.353325266Z level=info msg="Executing migration" id="copy data_keys id column values into name" grafana | logger=migrator t=2024-04-24T09:37:40.353758551Z level=info msg="Migration successfully executed" id="copy data_keys id column values into name" duration=438.206µs grafana | logger=migrator t=2024-04-24T09:37:40.357794497Z level=info msg="Executing migration" id="rename data_keys name column to label" grafana | logger=migrator t=2024-04-24T09:37:40.393526977Z level=info msg="Migration successfully executed" id="rename data_keys name column to label" duration=35.729161ms grafana | logger=migrator t=2024-04-24T09:37:40.404531725Z level=info msg="Executing migration" id="rename data_keys id column back to name" zookeeper | ===> User zookeeper | uid=1000(appuser) gid=1000(appuser) groups=1000(appuser) zookeeper | ===> Configuring ... zookeeper | ===> Running preflight checks ... zookeeper | ===> Check if /var/lib/zookeeper/data is writable ... zookeeper | ===> Check if /var/lib/zookeeper/log is writable ... zookeeper | ===> Launching ... zookeeper | ===> Launching zookeeper ... zookeeper | [2024-04-24 09:37:44,445] INFO Reading configuration from: /etc/kafka/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,452] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,452] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,452] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,452] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,453] INFO autopurge.snapRetainCount set to 3 (org.apache.zookeeper.server.DatadirCleanupManager) zookeeper | [2024-04-24 09:37:44,454] INFO autopurge.purgeInterval set to 0 (org.apache.zookeeper.server.DatadirCleanupManager) zookeeper | [2024-04-24 09:37:44,454] INFO Purge task is not scheduled. (org.apache.zookeeper.server.DatadirCleanupManager) zookeeper | [2024-04-24 09:37:44,454] WARN Either no config or no quorum defined in config, running in standalone mode (org.apache.zookeeper.server.quorum.QuorumPeerMain) zookeeper | [2024-04-24 09:37:44,455] INFO Log4j 1.2 jmx support not found; jmx disabled. (org.apache.zookeeper.jmx.ManagedUtil) zookeeper | [2024-04-24 09:37:44,455] INFO Reading configuration from: /etc/kafka/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,456] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,456] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,456] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,456] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) zookeeper | [2024-04-24 09:37:44,456] INFO Starting server (org.apache.zookeeper.server.ZooKeeperServerMain) zookeeper | [2024-04-24 09:37:44,467] INFO ServerMetrics initialized with provider org.apache.zookeeper.metrics.impl.DefaultMetricsProvider@77eca502 (org.apache.zookeeper.server.ServerMetrics) zookeeper | [2024-04-24 09:37:44,469] INFO ACL digest algorithm is: SHA1 (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) zookeeper | [2024-04-24 09:37:44,469] INFO zookeeper.DigestAuthenticationProvider.enabled = true (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) zookeeper | [2024-04-24 09:37:44,471] INFO zookeeper.snapshot.trust.empty : false (org.apache.zookeeper.server.persistence.FileTxnSnapLog) zookeeper | [2024-04-24 09:37:44,481] INFO (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,481] INFO ______ _ (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,481] INFO |___ / | | (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,481] INFO / / ___ ___ | | __ ___ ___ _ __ ___ _ __ (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,481] INFO / / / _ \ / _ \ | |/ / / _ \ / _ \ | '_ \ / _ \ | '__| (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,481] INFO / /__ | (_) | | (_) | | < | __/ | __/ | |_) | | __/ | | (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,481] INFO /_____| \___/ \___/ |_|\_\ \___| \___| | .__/ \___| |_| (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,481] INFO | | (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,481] INFO |_| (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,481] INFO (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:zookeeper.version=3.8.4-9316c2a7a97e1666d8f4593f34dd6fc36ecc436c, built on 2024-02-12 22:16 UTC (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:host.name=737a1873bb73 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:java.version=11.0.22 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:java.vendor=Azul Systems, Inc. (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:java.home=/usr/lib/jvm/java-11-zulu-openjdk-ca (org.apache.zookeeper.server.ZooKeeperServer) grafana | logger=migrator t=2024-04-24T09:37:40.473053989Z level=info msg="Migration successfully executed" id="rename data_keys id column back to name" duration=68.519224ms grafana | logger=migrator t=2024-04-24T09:37:40.48468706Z level=info msg="Executing migration" id="create kv_store table v1" grafana | logger=migrator t=2024-04-24T09:37:40.486547197Z level=info msg="Migration successfully executed" id="create kv_store table v1" duration=1.863537ms grafana | logger=migrator t=2024-04-24T09:37:40.497306616Z level=info msg="Executing migration" id="add index kv_store.org_id-namespace-key" grafana | logger=migrator t=2024-04-24T09:37:40.49879303Z level=info msg="Migration successfully executed" id="add index kv_store.org_id-namespace-key" duration=1.488654ms grafana | logger=migrator t=2024-04-24T09:37:40.524496838Z level=info msg="Executing migration" id="update dashboard_uid and panel_id from existing annotations" grafana | logger=migrator t=2024-04-24T09:37:40.525069079Z level=info msg="Migration successfully executed" id="update dashboard_uid and panel_id from existing annotations" duration=576.19µs grafana | logger=migrator t=2024-04-24T09:37:40.570633284Z level=info msg="Executing migration" id="create permission table" grafana | logger=migrator t=2024-04-24T09:37:40.572930277Z level=info msg="Migration successfully executed" id="create permission table" duration=2.299573ms grafana | logger=migrator t=2024-04-24T09:37:40.582184381Z level=info msg="Executing migration" id="add unique index permission.role_id" grafana | logger=migrator t=2024-04-24T09:37:40.583632844Z level=info msg="Migration successfully executed" id="add unique index permission.role_id" duration=1.449703ms grafana | logger=migrator t=2024-04-24T09:37:40.594413943Z level=info msg="Executing migration" id="add unique index role_id_action_scope" grafana | logger=migrator t=2024-04-24T09:37:40.595826044Z level=info msg="Migration successfully executed" id="add unique index role_id_action_scope" duration=1.413851ms grafana | logger=migrator t=2024-04-24T09:37:40.606727818Z level=info msg="Executing migration" id="create role table" grafana | logger=migrator t=2024-04-24T09:37:40.608897156Z level=info msg="Migration successfully executed" id="create role table" duration=2.174408ms grafana | logger=migrator t=2024-04-24T09:37:40.615417072Z level=info msg="Executing migration" id="add column display_name" grafana | logger=migrator t=2024-04-24T09:37:40.621946108Z level=info msg="Migration successfully executed" id="add column display_name" duration=6.505766ms grafana | logger=migrator t=2024-04-24T09:37:40.625578089Z level=info msg="Executing migration" id="add column group_name" grafana | logger=migrator t=2024-04-24T09:37:40.634553923Z level=info msg="Migration successfully executed" id="add column group_name" duration=8.970644ms grafana | logger=migrator t=2024-04-24T09:37:40.642472539Z level=info msg="Executing migration" id="add index role.org_id" grafana | logger=migrator t=2024-04-24T09:37:40.64363002Z level=info msg="Migration successfully executed" id="add index role.org_id" duration=1.161431ms grafana | logger=migrator t=2024-04-24T09:37:40.648295499Z level=info msg="Executing migration" id="add unique index role_org_id_name" grafana | logger=migrator t=2024-04-24T09:37:40.650343934Z level=info msg="Migration successfully executed" id="add unique index role_org_id_name" duration=2.052964ms grafana | logger=migrator t=2024-04-24T09:37:40.65773523Z level=info msg="Executing migration" id="add index role_org_id_uid" grafana | logger=migrator t=2024-04-24T09:37:40.658979275Z level=info msg="Migration successfully executed" id="add index role_org_id_uid" duration=1.255056ms grafana | logger=migrator t=2024-04-24T09:37:40.665713528Z level=info msg="Executing migration" id="create team role table" grafana | logger=migrator t=2024-04-24T09:37:40.66712336Z level=info msg="Migration successfully executed" id="create team role table" duration=1.410762ms grafana | logger=migrator t=2024-04-24T09:37:40.67323249Z level=info msg="Executing migration" id="add index team_role.org_id" grafana | logger=migrator t=2024-04-24T09:37:40.675496502Z level=info msg="Migration successfully executed" id="add index team_role.org_id" duration=2.264772ms grafana | logger=migrator t=2024-04-24T09:37:40.685404279Z level=info msg="Executing migration" id="add unique index team_role_org_id_team_id_role_id" grafana | logger=migrator t=2024-04-24T09:37:40.686827731Z level=info msg="Migration successfully executed" id="add unique index team_role_org_id_team_id_role_id" duration=1.426442ms grafana | logger=migrator t=2024-04-24T09:37:40.691168628Z level=info msg="Executing migration" id="add index team_role.team_id" grafana | logger=migrator t=2024-04-24T09:37:40.692033219Z level=info msg="Migration successfully executed" id="add index team_role.team_id" duration=863.752µs grafana | logger=migrator t=2024-04-24T09:37:40.697020629Z level=info msg="Executing migration" id="create user role table" grafana | logger=migrator t=2024-04-24T09:37:40.698073737Z level=info msg="Migration successfully executed" id="create user role table" duration=1.052148ms grafana | logger=migrator t=2024-04-24T09:37:40.704178038Z level=info msg="Executing migration" id="add index user_role.org_id" grafana | logger=migrator t=2024-04-24T09:37:40.705715734Z level=info msg="Migration successfully executed" id="add index user_role.org_id" duration=1.541796ms grafana | logger=migrator t=2024-04-24T09:37:40.710658892Z level=info msg="Executing migration" id="add unique index user_role_org_id_user_id_role_id" grafana | logger=migrator t=2024-04-24T09:37:40.711771172Z level=info msg="Migration successfully executed" id="add unique index user_role_org_id_user_id_role_id" duration=1.115541ms grafana | logger=migrator t=2024-04-24T09:37:40.722985297Z level=info msg="Executing migration" id="add index user_role.user_id" grafana | logger=migrator t=2024-04-24T09:37:40.724412048Z level=info msg="Migration successfully executed" id="add index user_role.user_id" duration=1.430391ms grafana | logger=migrator t=2024-04-24T09:37:40.731151042Z level=info msg="Executing migration" id="create builtin role table" grafana | logger=migrator t=2024-04-24T09:37:40.73220844Z level=info msg="Migration successfully executed" id="create builtin role table" duration=1.058418ms grafana | logger=migrator t=2024-04-24T09:37:40.737769341Z level=info msg="Executing migration" id="add index builtin_role.role_id" grafana | logger=migrator t=2024-04-24T09:37:40.739759613Z level=info msg="Migration successfully executed" id="add index builtin_role.role_id" duration=1.987822ms grafana | logger=migrator t=2024-04-24T09:37:40.812241361Z level=info msg="Executing migration" id="add index builtin_role.name" grafana | logger=migrator t=2024-04-24T09:37:40.814001554Z level=info msg="Migration successfully executed" id="add index builtin_role.name" duration=1.764433ms grafana | logger=migrator t=2024-04-24T09:37:40.823325301Z level=info msg="Executing migration" id="Add column org_id to builtin_role table" grafana | logger=migrator t=2024-04-24T09:37:40.831424113Z level=info msg="Migration successfully executed" id="Add column org_id to builtin_role table" duration=8.099573ms grafana | logger=migrator t=2024-04-24T09:37:40.835664047Z level=info msg="Executing migration" id="add index builtin_role.org_id" grafana | logger=migrator t=2024-04-24T09:37:40.837021405Z level=info msg="Migration successfully executed" id="add index builtin_role.org_id" duration=1.359328ms grafana | logger=migrator t=2024-04-24T09:37:40.841302701Z level=info msg="Executing migration" id="add unique index builtin_role_org_id_role_id_role" grafana | logger=migrator t=2024-04-24T09:37:40.843144207Z level=info msg="Migration successfully executed" id="add unique index builtin_role_org_id_role_id_role" duration=1.843457ms grafana | logger=migrator t=2024-04-24T09:37:40.848555793Z level=info msg="Executing migration" id="Remove unique index role_org_id_uid" zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:java.class.path=/usr/bin/../share/java/kafka/scala-logging_2.13-3.9.4.jar:/usr/bin/../share/java/kafka/jersey-common-2.39.1.jar:/usr/bin/../share/java/kafka/swagger-annotations-2.2.8.jar:/usr/bin/../share/java/kafka/connect-json-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-storage-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/commons-validator-1.7.jar:/usr/bin/../share/java/kafka/javax.servlet-api-3.1.0.jar:/usr/bin/../share/java/kafka/aopalliance-repackaged-2.6.1.jar:/usr/bin/../share/java/kafka/netty-transport-4.1.100.Final.jar:/usr/bin/../share/java/kafka/rocksdbjni-7.9.2.jar:/usr/bin/../share/java/kafka/jetty-http-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka-log4j-appender-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/commons-io-2.11.0.jar:/usr/bin/../share/java/kafka/kafka-clients-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/javax.activation-api-1.2.0.jar:/usr/bin/../share/java/kafka/jetty-security-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/commons-cli-1.4.jar:/usr/bin/../share/java/kafka/jetty-server-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/slf4j-reload4j-1.7.36.jar:/usr/bin/../share/java/kafka/scala-reflect-2.13.11.jar:/usr/bin/../share/java/kafka/kafka-server-common-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-group-coordinator-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-buffer-4.1.100.Final.jar:/usr/bin/../share/java/kafka/connect-runtime-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/trogdor-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.ws.rs-api-2.1.6.jar:/usr/bin/../share/java/kafka/jakarta.annotation-api-1.3.5.jar:/usr/bin/../share/java/kafka/kafka-streams-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/scala-java8-compat_2.13-1.0.2.jar:/usr/bin/../share/java/kafka/javax.ws.rs-api-2.1.1.jar:/usr/bin/../share/java/kafka/zookeeper-jute-3.8.4.jar:/usr/bin/../share/java/kafka/netty-resolver-4.1.100.Final.jar:/usr/bin/../share/java/kafka/hk2-api-2.6.1.jar:/usr/bin/../share/java/kafka/kafka-storage-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/zstd-jni-1.5.5-1.jar:/usr/bin/../share/java/kafka/jackson-dataformat-csv-2.13.5.jar:/usr/bin/../share/java/kafka/kafka.jar:/usr/bin/../share/java/kafka/scala-library-2.13.11.jar:/usr/bin/../share/java/kafka/jakarta.inject-2.6.1.jar:/usr/bin/../share/java/kafka/jakarta.xml.bind-api-2.3.3.jar:/usr/bin/../share/java/kafka/jetty-continuation-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/connect-transforms-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jose4j-0.9.4.jar:/usr/bin/../share/java/kafka/hk2-locator-2.6.1.jar:/usr/bin/../share/java/kafka/reflections-0.10.2.jar:/usr/bin/../share/java/kafka/slf4j-api-1.7.36.jar:/usr/bin/../share/java/kafka/paranamer-2.8.jar:/usr/bin/../share/java/kafka/commons-beanutils-1.9.4.jar:/usr/bin/../share/java/kafka/jaxb-api-2.3.1.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-2.39.1.jar:/usr/bin/../share/java/kafka/hk2-utils-2.6.1.jar:/usr/bin/../share/java/kafka/jackson-module-scala_2.13-2.13.5.jar:/usr/bin/../share/java/kafka/reload4j-1.2.25.jar:/usr/bin/../share/java/kafka/connect-mirror-client-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-servlet-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jackson-core-2.13.5.jar:/usr/bin/../share/java/kafka/jersey-hk2-2.39.1.jar:/usr/bin/../share/java/kafka/jackson-databind-2.13.5.jar:/usr/bin/../share/java/kafka/jetty-servlets-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/jersey-client-2.39.1.jar:/usr/bin/../share/java/kafka/osgi-resource-locator-1.0.3.jar:/usr/bin/../share/java/kafka/commons-digester-2.1.jar:/usr/bin/../share/java/kafka/netty-transport-native-epoll-4.1.100.Final.jar:/usr/bin/../share/java/kafka/argparse4j-0.7.0.jar:/usr/bin/../share/java/kafka/connect-mirror-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jackson-datatype-jdk8-2.13.5.jar:/usr/bin/../share/java/kafka/audience-annotations-0.12.0.jar:/usr/bin/../share/java/kafka/jackson-module-jaxb-annotations-2.13.5.jar:/usr/bin/../share/java/kafka/kafka-raft-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/javax.annotation-api-1.3.2.jar:/usr/bin/../share/java/kafka/maven-artifact-3.8.8.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-json-provider-2.13.5.jar:/usr/bin/../share/java/kafka/jakarta.validation-api-2.0.2.jar:/usr/bin/../share/java/kafka/zookeeper-3.8.4.jar:/usr/bin/../share/java/kafka/jersey-server-2.39.1.jar:/usr/bin/../share/java/kafka/commons-lang3-3.8.1.jar:/usr/bin/../share/java/kafka/jopt-simple-5.0.4.jar:/usr/bin/../share/java/kafka/error_prone_annotations-2.10.0.jar:/usr/bin/../share/java/kafka/lz4-java-1.8.0.jar:/usr/bin/../share/java/kafka/jetty-util-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka-tools-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jakarta.activation-api-1.2.2.jar:/usr/bin/../share/java/kafka/jersey-container-servlet-core-2.39.1.jar:/usr/bin/../share/java/kafka/checker-qual-3.19.0.jar:/usr/bin/../share/java/kafka/kafka-metadata-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/pcollections-4.0.1.jar:/usr/bin/../share/java/kafka/jackson-jaxrs-base-2.13.5.jar:/usr/bin/../share/java/kafka/commons-logging-1.2.jar:/usr/bin/../share/java/kafka/jsr305-3.0.2.jar:/usr/bin/../share/java/kafka/jetty-io-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/netty-codec-4.1.100.Final.jar:/usr/bin/../share/java/kafka/netty-transport-native-unix-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/scala-collection-compat_2.13-2.10.0.jar:/usr/bin/../share/java/kafka/metrics-core-2.2.0.jar:/usr/bin/../share/java/kafka/kafka-streams-test-utils-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/kafka-streams-examples-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-handler-4.1.100.Final.jar:/usr/bin/../share/java/kafka/commons-collections-3.2.2.jar:/usr/bin/../share/java/kafka/javassist-3.29.2-GA.jar:/usr/bin/../share/java/kafka/caffeine-2.9.3.jar:/usr/bin/../share/java/kafka/plexus-utils-3.3.1.jar:/usr/bin/../share/java/kafka/kafka-tools-api-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/activation-1.1.1.jar:/usr/bin/../share/java/kafka/netty-common-4.1.100.Final.jar:/usr/bin/../share/java/kafka/kafka-streams-scala_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/metrics-core-4.1.12.1.jar:/usr/bin/../share/java/kafka/jline-3.25.1.jar:/usr/bin/../share/java/kafka/jetty-client-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/kafka_2.13-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/connect-basic-auth-extension-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/jetty-util-ajax-9.4.54.v20240208.jar:/usr/bin/../share/java/kafka/snappy-java-1.1.10.5.jar:/usr/bin/../share/java/kafka/kafka-shell-7.6.1-ccs.jar:/usr/bin/../share/java/kafka/netty-transport-classes-epoll-4.1.100.Final.jar:/usr/bin/../share/java/confluent-telemetry/* (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:java.io.tmpdir=/tmp (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:java.compiler= (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:os.name=Linux (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:os.arch=amd64 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:os.version=4.15.0-192-generic (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,482] INFO Server environment:user.name=appuser (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO Server environment:user.home=/home/appuser (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO Server environment:user.dir=/home/appuser (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO Server environment:os.memory.free=491MB (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO Server environment:os.memory.max=512MB (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO Server environment:os.memory.total=512MB (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO zookeeper.enableEagerACLCheck = false (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO zookeeper.digest.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO zookeeper.closeSessionTxn.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO zookeeper.flushDelay = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO zookeeper.maxWriteQueuePollTime = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO zookeeper.maxBatchSize=1000 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,483] INFO zookeeper.intBufferStartingSizeBytes = 1024 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,484] INFO Weighed connection throttling is disabled (org.apache.zookeeper.server.BlueThrottle) zookeeper | [2024-04-24 09:37:44,485] INFO minSessionTimeout set to 4000 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,485] INFO maxSessionTimeout set to 40000 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,486] INFO getData response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) zookeeper | [2024-04-24 09:37:44,486] INFO getChildren response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) zookeeper | [2024-04-24 09:37:44,487] INFO zookeeper.pathStats.slotCapacity = 60 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2024-04-24 09:37:44,487] INFO zookeeper.pathStats.slotDuration = 15 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2024-04-24 09:37:44,487] INFO zookeeper.pathStats.maxDepth = 6 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2024-04-24 09:37:44,487] INFO zookeeper.pathStats.initialDelay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2024-04-24 09:37:44,487] INFO zookeeper.pathStats.delay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2024-04-24 09:37:44,488] INFO zookeeper.pathStats.enabled = false (org.apache.zookeeper.server.util.RequestPathMetricsCollector) zookeeper | [2024-04-24 09:37:44,490] INFO The max bytes for all large requests are set to 104857600 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,490] INFO The large request threshold is set to -1 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,490] INFO zookeeper.enforce.auth.enabled = false (org.apache.zookeeper.server.AuthenticationHelper) zookeeper | [2024-04-24 09:37:44,491] INFO zookeeper.enforce.auth.schemes = [] (org.apache.zookeeper.server.AuthenticationHelper) zookeeper | [2024-04-24 09:37:44,491] INFO Created server with tickTime 2000 ms minSessionTimeout 4000 ms maxSessionTimeout 40000 ms clientPortListenBacklog -1 datadir /var/lib/zookeeper/log/version-2 snapdir /var/lib/zookeeper/data/version-2 (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,514] INFO Logging initialized @620ms to org.eclipse.jetty.util.log.Slf4jLog (org.eclipse.jetty.util.log) zookeeper | [2024-04-24 09:37:44,625] WARN o.e.j.s.ServletContextHandler@6d5620ce{/,null,STOPPED} contextPath ends with /* (org.eclipse.jetty.server.handler.ContextHandler) zookeeper | [2024-04-24 09:37:44,625] WARN Empty contextPath (org.eclipse.jetty.server.handler.ContextHandler) zookeeper | [2024-04-24 09:37:44,653] INFO jetty-9.4.54.v20240208; built: 2024-02-08T19:42:39.027Z; git: cef3fbd6d736a21e7d541a5db490381d95a2047d; jvm 11.0.22+7-LTS (org.eclipse.jetty.server.Server) zookeeper | [2024-04-24 09:37:44,685] INFO DefaultSessionIdManager workerName=node0 (org.eclipse.jetty.server.session) zookeeper | [2024-04-24 09:37:44,686] INFO No SessionScavenger set, using defaults (org.eclipse.jetty.server.session) zookeeper | [2024-04-24 09:37:44,687] INFO node0 Scavenging every 660000ms (org.eclipse.jetty.server.session) zookeeper | [2024-04-24 09:37:44,690] WARN ServletContext@o.e.j.s.ServletContextHandler@6d5620ce{/,null,STARTING} has uncovered http methods for path: /* (org.eclipse.jetty.security.SecurityHandler) zookeeper | [2024-04-24 09:37:44,699] INFO Started o.e.j.s.ServletContextHandler@6d5620ce{/,null,AVAILABLE} (org.eclipse.jetty.server.handler.ContextHandler) zookeeper | [2024-04-24 09:37:44,719] INFO Started ServerConnector@4d1bf319{HTTP/1.1, (http/1.1)}{0.0.0.0:8080} (org.eclipse.jetty.server.AbstractConnector) zookeeper | [2024-04-24 09:37:44,719] INFO Started @825ms (org.eclipse.jetty.server.Server) zookeeper | [2024-04-24 09:37:44,719] INFO Started AdminServer on address 0.0.0.0, port 8080 and command URL /commands (org.apache.zookeeper.server.admin.JettyAdminServer) zookeeper | [2024-04-24 09:37:44,726] INFO Using org.apache.zookeeper.server.NIOServerCnxnFactory as server connection factory (org.apache.zookeeper.server.ServerCnxnFactory) zookeeper | [2024-04-24 09:37:44,727] WARN maxCnxns is not configured, using default value 0. (org.apache.zookeeper.server.ServerCnxnFactory) zookeeper | [2024-04-24 09:37:44,729] INFO Configuring NIO connection handler with 10s sessionless connection timeout, 2 selector thread(s), 16 worker threads, and 64 kB direct buffers. (org.apache.zookeeper.server.NIOServerCnxnFactory) zookeeper | [2024-04-24 09:37:44,732] INFO binding to port 0.0.0.0/0.0.0.0:2181 (org.apache.zookeeper.server.NIOServerCnxnFactory) zookeeper | [2024-04-24 09:37:44,767] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) zookeeper | [2024-04-24 09:37:44,767] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) zookeeper | [2024-04-24 09:37:44,769] INFO zookeeper.snapshotSizeFactor = 0.33 (org.apache.zookeeper.server.ZKDatabase) zookeeper | [2024-04-24 09:37:44,769] INFO zookeeper.commitLogCount=500 (org.apache.zookeeper.server.ZKDatabase) zookeeper | [2024-04-24 09:37:44,777] INFO zookeeper.snapshot.compression.method = CHECKED (org.apache.zookeeper.server.persistence.SnapStream) zookeeper | [2024-04-24 09:37:44,777] INFO Snapshotting: 0x0 to /var/lib/zookeeper/data/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) zookeeper | [2024-04-24 09:37:44,782] INFO Snapshot loaded in 13 ms, highest zxid is 0x0, digest is 1371985504 (org.apache.zookeeper.server.ZKDatabase) zookeeper | [2024-04-24 09:37:44,784] INFO Snapshotting: 0x0 to /var/lib/zookeeper/data/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) zookeeper | [2024-04-24 09:37:44,785] INFO Snapshot taken in 1 ms (org.apache.zookeeper.server.ZooKeeperServer) zookeeper | [2024-04-24 09:37:44,803] INFO PrepRequestProcessor (sid:0) started, reconfigEnabled=false (org.apache.zookeeper.server.PrepRequestProcessor) zookeeper | [2024-04-24 09:37:44,804] INFO zookeeper.request_throttler.shutdownTimeout = 10000 ms (org.apache.zookeeper.server.RequestThrottler) zookeeper | [2024-04-24 09:37:44,821] INFO Using checkIntervalMs=60000 maxPerMinute=10000 maxNeverUsedIntervalMs=0 (org.apache.zookeeper.server.ContainerManager) zookeeper | [2024-04-24 09:37:44,822] INFO ZooKeeper audit is disabled. (org.apache.zookeeper.audit.ZKAuditProvider) zookeeper | [2024-04-24 09:37:47,873] INFO Creating new log file: log.1 (org.apache.zookeeper.server.persistence.FileTxnLog) grafana | logger=migrator t=2024-04-24T09:37:40.849921082Z level=info msg="Migration successfully executed" id="Remove unique index role_org_id_uid" duration=1.406752ms grafana | logger=migrator t=2024-04-24T09:37:40.853933826Z level=info msg="Executing migration" id="add unique index role.uid" grafana | logger=migrator t=2024-04-24T09:37:40.854912172Z level=info msg="Migration successfully executed" id="add unique index role.uid" duration=978.346µs grafana | logger=migrator t=2024-04-24T09:37:40.85846296Z level=info msg="Executing migration" id="create seed assignment table" grafana | logger=migrator t=2024-04-24T09:37:40.859226128Z level=info msg="Migration successfully executed" id="create seed assignment table" duration=758.817µs grafana | logger=migrator t=2024-04-24T09:37:40.865329608Z level=info msg="Executing migration" id="add unique index builtin_role_role_name" grafana | logger=migrator t=2024-04-24T09:37:40.866526522Z level=info msg="Migration successfully executed" id="add unique index builtin_role_role_name" duration=1.198544ms grafana | logger=migrator t=2024-04-24T09:37:40.870450923Z level=info msg="Executing migration" id="add column hidden to role table" grafana | logger=migrator t=2024-04-24T09:37:40.878924679Z level=info msg="Migration successfully executed" id="add column hidden to role table" duration=8.468016ms grafana | logger=migrator t=2024-04-24T09:37:40.884256252Z level=info msg="Executing migration" id="permission kind migration" grafana | logger=migrator t=2024-04-24T09:37:40.892995267Z level=info msg="Migration successfully executed" id="permission kind migration" duration=8.736166ms grafana | logger=migrator t=2024-04-24T09:37:40.898159504Z level=info msg="Executing migration" id="permission attribute migration" grafana | logger=migrator t=2024-04-24T09:37:40.904822845Z level=info msg="Migration successfully executed" id="permission attribute migration" duration=6.668221ms grafana | logger=migrator t=2024-04-24T09:37:40.908655863Z level=info msg="Executing migration" id="permission identifier migration" grafana | logger=migrator t=2024-04-24T09:37:40.914758513Z level=info msg="Migration successfully executed" id="permission identifier migration" duration=6.1002ms grafana | logger=migrator t=2024-04-24T09:37:40.928006732Z level=info msg="Executing migration" id="add permission identifier index" grafana | logger=migrator t=2024-04-24T09:37:40.929416963Z level=info msg="Migration successfully executed" id="add permission identifier index" duration=1.412401ms grafana | logger=migrator t=2024-04-24T09:37:40.933323794Z level=info msg="Executing migration" id="add permission action scope role_id index" grafana | logger=migrator t=2024-04-24T09:37:40.935362867Z level=info msg="Migration successfully executed" id="add permission action scope role_id index" duration=2.038643ms grafana | logger=migrator t=2024-04-24T09:37:40.941052554Z level=info msg="Executing migration" id="remove permission role_id action scope index" grafana | logger=migrator t=2024-04-24T09:37:40.942318099Z level=info msg="Migration successfully executed" id="remove permission role_id action scope index" duration=1.271636ms grafana | logger=migrator t=2024-04-24T09:37:40.946088715Z level=info msg="Executing migration" id="create query_history table v1" grafana | logger=migrator t=2024-04-24T09:37:40.947125453Z level=info msg="Migration successfully executed" id="create query_history table v1" duration=1.036998ms grafana | logger=migrator t=2024-04-24T09:37:40.951992058Z level=info msg="Executing migration" id="add index query_history.org_id-created_by-datasource_uid" grafana | logger=migrator t=2024-04-24T09:37:40.953223273Z level=info msg="Migration successfully executed" id="add index query_history.org_id-created_by-datasource_uid" duration=1.231275ms grafana | logger=migrator t=2024-04-24T09:37:40.956691698Z level=info msg="Executing migration" id="alter table query_history alter column created_by type to bigint" grafana | logger=migrator t=2024-04-24T09:37:40.956802422Z level=info msg="Migration successfully executed" id="alter table query_history alter column created_by type to bigint" duration=111.914µs grafana | logger=migrator t=2024-04-24T09:37:40.959877633Z level=info msg="Executing migration" id="rbac disabled migrator" grafana | logger=migrator t=2024-04-24T09:37:40.959960366Z level=info msg="Migration successfully executed" id="rbac disabled migrator" duration=83.773µs grafana | logger=migrator t=2024-04-24T09:37:40.96363966Z level=info msg="Executing migration" id="teams permissions migration" grafana | logger=migrator t=2024-04-24T09:37:40.964437528Z level=info msg="Migration successfully executed" id="teams permissions migration" duration=801.659µs grafana | logger=migrator t=2024-04-24T09:37:40.969075056Z level=info msg="Executing migration" id="dashboard permissions" grafana | logger=migrator t=2024-04-24T09:37:40.969817872Z level=info msg="Migration successfully executed" id="dashboard permissions" duration=744.326µs grafana | logger=migrator t=2024-04-24T09:37:40.973754654Z level=info msg="Executing migration" id="dashboard permissions uid scopes" grafana | logger=migrator t=2024-04-24T09:37:40.97447319Z level=info msg="Migration successfully executed" id="dashboard permissions uid scopes" duration=718.946µs grafana | logger=migrator t=2024-04-24T09:37:40.978276548Z level=info msg="Executing migration" id="drop managed folder create actions" grafana | logger=migrator t=2024-04-24T09:37:40.978643491Z level=info msg="Migration successfully executed" id="drop managed folder create actions" duration=367.033µs grafana | logger=migrator t=2024-04-24T09:37:40.982570663Z level=info msg="Executing migration" id="alerting notification permissions" grafana | logger=migrator t=2024-04-24T09:37:40.983483866Z level=info msg="Migration successfully executed" id="alerting notification permissions" duration=913.834µs grafana | logger=migrator t=2024-04-24T09:37:40.988497376Z level=info msg="Executing migration" id="create query_history_star table v1" grafana | logger=migrator t=2024-04-24T09:37:40.990019282Z level=info msg="Migration successfully executed" id="create query_history_star table v1" duration=1.523606ms grafana | logger=migrator t=2024-04-24T09:37:40.994313827Z level=info msg="Executing migration" id="add index query_history.user_id-query_uid" grafana | logger=migrator t=2024-04-24T09:37:40.995690227Z level=info msg="Migration successfully executed" id="add index query_history.user_id-query_uid" duration=1.37914ms grafana | logger=migrator t=2024-04-24T09:37:40.999335978Z level=info msg="Executing migration" id="add column org_id in query_history_star" grafana | logger=migrator t=2024-04-24T09:37:41.007877546Z level=info msg="Migration successfully executed" id="add column org_id in query_history_star" duration=8.541758ms grafana | logger=migrator t=2024-04-24T09:37:41.012585166Z level=info msg="Executing migration" id="alter table query_history_star_mig column user_id type to bigint" grafana | logger=migrator t=2024-04-24T09:37:41.012656668Z level=info msg="Migration successfully executed" id="alter table query_history_star_mig column user_id type to bigint" duration=71.812µs grafana | logger=migrator t=2024-04-24T09:37:41.016267298Z level=info msg="Executing migration" id="create correlation table v1" grafana | logger=migrator t=2024-04-24T09:37:41.0171336Z level=info msg="Migration successfully executed" id="create correlation table v1" duration=865.962µs grafana | logger=migrator t=2024-04-24T09:37:41.022747852Z level=info msg="Executing migration" id="add index correlations.uid" grafana | logger=migrator t=2024-04-24T09:37:41.024618019Z level=info msg="Migration successfully executed" id="add index correlations.uid" duration=1.873836ms grafana | logger=migrator t=2024-04-24T09:37:41.029770824Z level=info msg="Executing migration" id="add index correlations.source_uid" kafka | [2024-04-24 09:38:19,019] INFO Created log for partition __consumer_offsets-40 in /var/lib/kafka/data/__consumer_offsets-40 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,019] INFO [Partition __consumer_offsets-40 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-40 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,020] INFO [Partition __consumer_offsets-40 broker=1] Log loaded for partition __consumer_offsets-40 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,020] INFO [Broker id=1] Leader __consumer_offsets-40 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,027] INFO [LogLoader partition=__consumer_offsets-47, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,027] INFO Created log for partition __consumer_offsets-47 in /var/lib/kafka/data/__consumer_offsets-47 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,028] INFO [Partition __consumer_offsets-47 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-47 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,028] INFO [Partition __consumer_offsets-47 broker=1] Log loaded for partition __consumer_offsets-47 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,028] INFO [Broker id=1] Leader __consumer_offsets-47 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,036] INFO [LogLoader partition=__consumer_offsets-17, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,036] INFO Created log for partition __consumer_offsets-17 in /var/lib/kafka/data/__consumer_offsets-17 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,036] INFO [Partition __consumer_offsets-17 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-17 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,037] INFO [Partition __consumer_offsets-17 broker=1] Log loaded for partition __consumer_offsets-17 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,037] INFO [Broker id=1] Leader __consumer_offsets-17 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,073] INFO [LogLoader partition=__consumer_offsets-32, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,074] INFO Created log for partition __consumer_offsets-32 in /var/lib/kafka/data/__consumer_offsets-32 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,075] INFO [Partition __consumer_offsets-32 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-32 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,075] INFO [Partition __consumer_offsets-32 broker=1] Log loaded for partition __consumer_offsets-32 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,075] INFO [Broker id=1] Leader __consumer_offsets-32 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,089] INFO [LogLoader partition=__consumer_offsets-37, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,090] INFO Created log for partition __consumer_offsets-37 in /var/lib/kafka/data/__consumer_offsets-37 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,091] INFO [Partition __consumer_offsets-37 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-37 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,091] INFO [Partition __consumer_offsets-37 broker=1] Log loaded for partition __consumer_offsets-37 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,091] INFO [Broker id=1] Leader __consumer_offsets-37 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,101] INFO [LogLoader partition=__consumer_offsets-7, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,102] INFO Created log for partition __consumer_offsets-7 in /var/lib/kafka/data/__consumer_offsets-7 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,102] INFO [Partition __consumer_offsets-7 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-7 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,102] INFO [Partition __consumer_offsets-7 broker=1] Log loaded for partition __consumer_offsets-7 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,103] INFO [Broker id=1] Leader __consumer_offsets-7 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,110] INFO [LogLoader partition=__consumer_offsets-22, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,111] INFO Created log for partition __consumer_offsets-22 in /var/lib/kafka/data/__consumer_offsets-22 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,111] INFO [Partition __consumer_offsets-22 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-22 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,111] INFO [Partition __consumer_offsets-22 broker=1] Log loaded for partition __consumer_offsets-22 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,112] INFO [Broker id=1] Leader __consumer_offsets-22 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) grafana | logger=migrator t=2024-04-24T09:37:41.031100813Z level=info msg="Migration successfully executed" id="add index correlations.source_uid" duration=1.331209ms grafana | logger=migrator t=2024-04-24T09:37:41.035009523Z level=info msg="Executing migration" id="add correlation config column" grafana | logger=migrator t=2024-04-24T09:37:41.046790827Z level=info msg="Migration successfully executed" id="add correlation config column" duration=11.778994ms grafana | logger=migrator t=2024-04-24T09:37:41.050959117Z level=info msg="Executing migration" id="drop index IDX_correlation_uid - v1" grafana | logger=migrator t=2024-04-24T09:37:41.051939252Z level=info msg="Migration successfully executed" id="drop index IDX_correlation_uid - v1" duration=981.325µs grafana | logger=migrator t=2024-04-24T09:37:41.056322191Z level=info msg="Executing migration" id="drop index IDX_correlation_source_uid - v1" grafana | logger=migrator t=2024-04-24T09:37:41.057520903Z level=info msg="Migration successfully executed" id="drop index IDX_correlation_source_uid - v1" duration=1.198042ms grafana | logger=migrator t=2024-04-24T09:37:41.061193995Z level=info msg="Executing migration" id="Rename table correlation to correlation_tmp_qwerty - v1" grafana | logger=migrator t=2024-04-24T09:37:41.083101274Z level=info msg="Migration successfully executed" id="Rename table correlation to correlation_tmp_qwerty - v1" duration=21.902339ms grafana | logger=migrator t=2024-04-24T09:37:41.087114839Z level=info msg="Executing migration" id="create correlation v2" grafana | logger=migrator t=2024-04-24T09:37:41.088046732Z level=info msg="Migration successfully executed" id="create correlation v2" duration=931.943µs grafana | logger=migrator t=2024-04-24T09:37:41.09297324Z level=info msg="Executing migration" id="create index IDX_correlation_uid - v2" grafana | logger=migrator t=2024-04-24T09:37:41.094752663Z level=info msg="Migration successfully executed" id="create index IDX_correlation_uid - v2" duration=1.782073ms grafana | logger=migrator t=2024-04-24T09:37:41.098766748Z level=info msg="Executing migration" id="create index IDX_correlation_source_uid - v2" grafana | logger=migrator t=2024-04-24T09:37:41.100656266Z level=info msg="Migration successfully executed" id="create index IDX_correlation_source_uid - v2" duration=1.889078ms grafana | logger=migrator t=2024-04-24T09:37:41.105608634Z level=info msg="Executing migration" id="create index IDX_correlation_org_id - v2" grafana | logger=migrator t=2024-04-24T09:37:41.1068605Z level=info msg="Migration successfully executed" id="create index IDX_correlation_org_id - v2" duration=1.251686ms grafana | logger=migrator t=2024-04-24T09:37:41.112074168Z level=info msg="Executing migration" id="copy correlation v1 to v2" grafana | logger=migrator t=2024-04-24T09:37:41.112522954Z level=info msg="Migration successfully executed" id="copy correlation v1 to v2" duration=448.786µs grafana | logger=migrator t=2024-04-24T09:37:41.116378492Z level=info msg="Executing migration" id="drop correlation_tmp_qwerty" grafana | logger=migrator t=2024-04-24T09:37:41.1176903Z level=info msg="Migration successfully executed" id="drop correlation_tmp_qwerty" duration=1.310968ms grafana | logger=migrator t=2024-04-24T09:37:41.122567465Z level=info msg="Executing migration" id="add provisioning column" grafana | logger=migrator t=2024-04-24T09:37:41.131419033Z level=info msg="Migration successfully executed" id="add provisioning column" duration=8.848308ms grafana | logger=migrator t=2024-04-24T09:37:41.135094706Z level=info msg="Executing migration" id="create entity_events table" grafana | logger=migrator t=2024-04-24T09:37:41.136084282Z level=info msg="Migration successfully executed" id="create entity_events table" duration=995.125µs grafana | logger=migrator t=2024-04-24T09:37:41.140047415Z level=info msg="Executing migration" id="create dashboard public config v1" grafana | logger=migrator t=2024-04-24T09:37:41.141300259Z level=info msg="Migration successfully executed" id="create dashboard public config v1" duration=1.253204ms grafana | logger=migrator t=2024-04-24T09:37:41.146539758Z level=info msg="Executing migration" id="drop index UQE_dashboard_public_config_uid - v1" grafana | logger=migrator t=2024-04-24T09:37:41.14743928Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop index UQE_dashboard_public_config_uid - v1" grafana | logger=migrator t=2024-04-24T09:37:41.151432455Z level=info msg="Executing migration" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v1" grafana | logger=migrator t=2024-04-24T09:37:41.152291745Z level=warn msg="Skipping migration: Already executed, but not recorded in migration log" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v1" grafana | logger=migrator t=2024-04-24T09:37:41.156388992Z level=info msg="Executing migration" id="Drop old dashboard public config table" grafana | logger=migrator t=2024-04-24T09:37:41.157449811Z level=info msg="Migration successfully executed" id="Drop old dashboard public config table" duration=1.062419ms grafana | logger=migrator t=2024-04-24T09:37:41.162217362Z level=info msg="Executing migration" id="recreate dashboard public config v1" grafana | logger=migrator t=2024-04-24T09:37:41.163919464Z level=info msg="Migration successfully executed" id="recreate dashboard public config v1" duration=1.698961ms grafana | logger=migrator t=2024-04-24T09:37:41.168415146Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_uid - v1" grafana | logger=migrator t=2024-04-24T09:37:41.170409837Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_uid - v1" duration=1.997451ms grafana | logger=migrator t=2024-04-24T09:37:41.174869858Z level=info msg="Executing migration" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v1" grafana | logger=migrator t=2024-04-24T09:37:41.176284188Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v1" duration=1.41607ms grafana | logger=migrator t=2024-04-24T09:37:41.181184105Z level=info msg="Executing migration" id="drop index UQE_dashboard_public_config_uid - v2" grafana | logger=migrator t=2024-04-24T09:37:41.182485382Z level=info msg="Migration successfully executed" id="drop index UQE_dashboard_public_config_uid - v2" duration=1.303087ms grafana | logger=migrator t=2024-04-24T09:37:41.1865913Z level=info msg="Executing migration" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v2" grafana | logger=migrator t=2024-04-24T09:37:41.187878187Z level=info msg="Migration successfully executed" id="drop index IDX_dashboard_public_config_org_id_dashboard_uid - v2" duration=1.287146ms grafana | logger=migrator t=2024-04-24T09:37:41.191903841Z level=info msg="Executing migration" id="Drop public config table" grafana | logger=migrator t=2024-04-24T09:37:41.192846605Z level=info msg="Migration successfully executed" id="Drop public config table" duration=942.444µs grafana | logger=migrator t=2024-04-24T09:37:41.1976975Z level=info msg="Executing migration" id="Recreate dashboard public config v2" grafana | logger=migrator t=2024-04-24T09:37:41.199118421Z level=info msg="Migration successfully executed" id="Recreate dashboard public config v2" duration=1.419722ms grafana | logger=migrator t=2024-04-24T09:37:41.205973428Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_uid - v2" grafana | logger=migrator t=2024-04-24T09:37:41.207375769Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_uid - v2" duration=1.403671ms grafana | logger=migrator t=2024-04-24T09:37:41.21186667Z level=info msg="Executing migration" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v2" kafka | [2024-04-24 09:38:19,125] INFO [LogLoader partition=__consumer_offsets-29, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,127] INFO Created log for partition __consumer_offsets-29 in /var/lib/kafka/data/__consumer_offsets-29 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,127] INFO [Partition __consumer_offsets-29 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-29 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,128] INFO [Partition __consumer_offsets-29 broker=1] Log loaded for partition __consumer_offsets-29 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,128] INFO [Broker id=1] Leader __consumer_offsets-29 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,136] INFO [LogLoader partition=__consumer_offsets-44, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,137] INFO Created log for partition __consumer_offsets-44 in /var/lib/kafka/data/__consumer_offsets-44 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,137] INFO [Partition __consumer_offsets-44 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-44 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,138] INFO [Partition __consumer_offsets-44 broker=1] Log loaded for partition __consumer_offsets-44 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,138] INFO [Broker id=1] Leader __consumer_offsets-44 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,149] INFO [LogLoader partition=__consumer_offsets-14, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,150] INFO Created log for partition __consumer_offsets-14 in /var/lib/kafka/data/__consumer_offsets-14 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,150] INFO [Partition __consumer_offsets-14 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-14 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,150] INFO [Partition __consumer_offsets-14 broker=1] Log loaded for partition __consumer_offsets-14 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,150] INFO [Broker id=1] Leader __consumer_offsets-14 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,164] INFO [LogLoader partition=__consumer_offsets-23, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,165] INFO Created log for partition __consumer_offsets-23 in /var/lib/kafka/data/__consumer_offsets-23 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,165] INFO [Partition __consumer_offsets-23 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-23 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,166] INFO [Partition __consumer_offsets-23 broker=1] Log loaded for partition __consumer_offsets-23 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,166] INFO [Broker id=1] Leader __consumer_offsets-23 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,174] INFO [LogLoader partition=__consumer_offsets-38, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,175] INFO Created log for partition __consumer_offsets-38 in /var/lib/kafka/data/__consumer_offsets-38 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,175] INFO [Partition __consumer_offsets-38 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-38 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,175] INFO [Partition __consumer_offsets-38 broker=1] Log loaded for partition __consumer_offsets-38 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,175] INFO [Broker id=1] Leader __consumer_offsets-38 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,190] INFO [LogLoader partition=__consumer_offsets-8, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,191] INFO Created log for partition __consumer_offsets-8 in /var/lib/kafka/data/__consumer_offsets-8 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,191] INFO [Partition __consumer_offsets-8 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-8 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,191] INFO [Partition __consumer_offsets-8 broker=1] Log loaded for partition __consumer_offsets-8 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,191] INFO [Broker id=1] Leader __consumer_offsets-8 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,199] INFO [LogLoader partition=policy-pdp-pap-0, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,200] INFO Created log for partition policy-pdp-pap-0 in /var/lib/kafka/data/policy-pdp-pap-0 with properties {} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,200] INFO [Partition policy-pdp-pap-0 broker=1] No checkpointed highwatermark is found for partition policy-pdp-pap-0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,200] INFO [Partition policy-pdp-pap-0 broker=1] Log loaded for partition policy-pdp-pap-0 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,200] INFO [Broker id=1] Leader policy-pdp-pap-0 with topic id Some(lm-F3iiTSNeAYdJn1IimsA) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) grafana | logger=migrator t=2024-04-24T09:37:41.21323946Z level=info msg="Migration successfully executed" id="create index IDX_dashboard_public_config_org_id_dashboard_uid - v2" duration=1.3728ms grafana | logger=migrator t=2024-04-24T09:37:41.21686683Z level=info msg="Executing migration" id="create index UQE_dashboard_public_config_access_token - v2" grafana | logger=migrator t=2024-04-24T09:37:41.218155546Z level=info msg="Migration successfully executed" id="create index UQE_dashboard_public_config_access_token - v2" duration=1.288616ms grafana | logger=migrator t=2024-04-24T09:37:41.224249126Z level=info msg="Executing migration" id="Rename table dashboard_public_config to dashboard_public - v2" grafana | logger=migrator t=2024-04-24T09:37:41.252435141Z level=info msg="Migration successfully executed" id="Rename table dashboard_public_config to dashboard_public - v2" duration=28.184904ms grafana | logger=migrator t=2024-04-24T09:37:41.25658963Z level=info msg="Executing migration" id="add annotations_enabled column" grafana | logger=migrator t=2024-04-24T09:37:41.263205558Z level=info msg="Migration successfully executed" id="add annotations_enabled column" duration=6.612679ms grafana | logger=migrator t=2024-04-24T09:37:41.267102938Z level=info msg="Executing migration" id="add time_selection_enabled column" grafana | logger=migrator t=2024-04-24T09:37:41.27661052Z level=info msg="Migration successfully executed" id="add time_selection_enabled column" duration=9.499682ms grafana | logger=migrator t=2024-04-24T09:37:41.280860944Z level=info msg="Executing migration" id="delete orphaned public dashboards" grafana | logger=migrator t=2024-04-24T09:37:41.281190035Z level=info msg="Migration successfully executed" id="delete orphaned public dashboards" duration=331.851µs grafana | logger=migrator t=2024-04-24T09:37:41.330257301Z level=info msg="Executing migration" id="add share column" grafana | logger=migrator t=2024-04-24T09:37:41.338307351Z level=info msg="Migration successfully executed" id="add share column" duration=8.04852ms grafana | logger=migrator t=2024-04-24T09:37:41.377261024Z level=info msg="Executing migration" id="backfill empty share column fields with default of public" grafana | logger=migrator t=2024-04-24T09:37:41.379809155Z level=info msg="Migration successfully executed" id="backfill empty share column fields with default of public" duration=2.549292ms grafana | logger=migrator t=2024-04-24T09:37:41.384345539Z level=info msg="Executing migration" id="create file table" grafana | logger=migrator t=2024-04-24T09:37:41.385440518Z level=info msg="Migration successfully executed" id="create file table" duration=1.094029ms grafana | logger=migrator t=2024-04-24T09:37:41.390256222Z level=info msg="Executing migration" id="file table idx: path natural pk" grafana | logger=migrator t=2024-04-24T09:37:41.392340057Z level=info msg="Migration successfully executed" id="file table idx: path natural pk" duration=2.083805ms grafana | logger=migrator t=2024-04-24T09:37:41.396453495Z level=info msg="Executing migration" id="file table idx: parent_folder_path_hash fast folder retrieval" grafana | logger=migrator t=2024-04-24T09:37:41.398159056Z level=info msg="Migration successfully executed" id="file table idx: parent_folder_path_hash fast folder retrieval" duration=1.707051ms grafana | logger=migrator t=2024-04-24T09:37:41.401983524Z level=info msg="Executing migration" id="create file_meta table" grafana | logger=migrator t=2024-04-24T09:37:41.402922367Z level=info msg="Migration successfully executed" id="create file_meta table" duration=938.663µs grafana | logger=migrator t=2024-04-24T09:37:41.407914277Z level=info msg="Executing migration" id="file table idx: path key" grafana | logger=migrator t=2024-04-24T09:37:41.409235105Z level=info msg="Migration successfully executed" id="file table idx: path key" duration=1.320298ms grafana | logger=migrator t=2024-04-24T09:37:41.414423081Z level=info msg="Executing migration" id="set path collation in file table" grafana | logger=migrator t=2024-04-24T09:37:41.414619679Z level=info msg="Migration successfully executed" id="set path collation in file table" duration=195.968µs grafana | logger=migrator t=2024-04-24T09:37:41.418734907Z level=info msg="Executing migration" id="migrate contents column to mediumblob for MySQL" grafana | logger=migrator t=2024-04-24T09:37:41.418969715Z level=info msg="Migration successfully executed" id="migrate contents column to mediumblob for MySQL" duration=155.865µs grafana | logger=migrator t=2024-04-24T09:37:41.42324628Z level=info msg="Executing migration" id="managed permissions migration" grafana | logger=migrator t=2024-04-24T09:37:41.424357459Z level=info msg="Migration successfully executed" id="managed permissions migration" duration=1.110469ms grafana | logger=migrator t=2024-04-24T09:37:41.429577057Z level=info msg="Executing migration" id="managed folder permissions alert actions migration" grafana | logger=migrator t=2024-04-24T09:37:41.430197889Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions migration" duration=620.552µs grafana | logger=migrator t=2024-04-24T09:37:41.434417362Z level=info msg="Executing migration" id="RBAC action name migrator" grafana | logger=migrator t=2024-04-24T09:37:41.435956497Z level=info msg="Migration successfully executed" id="RBAC action name migrator" duration=1.538505ms grafana | logger=migrator t=2024-04-24T09:37:41.439824867Z level=info msg="Executing migration" id="Add UID column to playlist" grafana | logger=migrator t=2024-04-24T09:37:41.44881389Z level=info msg="Migration successfully executed" id="Add UID column to playlist" duration=8.985294ms grafana | logger=migrator t=2024-04-24T09:37:41.452624327Z level=info msg="Executing migration" id="Update uid column values in playlist" grafana | logger=migrator t=2024-04-24T09:37:41.452867715Z level=info msg="Migration successfully executed" id="Update uid column values in playlist" duration=243.028µs grafana | logger=migrator t=2024-04-24T09:37:41.456951473Z level=info msg="Executing migration" id="Add index for uid in playlist" grafana | logger=migrator t=2024-04-24T09:37:41.457947888Z level=info msg="Migration successfully executed" id="Add index for uid in playlist" duration=993.545µs grafana | logger=migrator t=2024-04-24T09:37:41.461706703Z level=info msg="Executing migration" id="update group index for alert rules" grafana | logger=migrator t=2024-04-24T09:37:41.462529114Z level=info msg="Migration successfully executed" id="update group index for alert rules" duration=823.231µs grafana | logger=migrator t=2024-04-24T09:37:41.466389792Z level=info msg="Executing migration" id="managed folder permissions alert actions repeated migration" grafana | logger=migrator t=2024-04-24T09:37:41.466914541Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions repeated migration" duration=526.739µs grafana | logger=migrator t=2024-04-24T09:37:41.472862765Z level=info msg="Executing migration" id="admin only folder/dashboard permission" grafana | logger=migrator t=2024-04-24T09:37:41.473451917Z level=info msg="Migration successfully executed" id="admin only folder/dashboard permission" duration=587.902µs grafana | logger=migrator t=2024-04-24T09:37:41.476846969Z level=info msg="Executing migration" id="add action column to seed_assignment" grafana | logger=migrator t=2024-04-24T09:37:41.488310802Z level=info msg="Migration successfully executed" id="add action column to seed_assignment" duration=11.464073ms grafana | logger=migrator t=2024-04-24T09:37:41.492040676Z level=info msg="Executing migration" id="add scope column to seed_assignment" grafana | logger=migrator t=2024-04-24T09:37:41.500545692Z level=info msg="Migration successfully executed" id="add scope column to seed_assignment" duration=8.503936ms grafana | logger=migrator t=2024-04-24T09:37:41.504292137Z level=info msg="Executing migration" id="remove unique index builtin_role_role_name before nullable update" kafka | [2024-04-24 09:38:19,214] INFO [LogLoader partition=__consumer_offsets-45, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,215] INFO Created log for partition __consumer_offsets-45 in /var/lib/kafka/data/__consumer_offsets-45 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,215] INFO [Partition __consumer_offsets-45 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-45 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,215] INFO [Partition __consumer_offsets-45 broker=1] Log loaded for partition __consumer_offsets-45 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,215] INFO [Broker id=1] Leader __consumer_offsets-45 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,225] INFO [LogLoader partition=__consumer_offsets-15, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,225] INFO Created log for partition __consumer_offsets-15 in /var/lib/kafka/data/__consumer_offsets-15 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,225] INFO [Partition __consumer_offsets-15 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-15 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,225] INFO [Partition __consumer_offsets-15 broker=1] Log loaded for partition __consumer_offsets-15 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,225] INFO [Broker id=1] Leader __consumer_offsets-15 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,235] INFO [LogLoader partition=__consumer_offsets-30, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,237] INFO Created log for partition __consumer_offsets-30 in /var/lib/kafka/data/__consumer_offsets-30 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,237] INFO [Partition __consumer_offsets-30 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-30 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,237] INFO [Partition __consumer_offsets-30 broker=1] Log loaded for partition __consumer_offsets-30 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,237] INFO [Broker id=1] Leader __consumer_offsets-30 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,246] INFO [LogLoader partition=__consumer_offsets-0, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,247] INFO Created log for partition __consumer_offsets-0 in /var/lib/kafka/data/__consumer_offsets-0 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,247] INFO [Partition __consumer_offsets-0 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,247] INFO [Partition __consumer_offsets-0 broker=1] Log loaded for partition __consumer_offsets-0 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,248] INFO [Broker id=1] Leader __consumer_offsets-0 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,255] INFO [LogLoader partition=__consumer_offsets-35, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,260] INFO Created log for partition __consumer_offsets-35 in /var/lib/kafka/data/__consumer_offsets-35 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,260] INFO [Partition __consumer_offsets-35 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-35 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,260] INFO [Partition __consumer_offsets-35 broker=1] Log loaded for partition __consumer_offsets-35 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,260] INFO [Broker id=1] Leader __consumer_offsets-35 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,266] INFO [LogLoader partition=__consumer_offsets-5, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,266] INFO Created log for partition __consumer_offsets-5 in /var/lib/kafka/data/__consumer_offsets-5 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,266] INFO [Partition __consumer_offsets-5 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-5 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,267] INFO [Partition __consumer_offsets-5 broker=1] Log loaded for partition __consumer_offsets-5 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,267] INFO [Broker id=1] Leader __consumer_offsets-5 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,276] INFO [LogLoader partition=__consumer_offsets-20, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,278] INFO Created log for partition __consumer_offsets-20 in /var/lib/kafka/data/__consumer_offsets-20 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,278] INFO [Partition __consumer_offsets-20 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-20 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,278] INFO [Partition __consumer_offsets-20 broker=1] Log loaded for partition __consumer_offsets-20 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,278] INFO [Broker id=1] Leader __consumer_offsets-20 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) grafana | logger=migrator t=2024-04-24T09:37:41.505167278Z level=info msg="Migration successfully executed" id="remove unique index builtin_role_role_name before nullable update" duration=875.081µs grafana | logger=migrator t=2024-04-24T09:37:41.509756164Z level=info msg="Executing migration" id="update seed_assignment role_name column to nullable" grafana | logger=migrator t=2024-04-24T09:37:41.58465929Z level=info msg="Migration successfully executed" id="update seed_assignment role_name column to nullable" duration=74.900016ms grafana | logger=migrator t=2024-04-24T09:37:41.589202194Z level=info msg="Executing migration" id="add unique index builtin_role_name back" grafana | logger=migrator t=2024-04-24T09:37:41.590088575Z level=info msg="Migration successfully executed" id="add unique index builtin_role_name back" duration=886.491µs grafana | logger=migrator t=2024-04-24T09:37:41.594340699Z level=info msg="Executing migration" id="add unique index builtin_role_action_scope" grafana | logger=migrator t=2024-04-24T09:37:41.59606987Z level=info msg="Migration successfully executed" id="add unique index builtin_role_action_scope" duration=1.721871ms grafana | logger=migrator t=2024-04-24T09:37:41.600314464Z level=info msg="Executing migration" id="add primary key to seed_assigment" grafana | logger=migrator t=2024-04-24T09:37:41.627408799Z level=info msg="Migration successfully executed" id="add primary key to seed_assigment" duration=27.096055ms grafana | logger=migrator t=2024-04-24T09:37:41.631655922Z level=info msg="Executing migration" id="add origin column to seed_assignment" grafana | logger=migrator t=2024-04-24T09:37:41.638124184Z level=info msg="Migration successfully executed" id="add origin column to seed_assignment" duration=6.467402ms grafana | logger=migrator t=2024-04-24T09:37:41.645130437Z level=info msg="Executing migration" id="add origin to plugin seed_assignment" grafana | logger=migrator t=2024-04-24T09:37:41.645559582Z level=info msg="Migration successfully executed" id="add origin to plugin seed_assignment" duration=428.745µs grafana | logger=migrator t=2024-04-24T09:37:41.652052136Z level=info msg="Executing migration" id="prevent seeding OnCall access" grafana | logger=migrator t=2024-04-24T09:37:41.652510602Z level=info msg="Migration successfully executed" id="prevent seeding OnCall access" duration=456.806µs grafana | logger=migrator t=2024-04-24T09:37:41.658708186Z level=info msg="Executing migration" id="managed folder permissions alert actions repeated fixed migration" grafana | logger=migrator t=2024-04-24T09:37:41.659153822Z level=info msg="Migration successfully executed" id="managed folder permissions alert actions repeated fixed migration" duration=445.416µs grafana | logger=migrator t=2024-04-24T09:37:41.664483304Z level=info msg="Executing migration" id="managed folder permissions library panel actions migration" grafana | logger=migrator t=2024-04-24T09:37:41.665052094Z level=info msg="Migration successfully executed" id="managed folder permissions library panel actions migration" duration=567.87µs grafana | logger=migrator t=2024-04-24T09:37:41.670154218Z level=info msg="Executing migration" id="migrate external alertmanagers to datsourcse" grafana | logger=migrator t=2024-04-24T09:37:41.670561513Z level=info msg="Migration successfully executed" id="migrate external alertmanagers to datsourcse" duration=403.594µs grafana | logger=migrator t=2024-04-24T09:37:41.674515144Z level=info msg="Executing migration" id="create folder table" grafana | logger=migrator t=2024-04-24T09:37:41.675572443Z level=info msg="Migration successfully executed" id="create folder table" duration=1.058909ms grafana | logger=migrator t=2024-04-24T09:37:41.679964001Z level=info msg="Executing migration" id="Add index for parent_uid" grafana | logger=migrator t=2024-04-24T09:37:41.681218046Z level=info msg="Migration successfully executed" id="Add index for parent_uid" duration=1.253225ms grafana | logger=migrator t=2024-04-24T09:37:41.686298269Z level=info msg="Executing migration" id="Add unique index for folder.uid and folder.org_id" grafana | logger=migrator t=2024-04-24T09:37:41.687556834Z level=info msg="Migration successfully executed" id="Add unique index for folder.uid and folder.org_id" duration=1.258755ms grafana | logger=migrator t=2024-04-24T09:37:41.692583085Z level=info msg="Executing migration" id="Update folder title length" grafana | logger=migrator t=2024-04-24T09:37:41.692616147Z level=info msg="Migration successfully executed" id="Update folder title length" duration=32.001µs grafana | logger=migrator t=2024-04-24T09:37:41.695714128Z level=info msg="Executing migration" id="Add unique index for folder.title and folder.parent_uid" grafana | logger=migrator t=2024-04-24T09:37:41.696981313Z level=info msg="Migration successfully executed" id="Add unique index for folder.title and folder.parent_uid" duration=1.267235ms grafana | logger=migrator t=2024-04-24T09:37:41.701826228Z level=info msg="Executing migration" id="Remove unique index for folder.title and folder.parent_uid" grafana | logger=migrator t=2024-04-24T09:37:41.703791289Z level=info msg="Migration successfully executed" id="Remove unique index for folder.title and folder.parent_uid" duration=1.965441ms grafana | logger=migrator t=2024-04-24T09:37:41.707743951Z level=info msg="Executing migration" id="Add unique index for title, parent_uid, and org_id" grafana | logger=migrator t=2024-04-24T09:37:41.708953175Z level=info msg="Migration successfully executed" id="Add unique index for title, parent_uid, and org_id" duration=1.208894ms grafana | logger=migrator t=2024-04-24T09:37:41.712977589Z level=info msg="Executing migration" id="Sync dashboard and folder table" grafana | logger=migrator t=2024-04-24T09:37:41.713503748Z level=info msg="Migration successfully executed" id="Sync dashboard and folder table" duration=525.859µs grafana | logger=migrator t=2024-04-24T09:37:41.718155766Z level=info msg="Executing migration" id="Remove ghost folders from the folder table" grafana | logger=migrator t=2024-04-24T09:37:41.718526179Z level=info msg="Migration successfully executed" id="Remove ghost folders from the folder table" duration=370.753µs grafana | logger=migrator t=2024-04-24T09:37:41.722364187Z level=info msg="Executing migration" id="Remove unique index UQE_folder_uid_org_id" grafana | logger=migrator t=2024-04-24T09:37:41.723496078Z level=info msg="Migration successfully executed" id="Remove unique index UQE_folder_uid_org_id" duration=1.132131ms grafana | logger=migrator t=2024-04-24T09:37:41.727518703Z level=info msg="Executing migration" id="Add unique index UQE_folder_org_id_uid" grafana | logger=migrator t=2024-04-24T09:37:41.729553806Z level=info msg="Migration successfully executed" id="Add unique index UQE_folder_org_id_uid" duration=2.034804ms grafana | logger=migrator t=2024-04-24T09:37:41.734754934Z level=info msg="Executing migration" id="Remove unique index UQE_folder_title_parent_uid_org_id" grafana | logger=migrator t=2024-04-24T09:37:41.735903024Z level=info msg="Migration successfully executed" id="Remove unique index UQE_folder_title_parent_uid_org_id" duration=1.147341ms grafana | logger=migrator t=2024-04-24T09:37:41.739563817Z level=info msg="Executing migration" id="Add unique index UQE_folder_org_id_parent_uid_title" grafana | logger=migrator t=2024-04-24T09:37:41.740821292Z level=info msg="Migration successfully executed" id="Add unique index UQE_folder_org_id_parent_uid_title" duration=1.256865ms grafana | logger=migrator t=2024-04-24T09:37:41.744410291Z level=info msg="Executing migration" id="Remove index IDX_folder_parent_uid_org_id" grafana | logger=migrator t=2024-04-24T09:37:41.745543212Z level=info msg="Migration successfully executed" id="Remove index IDX_folder_parent_uid_org_id" duration=1.130321ms grafana | logger=migrator t=2024-04-24T09:37:41.750256271Z level=info msg="Executing migration" id="create anon_device table" grafana | logger=migrator t=2024-04-24T09:37:41.751698573Z level=info msg="Migration successfully executed" id="create anon_device table" duration=1.438482ms grafana | logger=migrator t=2024-04-24T09:37:41.755852552Z level=info msg="Executing migration" id="add unique index anon_device.device_id" grafana | logger=migrator t=2024-04-24T09:37:41.757905677Z level=info msg="Migration successfully executed" id="add unique index anon_device.device_id" duration=2.053125ms grafana | logger=migrator t=2024-04-24T09:37:41.76270975Z level=info msg="Executing migration" id="add index anon_device.updated_at" grafana | logger=migrator t=2024-04-24T09:37:41.763960294Z level=info msg="Migration successfully executed" id="add index anon_device.updated_at" duration=1.250004ms grafana | logger=migrator t=2024-04-24T09:37:41.767925418Z level=info msg="Executing migration" id="create signing_key table" grafana | logger=migrator t=2024-04-24T09:37:41.76965988Z level=info msg="Migration successfully executed" id="create signing_key table" duration=1.733191ms grafana | logger=migrator t=2024-04-24T09:37:41.77383423Z level=info msg="Executing migration" id="add unique index signing_key.key_id" grafana | logger=migrator t=2024-04-24T09:37:41.775887525Z level=info msg="Migration successfully executed" id="add unique index signing_key.key_id" duration=2.053224ms grafana | logger=migrator t=2024-04-24T09:37:41.78078882Z level=info msg="Executing migration" id="set legacy alert migration status in kvstore" grafana | logger=migrator t=2024-04-24T09:37:41.782372588Z level=info msg="Migration successfully executed" id="set legacy alert migration status in kvstore" duration=1.580518ms grafana | logger=migrator t=2024-04-24T09:37:41.916949872Z level=info msg="Executing migration" id="migrate record of created folders during legacy migration to kvstore" grafana | logger=migrator t=2024-04-24T09:37:41.917741001Z level=info msg="Migration successfully executed" id="migrate record of created folders during legacy migration to kvstore" duration=799.468µs grafana | logger=migrator t=2024-04-24T09:37:41.923056392Z level=info msg="Executing migration" id="Add folder_uid for dashboard" grafana | logger=migrator t=2024-04-24T09:37:41.930351985Z level=info msg="Migration successfully executed" id="Add folder_uid for dashboard" duration=7.295193ms grafana | logger=migrator t=2024-04-24T09:37:41.934153352Z level=info msg="Executing migration" id="Populate dashboard folder_uid column" grafana | logger=migrator t=2024-04-24T09:37:41.935484269Z level=info msg="Migration successfully executed" id="Populate dashboard folder_uid column" duration=1.339048ms grafana | logger=migrator t=2024-04-24T09:37:41.939728832Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_folder_uid_title" grafana | logger=migrator t=2024-04-24T09:37:41.941712234Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_folder_uid_title" duration=1.981372ms grafana | logger=migrator t=2024-04-24T09:37:41.947381388Z level=info msg="Executing migration" id="Delete unique index for dashboard_org_id_folder_id_title" grafana | logger=migrator t=2024-04-24T09:37:41.94826648Z level=info msg="Migration successfully executed" id="Delete unique index for dashboard_org_id_folder_id_title" duration=882.801µs grafana | logger=migrator t=2024-04-24T09:37:41.951417613Z level=info msg="Executing migration" id="Delete unique index for dashboard_org_id_folder_uid_title" grafana | logger=migrator t=2024-04-24T09:37:41.952283714Z level=info msg="Migration successfully executed" id="Delete unique index for dashboard_org_id_folder_uid_title" duration=866.081µs grafana | logger=migrator t=2024-04-24T09:37:41.955546182Z level=info msg="Executing migration" id="Add unique index for dashboard_org_id_folder_uid_title_is_folder" grafana | logger=migrator t=2024-04-24T09:37:41.956496766Z level=info msg="Migration successfully executed" id="Add unique index for dashboard_org_id_folder_uid_title_is_folder" duration=950.094µs grafana | logger=migrator t=2024-04-24T09:37:41.961409563Z level=info msg="Executing migration" id="Restore index for dashboard_org_id_folder_id_title" grafana | logger=migrator t=2024-04-24T09:37:41.962267894Z level=info msg="Migration successfully executed" id="Restore index for dashboard_org_id_folder_id_title" duration=858.971µs grafana | logger=migrator t=2024-04-24T09:37:41.966655582Z level=info msg="Executing migration" id="create sso_setting table" grafana | logger=migrator t=2024-04-24T09:37:41.968331972Z level=info msg="Migration successfully executed" id="create sso_setting table" duration=1.67585ms grafana | logger=migrator t=2024-04-24T09:37:41.974454612Z level=info msg="Executing migration" id="copy kvstore migration status to each org" grafana | logger=migrator t=2024-04-24T09:37:41.975280542Z level=info msg="Migration successfully executed" id="copy kvstore migration status to each org" duration=828.11µs grafana | logger=migrator t=2024-04-24T09:37:41.980472609Z level=info msg="Executing migration" id="add back entry for orgid=0 migrated status" grafana | logger=migrator t=2024-04-24T09:37:41.980744539Z level=info msg="Migration successfully executed" id="add back entry for orgid=0 migrated status" duration=272.07µs grafana | logger=migrator t=2024-04-24T09:37:41.984480353Z level=info msg="Executing migration" id="alter kv_store.value to longtext" grafana | logger=migrator t=2024-04-24T09:37:41.984582327Z level=info msg="Migration successfully executed" id="alter kv_store.value to longtext" duration=102.534µs grafana | logger=migrator t=2024-04-24T09:37:41.988515358Z level=info msg="Executing migration" id="add notification_settings column to alert_rule table" grafana | logger=migrator t=2024-04-24T09:37:41.999779384Z level=info msg="Migration successfully executed" id="add notification_settings column to alert_rule table" duration=11.262705ms grafana | logger=migrator t=2024-04-24T09:37:42.004932605Z level=info msg="Executing migration" id="add notification_settings column to alert_rule_version table" grafana | logger=migrator t=2024-04-24T09:37:42.011536996Z level=info msg="Migration successfully executed" id="add notification_settings column to alert_rule_version table" duration=6.603972ms grafana | logger=migrator t=2024-04-24T09:37:42.015031716Z level=info msg="Executing migration" id="removing scope from alert.instances:read action migration" grafana | logger=migrator t=2024-04-24T09:37:42.015339558Z level=info msg="Migration successfully executed" id="removing scope from alert.instances:read action migration" duration=303.021µs grafana | logger=migrator t=2024-04-24T09:37:42.018752034Z level=info msg="migrations completed" performed=548 skipped=0 duration=4.723919646s grafana | logger=sqlstore t=2024-04-24T09:37:42.028006024Z level=info msg="Created default admin" user=admin grafana | logger=sqlstore t=2024-04-24T09:37:42.028299066Z level=info msg="Created default organization" grafana | logger=secrets t=2024-04-24T09:37:42.03392082Z level=info msg="Envelope encryption state" enabled=true currentprovider=secretKey.v1 grafana | logger=plugin.store t=2024-04-24T09:37:42.056082675Z level=info msg="Loading plugins..." kafka | [2024-04-24 09:38:19,301] INFO [LogLoader partition=__consumer_offsets-27, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,302] INFO Created log for partition __consumer_offsets-27 in /var/lib/kafka/data/__consumer_offsets-27 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,302] INFO [Partition __consumer_offsets-27 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-27 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,302] INFO [Partition __consumer_offsets-27 broker=1] Log loaded for partition __consumer_offsets-27 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,303] INFO [Broker id=1] Leader __consumer_offsets-27 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,312] INFO [LogLoader partition=__consumer_offsets-42, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,314] INFO Created log for partition __consumer_offsets-42 in /var/lib/kafka/data/__consumer_offsets-42 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,314] INFO [Partition __consumer_offsets-42 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-42 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,314] INFO [Partition __consumer_offsets-42 broker=1] Log loaded for partition __consumer_offsets-42 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,315] INFO [Broker id=1] Leader __consumer_offsets-42 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,324] INFO [LogLoader partition=__consumer_offsets-12, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,325] INFO Created log for partition __consumer_offsets-12 in /var/lib/kafka/data/__consumer_offsets-12 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,325] INFO [Partition __consumer_offsets-12 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-12 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,325] INFO [Partition __consumer_offsets-12 broker=1] Log loaded for partition __consumer_offsets-12 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,325] INFO [Broker id=1] Leader __consumer_offsets-12 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,337] INFO [LogLoader partition=__consumer_offsets-21, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,338] INFO Created log for partition __consumer_offsets-21 in /var/lib/kafka/data/__consumer_offsets-21 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,339] INFO [Partition __consumer_offsets-21 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-21 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,339] INFO [Partition __consumer_offsets-21 broker=1] Log loaded for partition __consumer_offsets-21 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,339] INFO [Broker id=1] Leader __consumer_offsets-21 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,346] INFO [LogLoader partition=__consumer_offsets-36, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,348] INFO Created log for partition __consumer_offsets-36 in /var/lib/kafka/data/__consumer_offsets-36 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,348] INFO [Partition __consumer_offsets-36 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-36 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,348] INFO [Partition __consumer_offsets-36 broker=1] Log loaded for partition __consumer_offsets-36 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,349] INFO [Broker id=1] Leader __consumer_offsets-36 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,356] INFO [LogLoader partition=__consumer_offsets-6, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,356] INFO Created log for partition __consumer_offsets-6 in /var/lib/kafka/data/__consumer_offsets-6 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,357] INFO [Partition __consumer_offsets-6 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-6 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,357] INFO [Partition __consumer_offsets-6 broker=1] Log loaded for partition __consumer_offsets-6 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,357] INFO [Broker id=1] Leader __consumer_offsets-6 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,364] INFO [LogLoader partition=__consumer_offsets-43, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,365] INFO Created log for partition __consumer_offsets-43 in /var/lib/kafka/data/__consumer_offsets-43 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,365] INFO [Partition __consumer_offsets-43 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-43 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,365] INFO [Partition __consumer_offsets-43 broker=1] Log loaded for partition __consumer_offsets-43 with initial high watermark 0 (kafka.cluster.Partition) grafana | logger=local.finder t=2024-04-24T09:37:42.104927045Z level=warn msg="Skipping finding plugins as directory does not exist" path=/usr/share/grafana/plugins-bundled grafana | logger=plugin.store t=2024-04-24T09:37:42.104980157Z level=info msg="Plugins loaded" count=55 duration=48.897632ms grafana | logger=query_data t=2024-04-24T09:37:42.111865862Z level=info msg="Query Service initialization" grafana | logger=live.push_http t=2024-04-24T09:37:42.116720996Z level=info msg="Live Push Gateway initialization" grafana | logger=ngalert.migration t=2024-04-24T09:37:42.125026998Z level=info msg=Starting grafana | logger=ngalert.migration t=2024-04-24T09:37:42.12558836Z level=info msg="Applying transition" currentType=Legacy desiredType=UnifiedAlerting cleanOnDowngrade=false cleanOnUpgrade=false grafana | logger=ngalert.migration orgID=1 t=2024-04-24T09:37:42.126092601Z level=info msg="Migrating alerts for organisation" grafana | logger=ngalert.migration orgID=1 t=2024-04-24T09:37:42.126859141Z level=info msg="Alerts found to migrate" alerts=0 grafana | logger=ngalert.migration t=2024-04-24T09:37:42.128961275Z level=info msg="Completed alerting migration" grafana | logger=ngalert.state.manager t=2024-04-24T09:37:42.1588934Z level=info msg="Running in alternative execution of Error/NoData mode" grafana | logger=infra.usagestats.collector t=2024-04-24T09:37:42.161223283Z level=info msg="registering usage stat providers" usageStatsProvidersLen=2 grafana | logger=provisioning.datasources t=2024-04-24T09:37:42.163405231Z level=info msg="inserting datasource from configuration" name=PolicyPrometheus uid=dkSf71fnz grafana | logger=provisioning.alerting t=2024-04-24T09:37:42.177367738Z level=info msg="starting to provision alerting" grafana | logger=provisioning.alerting t=2024-04-24T09:37:42.177394359Z level=info msg="finished to provision alerting" grafana | logger=grafanaStorageLogger t=2024-04-24T09:37:42.178262614Z level=info msg="Storage starting" grafana | logger=http.server t=2024-04-24T09:37:42.181477742Z level=info msg="HTTP Server Listen" address=[::]:3000 protocol=http subUrl= socket= grafana | logger=ngalert.state.manager t=2024-04-24T09:37:42.181590416Z level=info msg="Warming state cache for startup" grafana | logger=ngalert.multiorg.alertmanager t=2024-04-24T09:37:42.187530693Z level=info msg="Starting MultiOrg Alertmanager" grafana | logger=ngalert.state.manager t=2024-04-24T09:37:42.210765602Z level=info msg="State cache has been initialized" states=0 duration=29.169926ms grafana | logger=ngalert.scheduler t=2024-04-24T09:37:42.210847475Z level=info msg="Starting scheduler" tickInterval=10s maxAttempts=1 grafana | logger=ticker t=2024-04-24T09:37:42.210944369Z level=info msg=starting first_tick=2024-04-24T09:37:50Z grafana | logger=provisioning.dashboard t=2024-04-24T09:37:42.270263708Z level=info msg="starting to provision dashboards" grafana | logger=plugins.update.checker t=2024-04-24T09:37:42.274477996Z level=info msg="Update check succeeded" duration=96.113709ms grafana | logger=grafana.update.checker t=2024-04-24T09:37:42.288977515Z level=info msg="Update check succeeded" duration=111.395318ms grafana | logger=sqlstore.transactions t=2024-04-24T09:37:42.297245505Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=0 code="database is locked" grafana | logger=sqlstore.transactions t=2024-04-24T09:37:42.308056866Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=1 code="database is locked" grafana | logger=sqlstore.transactions t=2024-04-24T09:37:42.310238944Z level=info msg="Database locked, sleeping then retrying" error="database is locked" retry=0 code="database is locked" grafana | logger=provisioning.dashboard t=2024-04-24T09:37:42.768083917Z level=info msg="finished to provision dashboards" grafana | logger=grafana-apiserver t=2024-04-24T09:37:42.801397028Z level=info msg="Adding GroupVersion playlist.grafana.app v0alpha1 to ResourceManager" grafana | logger=grafana-apiserver t=2024-04-24T09:37:42.802053394Z level=info msg="Adding GroupVersion featuretoggle.grafana.app v0alpha1 to ResourceManager" grafana | logger=infra.usagestats t=2024-04-24T09:39:37.196894549Z level=info msg="Usage stats are ready to report" kafka | [2024-04-24 09:38:19,365] INFO [Broker id=1] Leader __consumer_offsets-43 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,376] INFO [LogLoader partition=__consumer_offsets-13, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,376] INFO Created log for partition __consumer_offsets-13 in /var/lib/kafka/data/__consumer_offsets-13 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,376] INFO [Partition __consumer_offsets-13 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-13 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,376] INFO [Partition __consumer_offsets-13 broker=1] Log loaded for partition __consumer_offsets-13 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,376] INFO [Broker id=1] Leader __consumer_offsets-13 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,386] INFO [LogLoader partition=__consumer_offsets-28, dir=/var/lib/kafka/data] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) kafka | [2024-04-24 09:38:19,386] INFO Created log for partition __consumer_offsets-28 in /var/lib/kafka/data/__consumer_offsets-28 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) kafka | [2024-04-24 09:38:19,386] INFO [Partition __consumer_offsets-28 broker=1] No checkpointed highwatermark is found for partition __consumer_offsets-28 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,386] INFO [Partition __consumer_offsets-28 broker=1] Log loaded for partition __consumer_offsets-28 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2024-04-24 09:38:19,386] INFO [Broker id=1] Leader __consumer_offsets-28 with topic id Some(kkBxd83QQh-C7hymHduj-A) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [1], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-3 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-18 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-41 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-10 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-33 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-48 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-19 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-34 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-4 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-11 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-26 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-49 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-39 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-9 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-24 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-31 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-46 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-1 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-16 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-2 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-25 (state.change.logger) kafka | [2024-04-24 09:38:19,391] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-40 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-47 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-17 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-32 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-37 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-7 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-22 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-29 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-44 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-14 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-23 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-38 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-8 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition policy-pdp-pap-0 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-45 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-15 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-30 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-0 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-35 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-5 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-20 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-27 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-42 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-12 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-21 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-36 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-6 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-43 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-13 (state.change.logger) kafka | [2024-04-24 09:38:19,392] TRACE [Broker id=1] Completed LeaderAndIsr request correlationId 1 from controller 1 epoch 1 for the become-leader transition for partition __consumer_offsets-28 (state.change.logger) kafka | [2024-04-24 09:38:19,400] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 3 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,411] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-3 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,415] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 18 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,415] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-18 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,415] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 41 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,415] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-41 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,415] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 10 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,415] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-10 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,416] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 33 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,416] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-33 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,416] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 48 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,416] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-48 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,416] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 19 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,416] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-19 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,416] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 34 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,416] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-34 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,417] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 4 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,417] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-4 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,417] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 11 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,417] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-11 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,417] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 26 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,417] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-26 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,417] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 49 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,417] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-49 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,418] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 39 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,418] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-39 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,418] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 9 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,418] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-9 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,418] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 24 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,418] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-24 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,418] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 31 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,418] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-31 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,419] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 46 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,419] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-46 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,419] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 1 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,419] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-1 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,419] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 16 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,419] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-16 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,419] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 2 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,419] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-2 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,419] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 25 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,420] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-25 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,420] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 40 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,420] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-40 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,420] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 47 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,420] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-47 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,422] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 17 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,422] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-17 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,422] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 32 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,422] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-32 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,422] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 37 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,422] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-37 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,422] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 7 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,422] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-7 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,422] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 22 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,422] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-22 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,422] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 29 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,423] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-29 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,420] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-3 in 8 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,425] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 44 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,425] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-44 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,425] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-18 in 10 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,425] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 14 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,425] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-14 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,426] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 23 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,426] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-23 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,426] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 38 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,426] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-38 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,426] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 8 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,426] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-8 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,426] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 45 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,426] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-45 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,426] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 15 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,426] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-15 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,426] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 30 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,426] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-30 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,426] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 0 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,427] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-0 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,427] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 35 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,427] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-35 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,427] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 5 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,427] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-5 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,427] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 20 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,426] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-41 in 11 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,427] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-20 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,427] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-10 in 12 milliseconds for epoch 0, of which 12 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,427] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-33 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,427] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 27 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,427] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-48 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,427] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-27 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,427] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-19 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 42 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-34 in 12 milliseconds for epoch 0, of which 12 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-42 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-4 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-11 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 12 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-26 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-12 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-49 in 10 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 21 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-39 in 10 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-21 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-9 in 10 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,428] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 36 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,429] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-36 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,429] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 6 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,429] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-6 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,429] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-24 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,429] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 43 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,429] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-43 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,429] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 13 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,429] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-31 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,429] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-13 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,429] INFO [GroupCoordinator 1]: Elected as the group coordinator for partition 28 in epoch 0 (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,429] INFO [GroupMetadataManager brokerId=1] Scheduling loading of offsets and group metadata from __consumer_offsets-28 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,429] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-46 in 10 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,430] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-1 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,430] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-16 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,430] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-2 in 11 milliseconds for epoch 0, of which 11 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,430] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-25 in 10 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,430] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-40 in 10 milliseconds for epoch 0, of which 10 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,430] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-47 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,431] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-17 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,431] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-32 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,431] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-37 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,431] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-7 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,431] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-22 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,431] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-29 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,431] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-44 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,432] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-14 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,432] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-23 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,432] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-38 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,432] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-8 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,432] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-45 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,433] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-15 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,433] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-30 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,433] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-0 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,433] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-35 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,433] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-5 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,433] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-20 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,433] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-27 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,434] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-42 in 6 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,434] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-12 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,434] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-21 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,434] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-36 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,434] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-6 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,434] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-43 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,435] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-13 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,435] INFO [GroupMetadataManager brokerId=1] Finished loading offsets and group metadata from __consumer_offsets-28 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) kafka | [2024-04-24 09:38:19,437] INFO [Broker id=1] Finished LeaderAndIsr request in 781ms correlationId 1 from controller 1 for 51 partitions (state.change.logger) kafka | [2024-04-24 09:38:19,443] TRACE [Controller id=1 epoch=1] Received response LeaderAndIsrResponseData(errorCode=0, partitionErrors=[], topics=[LeaderAndIsrTopicError(topicId=kkBxd83QQh-C7hymHduj-A, partitionErrors=[LeaderAndIsrPartitionError(topicName='', partitionIndex=13, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=46, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=9, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=42, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=21, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=17, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=30, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=26, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=5, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=38, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=1, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=34, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=16, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=45, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=12, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=41, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=24, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=20, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=49, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=0, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=29, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=25, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=8, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=37, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=4, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=33, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=15, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=48, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=11, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=44, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=23, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=19, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=32, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=28, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=7, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=40, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=3, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=36, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=47, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=14, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=43, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=10, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=22, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=18, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=31, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=27, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=39, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=6, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=35, errorCode=0), LeaderAndIsrPartitionError(topicName='', partitionIndex=2, errorCode=0)]), LeaderAndIsrTopicError(topicId=lm-F3iiTSNeAYdJn1IimsA, partitionErrors=[LeaderAndIsrPartitionError(topicName='', partitionIndex=0, errorCode=0)])]) for request LEADER_AND_ISR with correlation id 1 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) kafka | [2024-04-24 09:38:19,453] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='policy-pdp-pap', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition policy-pdp-pap-0 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=13, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-13 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=46, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-46 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=9, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-9 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=42, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-42 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=21, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-21 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=17, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-17 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=30, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-30 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=26, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-26 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=5, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-5 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=38, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-38 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=1, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-1 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=34, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-34 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=16, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-16 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=45, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-45 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,454] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=12, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-12 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=41, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-41 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=24, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-24 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=20, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-20 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=49, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-49 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=0, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-0 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=29, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-29 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=25, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-25 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=8, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-8 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=37, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-37 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=4, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-4 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=33, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-33 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=15, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-15 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=48, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-48 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=11, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-11 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=44, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-44 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=23, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-23 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,455] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=19, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-19 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=32, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-32 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=28, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-28 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=7, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-7 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=40, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-40 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=3, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-3 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=36, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-36 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=47, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-47 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=14, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-14 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=43, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-43 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=10, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-10 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=22, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-22 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=18, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-18 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=31, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-31 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=27, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-27 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=39, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-39 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=6, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-6 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,456] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=35, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-35 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,457] TRACE [Broker id=1] Cached leader info UpdateMetadataPartitionState(topicName='__consumer_offsets', partitionIndex=2, controllerEpoch=1, leader=1, leaderEpoch=0, isr=[1], zkVersion=0, replicas=[1], offlineReplicas=[]) for partition __consumer_offsets-2 in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,457] INFO [Broker id=1] Add 51 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 1 epoch 1 with correlation id 2 (state.change.logger) kafka | [2024-04-24 09:38:19,458] TRACE [Controller id=1 epoch=1] Received response UpdateMetadataResponseData(errorCode=0) for request UPDATE_METADATA with correlation id 2 sent to broker kafka:9092 (id: 1 rack: null) (state.change.logger) kafka | [2024-04-24 09:38:19,554] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group policy-pap in Empty state. Created a new member id consumer-policy-pap-4-5818d6ba-9a29-412c-a74d-37f074bc3da4 and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,554] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group 0d71b024-87db-4bab-bad3-947ec269530c in Empty state. Created a new member id consumer-0d71b024-87db-4bab-bad3-947ec269530c-3-d1876d69-1880-43fc-9d10-af88199c63eb and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,571] INFO [GroupCoordinator 1]: Preparing to rebalance group policy-pap in state PreparingRebalance with old generation 0 (__consumer_offsets-24) (reason: Adding new member consumer-policy-pap-4-5818d6ba-9a29-412c-a74d-37f074bc3da4 with group instance id None; client reason: rebalance failed due to MemberIdRequiredException) (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:19,572] INFO [GroupCoordinator 1]: Preparing to rebalance group 0d71b024-87db-4bab-bad3-947ec269530c in state PreparingRebalance with old generation 0 (__consumer_offsets-32) (reason: Adding new member consumer-0d71b024-87db-4bab-bad3-947ec269530c-3-d1876d69-1880-43fc-9d10-af88199c63eb with group instance id None; client reason: rebalance failed due to MemberIdRequiredException) (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:20,270] INFO [GroupCoordinator 1]: Dynamic member with unknown member id joins group ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0 in Empty state. Created a new member id consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2-dd69be57-ac85-4787-b12c-453fd9504f7c and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:20,275] INFO [GroupCoordinator 1]: Preparing to rebalance group ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0 in state PreparingRebalance with old generation 0 (__consumer_offsets-6) (reason: Adding new member consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2-dd69be57-ac85-4787-b12c-453fd9504f7c with group instance id None; client reason: rebalance failed due to MemberIdRequiredException) (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:22,583] INFO [GroupCoordinator 1]: Stabilized group policy-pap generation 1 (__consumer_offsets-24) with 1 members (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:22,590] INFO [GroupCoordinator 1]: Stabilized group 0d71b024-87db-4bab-bad3-947ec269530c generation 1 (__consumer_offsets-32) with 1 members (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:22,614] INFO [GroupCoordinator 1]: Assignment received from leader consumer-0d71b024-87db-4bab-bad3-947ec269530c-3-d1876d69-1880-43fc-9d10-af88199c63eb for group 0d71b024-87db-4bab-bad3-947ec269530c for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:22,614] INFO [GroupCoordinator 1]: Assignment received from leader consumer-policy-pap-4-5818d6ba-9a29-412c-a74d-37f074bc3da4 for group policy-pap for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:23,278] INFO [GroupCoordinator 1]: Stabilized group ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0 generation 1 (__consumer_offsets-6) with 1 members (kafka.coordinator.group.GroupCoordinator) kafka | [2024-04-24 09:38:23,293] INFO [GroupCoordinator 1]: Assignment received from leader consumer-ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0-2-dd69be57-ac85-4787-b12c-453fd9504f7c for group ea7e35d4-d1bc-4f0d-bd5b-308fdbd281e0 for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) ++ echo 'Tearing down containers...' Tearing down containers... ++ docker-compose down -v --remove-orphans Stopping policy-apex-pdp ... Stopping policy-pap ... Stopping kafka ... Stopping policy-api ... Stopping grafana ... Stopping simulator ... Stopping zookeeper ... Stopping mariadb ... Stopping prometheus ... Stopping grafana ... done Stopping prometheus ... done Stopping policy-apex-pdp ... done Stopping simulator ... done Stopping policy-pap ... done Stopping mariadb ... done Stopping kafka ... done Stopping zookeeper ... done Stopping policy-api ... done Removing policy-apex-pdp ... Removing policy-pap ... Removing kafka ... Removing policy-api ... Removing policy-db-migrator ... Removing grafana ... Removing simulator ... Removing zookeeper ... Removing mariadb ... Removing prometheus ... Removing policy-api ... done Removing grafana ... done Removing simulator ... done Removing policy-apex-pdp ... done Removing policy-db-migrator ... done Removing mariadb ... done Removing policy-pap ... done Removing prometheus ... done Removing kafka ... done Removing zookeeper ... done Removing network compose_default ++ cd /w/workspace/policy-pap-master-project-csit-pap + load_set + _setopts=hxB ++ echo braceexpand:hashall:interactive-comments:xtrace ++ tr : ' ' + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o braceexpand + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o hashall + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o interactive-comments + for i in $(echo "${SHELLOPTS}" | tr ':' ' ') + set +o xtrace ++ echo hxB ++ sed 's/./& /g' + for i in $(echo "$_setopts" | sed 's/./& /g') + set +h + for i in $(echo "$_setopts" | sed 's/./& /g') + set +x + rsync /w/workspace/policy-pap-master-project-csit-pap/compose/docker_compose.log /w/workspace/policy-pap-master-project-csit-pap/csit/archives/pap + [[ -n /tmp/tmp.UcCaUoaeAm ]] + rsync -av /tmp/tmp.UcCaUoaeAm/ /w/workspace/policy-pap-master-project-csit-pap/csit/archives/pap sending incremental file list ./ log.html output.xml report.html testplan.txt sent 918,977 bytes received 95 bytes 1,838,144.00 bytes/sec total size is 918,431 speedup is 1.00 + rm -rf /w/workspace/policy-pap-master-project-csit-pap/models + exit 0 $ ssh-agent -k unset SSH_AUTH_SOCK; unset SSH_AGENT_PID; echo Agent pid 2159 killed; [ssh-agent] Stopped. Robot results publisher started... INFO: Checking test criticality is deprecated and will be dropped in a future release! -Parsing output xml: Done! WARNING! Could not find file: **/log.html WARNING! Could not find file: **/report.html -Copying log files to build dir: Done! -Assigning results to build: Done! -Checking thresholds: Done! Done publishing Robot results. [PostBuildScript] - [INFO] Executing post build scripts. [policy-pap-master-project-csit-pap] $ /bin/bash /tmp/jenkins9314019032778867199.sh ---> sysstat.sh [policy-pap-master-project-csit-pap] $ /bin/bash /tmp/jenkins18294002054465843378.sh ---> package-listing.sh ++ tr '[:upper:]' '[:lower:]' ++ facter osfamily + OS_FAMILY=debian + workspace=/w/workspace/policy-pap-master-project-csit-pap + START_PACKAGES=/tmp/packages_start.txt + END_PACKAGES=/tmp/packages_end.txt + DIFF_PACKAGES=/tmp/packages_diff.txt + PACKAGES=/tmp/packages_start.txt + '[' /w/workspace/policy-pap-master-project-csit-pap ']' + PACKAGES=/tmp/packages_end.txt + case "${OS_FAMILY}" in + dpkg -l + grep '^ii' + '[' -f /tmp/packages_start.txt ']' + '[' -f /tmp/packages_end.txt ']' + diff /tmp/packages_start.txt /tmp/packages_end.txt + '[' /w/workspace/policy-pap-master-project-csit-pap ']' + mkdir -p /w/workspace/policy-pap-master-project-csit-pap/archives/ + cp -f /tmp/packages_diff.txt /tmp/packages_end.txt /tmp/packages_start.txt /w/workspace/policy-pap-master-project-csit-pap/archives/ [policy-pap-master-project-csit-pap] $ /bin/bash /tmp/jenkins2875920612796038602.sh ---> capture-instance-metadata.sh Setup pyenv: system 3.8.13 3.9.13 * 3.10.6 (set by /w/workspace/policy-pap-master-project-csit-pap/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-8CY4 from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-8CY4/bin to PATH INFO: Running in OpenStack, capturing instance metadata [policy-pap-master-project-csit-pap] $ /bin/bash /tmp/jenkins17120530160367500595.sh provisioning config files... copy managed file [jenkins-log-archives-settings] to file:/w/workspace/policy-pap-master-project-csit-pap@tmp/config1344521399750737532tmp Regular expression run condition: Expression=[^.*logs-s3.*], Label=[] Run condition [Regular expression match] preventing perform for step [Provide Configuration files] [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content SERVER_ID=logs [EnvInject] - Variables injected successfully. [policy-pap-master-project-csit-pap] $ /bin/bash /tmp/jenkins11124234703522775343.sh ---> create-netrc.sh [policy-pap-master-project-csit-pap] $ /bin/bash /tmp/jenkins11111769779688532314.sh ---> python-tools-install.sh Setup pyenv: system 3.8.13 3.9.13 * 3.10.6 (set by /w/workspace/policy-pap-master-project-csit-pap/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-8CY4 from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-8CY4/bin to PATH [policy-pap-master-project-csit-pap] $ /bin/bash /tmp/jenkins16096346737430137521.sh ---> sudo-logs.sh Archiving 'sudo' log.. [policy-pap-master-project-csit-pap] $ /bin/bash /tmp/jenkins13116475427522314579.sh ---> job-cost.sh Setup pyenv: system 3.8.13 3.9.13 * 3.10.6 (set by /w/workspace/policy-pap-master-project-csit-pap/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-8CY4 from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: zipp==1.1.0 python-openstackclient urllib3~=1.26.15 lf-activate-venv(): INFO: Adding /tmp/venv-8CY4/bin to PATH INFO: No Stack... INFO: Retrieving Pricing Info for: v3-standard-8 INFO: Archiving Costs [policy-pap-master-project-csit-pap] $ /bin/bash -l /tmp/jenkins11737155645551054094.sh ---> logs-deploy.sh Setup pyenv: system 3.8.13 3.9.13 * 3.10.6 (set by /w/workspace/policy-pap-master-project-csit-pap/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-8CY4 from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-8CY4/bin to PATH INFO: Nexus URL https://nexus.onap.org path production/vex-yul-ecomp-jenkins-1/policy-pap-master-project-csit-pap/1658 INFO: archiving workspace using pattern(s): -p **/target/surefire-reports/*-output.txt Archives upload complete. INFO: archiving logs to Nexus ---> uname -a: Linux prd-ubuntu1804-docker-8c-8g-25507 4.15.0-192-generic #203-Ubuntu SMP Wed Aug 10 17:40:03 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux ---> lscpu: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 8 On-line CPU(s) list: 0-7 Thread(s) per core: 1 Core(s) per socket: 1 Socket(s): 8 NUMA node(s): 1 Vendor ID: AuthenticAMD CPU family: 23 Model: 49 Model name: AMD EPYC-Rome Processor Stepping: 0 CPU MHz: 2799.998 BogoMIPS: 5599.99 Virtualization: AMD-V Hypervisor vendor: KVM Virtualization type: full L1d cache: 32K L1i cache: 32K L2 cache: 512K L3 cache: 16384K NUMA node0 CPU(s): 0-7 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm rep_good nopl xtopology cpuid extd_apicid tsc_known_freq pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext perfctr_core ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves clzero xsaveerptr arat npt nrip_save umip rdpid arch_capabilities ---> nproc: 8 ---> df -h: Filesystem Size Used Avail Use% Mounted on udev 16G 0 16G 0% /dev tmpfs 3.2G 708K 3.2G 1% /run /dev/vda1 155G 14G 142G 9% / tmpfs 16G 0 16G 0% /dev/shm tmpfs 5.0M 0 5.0M 0% /run/lock tmpfs 16G 0 16G 0% /sys/fs/cgroup /dev/vda15 105M 4.4M 100M 5% /boot/efi tmpfs 3.2G 0 3.2G 0% /run/user/1001 ---> free -m: total used free shared buff/cache available Mem: 32167 848 25163 0 6155 30862 Swap: 1023 0 1023 ---> ip addr: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: ens3: mtu 1458 qdisc mq state UP group default qlen 1000 link/ether fa:16:3e:63:43:0c brd ff:ff:ff:ff:ff:ff inet 10.30.106.122/23 brd 10.30.107.255 scope global dynamic ens3 valid_lft 85847sec preferred_lft 85847sec inet6 fe80::f816:3eff:fe63:430c/64 scope link valid_lft forever preferred_lft forever 3: docker0: mtu 1500 qdisc noqueue state DOWN group default link/ether 02:42:38:51:99:13 brd ff:ff:ff:ff:ff:ff inet 10.250.0.254/24 brd 10.250.0.255 scope global docker0 valid_lft forever preferred_lft forever ---> sar -b -r -n DEV: Linux 4.15.0-192-generic (prd-ubuntu1804-docker-8c-8g-25507) 04/24/24 _x86_64_ (8 CPU) 09:32:24 LINUX RESTART (8 CPU) 09:33:01 tps rtps wtps bread/s bwrtn/s 09:34:01 87.29 53.86 33.43 3590.47 41929.28 09:35:02 86.79 17.56 69.23 1019.79 22339.75 09:36:01 112.71 14.08 98.63 1143.40 27971.53 09:37:01 120.58 9.65 110.93 1714.25 56527.38 09:38:01 433.38 11.76 421.61 779.67 112537.28 09:39:01 31.61 0.27 31.34 13.06 23204.18 09:40:01 14.56 0.00 14.56 0.00 18614.71 09:41:01 65.09 0.85 64.24 42.66 18850.56 Average: 119.01 13.50 105.51 1037.69 40272.08 09:33:01 kbmemfree kbavail kbmemused %memused kbbuffers kbcached kbcommit %commit kbactive kbinact kbdirty 09:34:01 30397416 31736864 2541804 7.72 45000 1618092 1421364 4.18 795628 1481992 55380 09:35:02 30151872 31725740 2787348 8.46 68996 1816136 1407216 4.14 845188 1652600 145332 09:36:01 29840288 31715876 3098932 9.41 85848 2085676 1410320 4.15 864428 1912012 123580 09:37:01 27215080 31662068 5724140 17.38 129312 4504256 1414284 4.16 1014324 4242104 1322608 09:38:01 24678888 30556540 8260332 25.08 155168 5835208 7552164 22.22 2261844 5406540 252 09:39:01 23696464 29686984 9242756 28.06 156872 5942956 8704140 25.61 3180496 5456868 204 09:40:01 23684064 29675468 9255156 28.10 157020 5943532 8736548 25.70 3193928 5457416 400 09:41:01 25732536 31565688 7206684 21.88 158480 5801252 1579276 4.65 1347888 5316876 9404 Average: 26924576 31040654 6014644 18.26 119587 4193388 4028164 11.85 1687966 3865801 207145 09:33:01 IFACE rxpck/s txpck/s rxkB/s txkB/s rxcmp/s txcmp/s rxmcst/s %ifutil 09:34:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:34:01 ens3 341.78 227.06 819.96 57.22 0.00 0.00 0.00 0.00 09:34:01 lo 0.67 0.67 0.07 0.07 0.00 0.00 0.00 0.00 09:35:02 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:35:02 ens3 44.12 31.52 713.68 7.02 0.00 0.00 0.00 0.00 09:35:02 lo 1.60 1.60 0.17 0.17 0.00 0.00 0.00 0.00 09:36:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:36:01 ens3 49.89 37.99 721.83 8.32 0.00 0.00 0.00 0.00 09:36:01 lo 1.08 1.08 0.12 0.12 0.00 0.00 0.00 0.00 09:37:01 br-a8723711ca17 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:37:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:37:01 ens3 628.21 325.25 17100.91 24.53 0.00 0.00 0.00 0.00 09:37:01 lo 8.67 8.67 0.84 0.84 0.00 0.00 0.00 0.00 09:38:01 br-a8723711ca17 0.72 0.70 0.05 0.31 0.00 0.00 0.00 0.00 09:38:01 veth8494eda 0.47 0.90 0.05 0.31 0.00 0.00 0.00 0.00 09:38:01 veth493b75e 1.32 1.80 0.15 0.18 0.00 0.00 0.00 0.00 09:38:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:39:01 br-a8723711ca17 1.77 2.03 1.74 1.69 0.00 0.00 0.00 0.00 09:39:01 veth8494eda 0.13 0.22 0.01 0.01 0.00 0.00 0.00 0.00 09:39:01 veth493b75e 4.23 4.97 0.70 0.77 0.00 0.00 0.00 0.00 09:39:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:40:01 br-a8723711ca17 0.82 0.83 0.10 0.08 0.00 0.00 0.00 0.00 09:40:01 veth8494eda 0.17 0.12 0.01 0.01 0.00 0.00 0.00 0.00 09:40:01 veth493b75e 0.17 0.35 0.01 0.02 0.00 0.00 0.00 0.00 09:40:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:41:01 docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 09:41:01 ens3 1543.51 893.40 33072.73 145.63 0.00 0.00 0.00 0.00 09:41:01 lo 35.66 35.66 6.27 6.27 0.00 0.00 0.00 0.00 Average: docker0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 Average: ens3 191.23 110.52 4133.91 18.05 0.00 0.00 0.00 0.00 Average: lo 4.04 4.04 0.75 0.75 0.00 0.00 0.00 0.00 ---> sar -P ALL: Linux 4.15.0-192-generic (prd-ubuntu1804-docker-8c-8g-25507) 04/24/24 _x86_64_ (8 CPU) 09:32:24 LINUX RESTART (8 CPU) 09:33:01 CPU %user %nice %system %iowait %steal %idle 09:34:01 all 4.93 0.00 0.61 16.39 0.03 78.04 09:34:01 0 4.63 0.00 0.80 0.28 0.03 94.25 09:34:01 1 3.45 0.00 0.38 0.18 0.03 95.95 09:34:01 2 6.04 0.00 0.72 1.64 0.02 91.59 09:34:01 3 4.39 0.00 0.55 46.25 0.02 48.78 09:34:01 4 6.40 0.00 0.42 72.34 0.03 20.81 09:34:01 5 5.75 0.00 0.85 6.65 0.02 86.72 09:34:01 6 4.23 0.00 0.59 1.77 0.02 93.40 09:34:01 7 4.50 0.00 0.64 2.07 0.05 92.74 09:35:02 all 7.83 0.00 0.63 5.13 0.03 86.38 09:35:02 0 1.22 0.00 0.40 12.92 0.02 85.44 09:35:02 1 0.83 0.00 0.22 0.50 0.00 98.45 09:35:02 2 0.18 0.00 0.17 0.03 0.02 99.60 09:35:02 3 7.71 0.00 0.47 0.55 0.02 91.26 09:35:02 4 15.39 0.00 0.95 23.73 0.03 59.89 09:35:02 5 13.99 0.00 0.98 0.73 0.02 84.28 09:35:02 6 20.31 0.00 1.57 2.60 0.07 75.45 09:35:02 7 3.10 0.00 0.25 0.00 0.00 96.64 09:36:01 all 8.69 0.00 0.63 4.96 0.03 85.69 09:36:01 0 5.97 0.00 0.64 5.21 0.00 88.18 09:36:01 1 0.73 0.00 0.27 0.25 0.02 98.73 09:36:01 2 6.15 0.00 0.77 21.40 0.03 71.64 09:36:01 3 18.55 0.00 1.05 7.98 0.05 72.36 09:36:01 4 6.83 0.00 0.47 0.64 0.03 92.02 09:36:01 5 1.12 0.00 0.14 0.00 0.02 98.73 09:36:01 6 8.75 0.00 0.78 0.26 0.02 90.20 09:36:01 7 21.37 0.00 0.93 3.95 0.05 73.69 09:37:01 all 9.64 0.00 4.23 9.43 0.06 76.64 09:37:01 0 10.20 0.00 3.54 12.36 0.07 73.82 09:37:01 1 8.92 0.00 4.90 14.98 0.07 71.14 09:37:01 2 9.13 0.00 4.30 11.17 0.05 75.35 09:37:01 3 9.42 0.00 3.94 7.51 0.05 79.07 09:37:01 4 9.06 0.00 4.35 6.13 0.07 80.39 09:37:01 5 10.32 0.00 5.01 2.17 0.07 82.43 09:37:01 6 10.53 0.00 3.68 1.14 0.07 84.59 09:37:01 7 9.50 0.00 4.08 19.99 0.07 66.36 09:38:01 all 17.20 0.00 4.99 6.78 0.08 70.95 09:38:01 0 9.85 0.00 3.78 5.44 0.07 80.86 09:38:01 1 19.62 0.00 4.89 27.97 0.10 47.42 09:38:01 2 17.21 0.00 5.58 1.57 0.08 75.55 09:38:01 3 17.97 0.00 4.18 2.87 0.08 74.89 09:38:01 4 17.41 0.00 5.16 3.37 0.07 73.99 09:38:01 5 20.18 0.00 4.77 2.50 0.08 72.46 09:38:01 6 16.69 0.00 5.50 2.19 0.08 75.54 09:38:01 7 18.77 0.00 6.05 8.48 0.08 66.62 09:39:01 all 21.22 0.00 2.02 0.27 0.07 76.41 09:39:01 0 23.07 0.00 2.10 0.00 0.10 74.72 09:39:01 1 24.30 0.00 2.39 1.87 0.07 71.37 09:39:01 2 26.59 0.00 2.45 0.17 0.07 70.73 09:39:01 3 21.61 0.00 1.81 0.02 0.05 76.51 09:39:01 4 17.47 0.00 1.77 0.03 0.07 80.66 09:39:01 5 21.53 0.00 2.09 0.00 0.05 76.33 09:39:01 6 20.54 0.00 1.92 0.02 0.08 77.44 09:39:01 7 14.65 0.00 1.64 0.03 0.08 83.59 09:40:01 all 1.05 0.00 0.18 0.73 0.05 97.99 09:40:01 0 1.42 0.00 0.20 0.00 0.07 98.32 09:40:01 1 1.29 0.00 0.13 5.73 0.03 92.82 09:40:01 2 0.87 0.00 0.20 0.13 0.05 98.75 09:40:01 3 0.90 0.00 0.18 0.00 0.03 98.88 09:40:01 4 0.95 0.00 0.15 0.00 0.05 98.85 09:40:01 5 0.83 0.00 0.20 0.02 0.07 98.88 09:40:01 6 1.37 0.00 0.23 0.00 0.08 98.31 09:40:01 7 0.74 0.00 0.13 0.02 0.03 99.08 09:41:01 all 3.42 0.00 0.65 2.23 0.04 93.66 09:41:01 0 1.95 0.00 0.62 0.42 0.05 96.96 09:41:01 1 1.74 0.00 0.57 12.70 0.05 84.94 09:41:01 2 13.53 0.00 0.77 1.10 0.05 84.54 09:41:01 3 1.60 0.00 0.60 1.12 0.03 96.65 09:41:01 4 3.40 0.00 0.74 0.10 0.03 95.73 09:41:01 5 1.25 0.00 0.64 1.29 0.03 96.79 09:41:01 6 1.14 0.00 0.72 0.43 0.02 97.70 09:41:01 7 2.76 0.00 0.60 0.68 0.03 95.92 Average: all 9.23 0.00 1.73 5.73 0.05 83.25 Average: 0 7.28 0.00 1.51 4.56 0.05 86.61 Average: 1 7.59 0.00 1.71 7.98 0.05 82.68 Average: 2 9.96 0.00 1.86 4.61 0.05 83.52 Average: 3 10.24 0.00 1.59 8.28 0.04 79.85 Average: 4 9.61 0.00 1.74 13.35 0.05 75.25 Average: 5 9.37 0.00 1.83 1.67 0.04 87.09 Average: 6 10.44 0.00 1.87 1.05 0.05 86.59 Average: 7 9.38 0.00 1.78 4.38 0.05 84.41