Last login: Wed Dec 30 08:55:48 on ttys000 gbrown1-m01:~ gbrown1$ cd datalearning/ .env gendata.py rentsite/ dbtest.py itinerary terraform.tfstate debezium/ landbot/ utils/ debezium.json postgresql/ docker-compose.yml rentbot/ gbrown1-m01:~ gbrown1$ cd datalearning/ .env gendata.py rentsite/ dbtest.py itinerary terraform.tfstate debezium/ landbot/ utils/ debezium.json postgresql/ docker-compose.yml rentbot/ gbrown1-m01:~ gbrown1$ cd datalearning/ gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> a313926ca0de Successfully built a313926ca0de Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> Using cache ---> 9309f5b45356 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Using cache ---> ccf337e03301 Successfully built ccf337e03301 Successfully tagged datalearning_landbot:latest datalearning_rentsite-db_1 is up-to-date datalearning_landbot-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_kafka_1 is up-to-date Starting datalearning_landbot_1 ... Starting datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ cd datalearning/ gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 76, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 67, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 51, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 55, in execute landbot_1 | result.raise_for_status() landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 943, in raise_for_status landbot_1 | raise HTTPError(http_error_msg, response=self) landbot_1 | requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://rentsite:8000/graphql landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 76, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 67, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 51, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 55, in execute landbot_1 | result.raise_for_status() landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 943, in raise_for_status landbot_1 | raise HTTPError(http_error_msg, response=self) landbot_1 | requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://rentsite:8000/graphql landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 76, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 67, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 51, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 55, in execute landbot_1 | result.raise_for_status() landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 943, in raise_for_status landbot_1 | raise HTTPError(http_error_msg, response=self) landbot_1 | requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://rentsite:8000/graphql gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 76, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 67, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 51, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 55, in execute landbot_1 | result.raise_for_status() landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 943, in raise_for_status landbot_1 | raise HTTPError(http_error_msg, response=self) landbot_1 | requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://rentsite:8000/graphql landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 76, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 67, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 51, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 55, in execute landbot_1 | result.raise_for_status() landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 943, in raise_for_status landbot_1 | raise HTTPError(http_error_msg, response=self) landbot_1 | requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://rentsite:8000/graphql landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 76, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 67, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 51, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 55, in execute landbot_1 | result.raise_for_status() landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 943, in raise_for_status landbot_1 | raise HTTPError(http_error_msg, response=self) landbot_1 | requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://rentsite:8000/graphql gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> a313926ca0de Successfully built a313926ca0de Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> e93fe209d419 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Running in 10751276f2c2 Removing intermediate container 10751276f2c2 ---> 11225d880bc0 Successfully built 11225d880bc0 Successfully tagged datalearning_landbot:latest datalearning_rentsite-db_1 is up-to-date datalearning_landbot-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_kafka_1 is up-to-date Recreating datalearning_landbot_1 ... Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 72, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 67, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 51, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 55, in execute landbot_1 | result.raise_for_status() landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 943, in raise_for_status landbot_1 | raise HTTPError(http_error_msg, response=self) landbot_1 | requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://rentsite:8000/graphql gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> a313926ca0de Successfully built a313926ca0de Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> 7349846ddce2 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Running in f6164ddca1df Removing intermediate container f6164ddca1df ---> 0dd29a830bc5 Successfully built 0dd29a830bc5 Successfully tagged datalearning_landbot:latest datalearning_rentsite-db_1 is up-to-date datalearning_landbot-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_kafka_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_connect_1 is up-to-date Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 52, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 55, in execute landbot_1 | result.raise_for_status() landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 943, in raise_for_status landbot_1 | raise HTTPError(http_error_msg, response=self) landbot_1 | requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://rentsite:8000/graphql landbot_1 | landbot_1 | During handling of the above exception, another exception occurred: landbot_1 | landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 77, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 72, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 53, in rentsite_apartment_create landbot_1 | except HTTPError as ex: landbot_1 | NameError: name 'HTTPError' is not defined gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> a313926ca0de Successfully built a313926ca0de Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> 2b48abbf88d0 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Running in 05571e817072 Removing intermediate container 05571e817072 ---> b5ab1bbac469 Successfully built b5ab1bbac469 Successfully tagged datalearning_landbot:latest datalearning_rentsite-db_1 is up-to-date datalearning_landbot-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_kafka_1 is up-to-date Recreating datalearning_landbot_1 ... Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 16, in landbot_1 | import urllib.HTTPError landbot_1 | ModuleNotFoundError: No module named 'urllib.HTTPError' gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> a313926ca0de Successfully built a313926ca0de Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> cb8cf5289d76 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Running in bcf35fa52543 Removing intermediate container bcf35fa52543 ---> 0910558fba0f Successfully built 0910558fba0f Successfully tagged datalearning_landbot:latest datalearning_rentsite-db_1 is up-to-date datalearning_landbot-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_kafka_1 is up-to-date Recreating datalearning_landbot_1 ... Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 16, in landbot_1 | import urllib.error.HTTPError landbot_1 | ModuleNotFoundError: No module named 'urllib.error.HTTPError'; 'urllib.error' is not a package gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> a313926ca0de Successfully built a313926ca0de Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> 55679dc362d3 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Running in 581473cfffa5 Removing intermediate container 581473cfffa5 ---> 4eb6543c1c12 Successfully built 4eb6543c1c12 Successfully tagged datalearning_landbot:latest datalearning_landbot-db_1 is up-to-date datalearning_rentsite-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_kafka_1 is up-to-date Recreating datalearning_landbot_1 ... Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 78, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 73, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 53, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 55, in execute landbot_1 | result.raise_for_status() landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/models.py", line 943, in raise_for_status landbot_1 | raise HTTPError(http_error_msg, response=self) landbot_1 | requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://rentsite:8000/graphql gbrown1-m01:datalearning gbrown1$ docker-compose logs rentsite Attaching to datalearning_rentsite_1 rentsite_1 | Watching for file changes with StatReloader rentsite_1 | Invalid HTTP_HOST header: 'rentsite:8000'. You may need to add 'rentsite' to ALLOWED_HOSTS. rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 22:34:59] "POST /graphql HTTP/1.1" 400 59169 rentsite_1 | Invalid HTTP_HOST header: 'rentsite:8000'. You may need to add 'rentsite' to ALLOWED_HOSTS. rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 22:39:34] "POST /graphql HTTP/1.1" 400 59169 rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 22:43:11] "POST /graphql HTTP/1.1" 400 332 rentsite_1 | Invalid HTTP_HOST header: 'rentsite:8000'. You may need to add 'rentsite' to ALLOWED_HOSTS. rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 22:43:47] "POST /graphql HTTP/1.1" 400 59169 rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 22:44:20] "POST /graphql HTTP/1.1" 400 133 rentsite_1 | Invalid HTTP_HOST header: 'rentsite:8000'. You may need to add 'rentsite' to ALLOWED_HOSTS. rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 22:50:50] "POST /graphql HTTP/1.1" 400 59169 rentsite_1 | Invalid HTTP_HOST header: 'rentsite:8000'. You may need to add 'rentsite' to ALLOWED_HOSTS. rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 22:52:42] "POST /graphql HTTP/1.1" 400 59169 rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 22:54:09] "POST /graphql HTTP/1.1" 400 133 rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 23:03:29] "POST /graphql HTTP/1.1" 400 133 rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 23:03:32] "POST /graphql HTTP/1.1" 400 133 rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 23:04:28] "POST /graphql HTTP/1.1" 400 190 rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 23:05:05] "POST /graphql HTTP/1.1" 400 199 rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 23:05:28] "POST /graphql HTTP/1.1" 400 203 rentsite_1 | Invalid HTTP_HOST header: 'rentsite:8000'. You may need to add 'rentsite' to ALLOWED_HOSTS. rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 23:07:00] "POST /graphql HTTP/1.1" 400 59169 rentsite_1 | Invalid HTTP_HOST header: 'rentsite:8000'. You may need to add 'rentsite' to ALLOWED_HOSTS. rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 23:17:27] "POST /graphql HTTP/1.1" 400 59169 rentsite_1 | Invalid HTTP_HOST header: 'rentsite:8000'. You may need to add 'rentsite' to ALLOWED_HOSTS. rentsite_1 | Bad Request: /graphql rentsite_1 | [30/Dec/2020 23:21:10] "POST /graphql HTTP/1.1" 400 59169 gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> 5e45547a8de6 Successfully built 5e45547a8de6 Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> Using cache ---> 55679dc362d3 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Using cache ---> 4eb6543c1c12 Successfully built 4eb6543c1c12 Successfully tagged datalearning_landbot:latest datalearning_rentsite-db_1 is up-to-date datalearning_landbot-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date Recreating datalearning_rentsite_1 ... datalearning_kafka_1 is up-to-date Recreating datalearning_rentsite_1 ... done Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | Traceback (most recent call last): landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 169, in _new_conn landbot_1 | conn = connection.create_connection( landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/util/connection.py", line 96, in create_connection landbot_1 | raise err landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/util/connection.py", line 86, in create_connection landbot_1 | sock.connect(sa) landbot_1 | ConnectionRefusedError: [Errno 111] Connection refused landbot_1 | landbot_1 | During handling of the above exception, another exception occurred: landbot_1 | landbot_1 | Traceback (most recent call last): landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 699, in urlopen landbot_1 | httplib_response = self._make_request( landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 394, in _make_request landbot_1 | conn.request(method, url, **httplib_request_kw) landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 234, in request landbot_1 | super(HTTPConnection, self).request(method, url, body=body, headers=headers) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 1255, in request landbot_1 | self._send_request(method, url, body, headers, encode_chunked) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 1301, in _send_request landbot_1 | self.endheaders(body, encode_chunked=encode_chunked) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 1250, in endheaders landbot_1 | self._send_output(message_body, encode_chunked=encode_chunked) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 1010, in _send_output landbot_1 | self.send(msg) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 950, in send landbot_1 | self.connect() landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 200, in connect landbot_1 | conn = self._new_conn() landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 181, in _new_conn landbot_1 | raise NewConnectionError( landbot_1 | urllib3.exceptions.NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused landbot_1 | landbot_1 | During handling of the above exception, another exception occurred: landbot_1 | landbot_1 | Traceback (most recent call last): landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 439, in send landbot_1 | resp = conn.urlopen( landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 755, in urlopen landbot_1 | retries = retries.increment( landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/util/retry.py", line 573, in increment landbot_1 | raise MaxRetryError(_pool, url, error or ResponseError(cause)) landbot_1 | urllib3.exceptions.MaxRetryError: HTTPConnectionPool(host='rentsite', port=8000): Max retries exceeded with url: /graphql (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) landbot_1 | landbot_1 | During handling of the above exception, another exception occurred: landbot_1 | landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 78, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 73, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 53, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 48, in execute landbot_1 | result = requests.post( landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/api.py", line 119, in post landbot_1 | return request('post', url, data=data, json=json, **kwargs) landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/api.py", line 61, in request landbot_1 | return session.request(method=method, url=url, **kwargs) landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 542, in request landbot_1 | resp = self.send(prep, **send_kwargs) landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 655, in send landbot_1 | r = adapter.send(request, **kwargs) landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 516, in send landbot_1 | raise ConnectionError(e, request=request) landbot_1 | requests.exceptions.ConnectionError: HTTPConnectionPool(host='rentsite', port=8000): Max retries exceeded with url: /graphql (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) gbrown1-m01:datalearning gbrown1$ docker-compose restart landbot Restarting datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | Traceback (most recent call last): landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 169, in _new_conn landbot_1 | conn = connection.create_connection( landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/util/connection.py", line 96, in create_connection landbot_1 | raise err landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/util/connection.py", line 86, in create_connection landbot_1 | sock.connect(sa) landbot_1 | ConnectionRefusedError: [Errno 111] Connection refused landbot_1 | landbot_1 | During handling of the above exception, another exception occurred: landbot_1 | landbot_1 | Traceback (most recent call last): landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 699, in urlopen landbot_1 | httplib_response = self._make_request( landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 394, in _make_request landbot_1 | conn.request(method, url, **httplib_request_kw) landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 234, in request landbot_1 | super(HTTPConnection, self).request(method, url, body=body, headers=headers) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 1255, in request landbot_1 | self._send_request(method, url, body, headers, encode_chunked) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 1301, in _send_request landbot_1 | self.endheaders(body, encode_chunked=encode_chunked) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 1250, in endheaders landbot_1 | self._send_output(message_body, encode_chunked=encode_chunked) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 1010, in _send_output landbot_1 | self.send(msg) landbot_1 | File "/usr/local/lib/python3.9/http/client.py", line 950, in send landbot_1 | self.connect() landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 200, in connect landbot_1 | conn = self._new_conn() landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connection.py", line 181, in _new_conn landbot_1 | raise NewConnectionError( landbot_1 | urllib3.exceptions.NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused landbot_1 | landbot_1 | During handling of the above exception, another exception occurred: landbot_1 | landbot_1 | Traceback (most recent call last): landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 439, in send landbot_1 | resp = conn.urlopen( landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/connectionpool.py", line 755, in urlopen landbot_1 | retries = retries.increment( landbot_1 | File "/usr/local/lib/python3.9/site-packages/urllib3/util/retry.py", line 573, in increment landbot_1 | raise MaxRetryError(_pool, url, error or ResponseError(cause)) landbot_1 | urllib3.exceptions.MaxRetryError: HTTPConnectionPool(host='rentsite', port=8000): Max retries exceeded with url: /graphql (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) landbot_1 | landbot_1 | During handling of the above exception, another exception occurred: landbot_1 | landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 78, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 73, in create_apartment landbot_1 | apartment.rentsite_id = rentsite_apartment_create(apartment) landbot_1 | File "//landbot.py", line 53, in rentsite_apartment_create landbot_1 | apartments = client.execute(query=create_apartment_mutation, variables=variables) landbot_1 | File "/usr/local/lib/python3.9/site-packages/python_graphql_client/graphql_client.py", line 48, in execute landbot_1 | result = requests.post( landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/api.py", line 119, in post landbot_1 | return request('post', url, data=data, json=json, **kwargs) landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/api.py", line 61, in request landbot_1 | return session.request(method=method, url=url, **kwargs) landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 542, in request landbot_1 | resp = self.send(prep, **send_kwargs) landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/sessions.py", line 655, in send landbot_1 | r = adapter.send(request, **kwargs) landbot_1 | File "/usr/local/lib/python3.9/site-packages/requests/adapters.py", line 516, in send landbot_1 | raise ConnectionError(e, request=request) landbot_1 | requests.exceptions.ConnectionError: HTTPConnectionPool(host='rentsite', port=8000): Max retries exceeded with url: /graphql (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 78, in landbot_1 | create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 74, in create_apartment landbot_1 | return db_create_apartment(apartment) landbot_1 | TypeError: db_create_apartment() missing 1 required positional argument: 'apartment' gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> 5e45547a8de6 Successfully built 5e45547a8de6 Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> fa2e71bdbbd3 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Running in 9b06da008559 Removing intermediate container 9b06da008559 ---> 2d72324dd21d Successfully built 2d72324dd21d Successfully tagged datalearning_landbot:latest datalearning_rentsite-db_1 is up-to-date datalearning_landbot-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_kafka_1 is up-to-date Recreating datalearning_landbot_1 ... Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> 5e45547a8de6 Successfully built 5e45547a8de6 Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> f57a3713f59b Step 8/8 : CMD ["python","-u","landbot.py"] ---> Running in 295645fd614b Removing intermediate container 295645fd614b ---> ac2b6ccf2adc Successfully built ac2b6ccf2adc Successfully tagged datalearning_landbot:latest datalearning_rentsite-db_1 is up-to-date datalearning_landbot-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_kafka_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_connect_1 is up-to-date Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | File "//landbot.py", line 79 landbot_1 | print apartment landbot_1 | ^ landbot_1 | SyntaxError: Missing parentheses in call to 'print'. Did you mean print(apartment)? gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> 5e45547a8de6 Successfully built 5e45547a8de6 Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> 4f4eb12d3970 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Running in 1023774564f3 Removing intermediate container 1023774564f3 ---> b03082de8531 Successfully built b03082de8531 Successfully tagged datalearning_landbot:latest datalearning_landbot-db_1 is up-to-date datalearning_rentsite-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_kafka_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_connect_1 is up-to-date Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose up -d --build Building rentsite Step 1/7 : FROM python:3-alpine ---> d4d4f50f871a Step 2/7 : RUN pip install pipenv ---> Using cache ---> 3a7692dd83a0 Step 3/7 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> 89c318cbe1ea Step 4/7 : COPY Pipfile . ---> Using cache ---> 63d381a8877c Step 5/7 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> bb2943819536 Step 6/7 : RUN pip install -r requirements.txt ---> Using cache ---> f11c6a840491 Step 7/7 : COPY . . ---> Using cache ---> 5e45547a8de6 Successfully built 5e45547a8de6 Successfully tagged datalearning_rentsite:latest Building landbot Step 1/8 : FROM python:3-alpine ---> d4d4f50f871a Step 2/8 : RUN apk add postgresql postgresql-contrib postgresql-dev g++ ---> Using cache ---> bf70b2e7fc73 Step 3/8 : RUN pip install pipenv ---> Using cache ---> 9b88652f7788 Step 4/8 : COPY Pipfile . ---> Using cache ---> f900b11193b7 Step 5/8 : RUN pipenv lock -r > requirements.txt ---> Using cache ---> cb496ec14885 Step 6/8 : RUN pip install -r requirements.txt ---> Using cache ---> fd6a2641c22b Step 7/8 : COPY . . ---> a07586c834f3 Step 8/8 : CMD ["python","-u","landbot.py"] ---> Running in 5b7425bded39 Removing intermediate container 5b7425bded39 ---> e497b38d3274 Successfully built e497b38d3274 Successfully tagged datalearning_landbot:latest datalearning_rentsite-db_1 is up-to-date datalearning_landbot-db_1 is up-to-date datalearning_zookeeper_1 is up-to-date datalearning_kafka_1 is up-to-date datalearning_rentsite_1 is up-to-date datalearning_connect_1 is up-to-date Recreating datalearning_landbot_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | gbrown1-m01:datalearning gbrown1$ docker ps CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 1315da9a5b8a datalearning_landbot "python -u landbot.py" 27 seconds ago Up 26 seconds datalearning_landbot_1 1a28df68654f datalearning_rentsite "python manage.py ru…" 6 minutes ago Up 6 minutes 0.0.0.0:8000->8000/tcp datalearning_rentsite_1 f0d08510159a debezium/connect:1.3 "/docker-entrypoint.…" About an hour ago Up About an hour 8778/tcp, 9092/tcp, 0.0.0.0:8083->8083/tcp, 9779/tcp datalearning_connect_1 14a8f729128e debezium/kafka:1.3 "/docker-entrypoint.…" About an hour ago Up About an hour 8778/tcp, 9779/tcp, 0.0.0.0:9092->9092/tcp datalearning_kafka_1 a074fc8b7ac5 debezium/zookeeper:1.3 "/docker-entrypoint.…" About an hour ago Up About an hour 0.0.0.0:2181->2181/tcp, 0.0.0.0:2888->2888/tcp, 8778/tcp, 0.0.0.0:3888->3888/tcp, 9779/tcp datalearning_zookeeper_1 f4c4e8aca13f postgres:12-alpine "docker-entrypoint.s…" About an hour ago Up About an hour 5432/tcp datalearning_landbot-db_1 55f646ca02c5 postgres:12-alpine "docker-entrypoint.s…" About an hour ago Up About an hour 5432/tcp datalearning_rentsite-db_1 gbrown1-m01:datalearning gbrown1$ docker ps CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 1a28df68654f datalearning_rentsite "python manage.py ru…" 6 minutes ago Up 6 minutes 0.0.0.0:8000->8000/tcp datalearning_rentsite_1 f0d08510159a debezium/connect:1.3 "/docker-entrypoint.…" About an hour ago Up About an hour 8778/tcp, 9092/tcp, 0.0.0.0:8083->8083/tcp, 9779/tcp datalearning_connect_1 14a8f729128e debezium/kafka:1.3 "/docker-entrypoint.…" About an hour ago Up About an hour 8778/tcp, 9779/tcp, 0.0.0.0:9092->9092/tcp datalearning_kafka_1 a074fc8b7ac5 debezium/zookeeper:1.3 "/docker-entrypoint.…" About an hour ago Up About an hour 0.0.0.0:2181->2181/tcp, 0.0.0.0:2888->2888/tcp, 8778/tcp, 0.0.0.0:3888->3888/tcp, 9779/tcp datalearning_zookeeper_1 f4c4e8aca13f postgres:12-alpine "docker-entrypoint.s…" About an hour ago Up About an hour 5432/tcp datalearning_landbot-db_1 55f646ca02c5 postgres:12-alpine "docker-entrypoint.s…" About an hour ago Up About an hour 5432/tcp datalearning_rentsite-db_1 gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 77, in landbot_1 | apartment = create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 73, in create_apartment landbot_1 | return db_create_apartment(engine,apartment) landbot_1 | File "//landbot.py", line 67, in db_create_apartment landbot_1 | session.commit() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 1046, in commit landbot_1 | self.transaction.commit() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 504, in commit landbot_1 | self._prepare_impl() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 483, in _prepare_impl landbot_1 | self.session.flush() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 2540, in flush landbot_1 | self._flush(objects) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 2682, in _flush landbot_1 | transaction.rollback(_capture_exception=True) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/langhelpers.py", line 68, in __exit__ landbot_1 | compat.raise_( landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 182, in raise_ landbot_1 | raise exception landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 2642, in _flush landbot_1 | flush_context.execute() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/unitofwork.py", line 422, in execute landbot_1 | rec.execute(self) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/unitofwork.py", line 586, in execute landbot_1 | persistence.save_obj( landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/persistence.py", line 205, in save_obj landbot_1 | for ( landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/persistence.py", line 373, in _organize_states_for_save landbot_1 | for state, dict_, mapper, connection in _connections_for_states( landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/persistence.py", line 1602, in _connections_for_states landbot_1 | connection = uowtransaction.transaction.connection(base_mapper) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 314, in connection landbot_1 | return self._connection_for_bind(bind, execution_options) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 421, in _connection_for_bind landbot_1 | conn = self._parent._connection_for_bind(bind, execution_options) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 433, in _connection_for_bind landbot_1 | conn = bind._contextual_connect() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2302, in _contextual_connect landbot_1 | self._wrap_pool_connect(self.pool.connect, None), landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2336, in _wrap_pool_connect landbot_1 | return fn() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/pool/base.py", line 364, in connect landbot_1 | return _ConnectionFairy._checkout(self) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/pool/base.py", line 778, in _checkout landbot_1 | fairy = _ConnectionRecord.checkout(pool) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/pool/base.py", line 495, in checkout landbot_1 | rec = pool._do_get() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/pool/impl.py", line 128, in _do_get landbot_1 | raise exc.TimeoutError( landbot_1 | sqlalchemy.exc.TimeoutError: QueuePool limit of size 5 overflow 10 reached, connection timed out, timeout 30 (Background on this error at: http://sqlalche.me/e/13/3o7r) gbrown1-m01:datalearning gbrown1$ docker-compose logs landbot Attaching to datalearning_landbot_1 landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | landbot_1 | Traceback (most recent call last): landbot_1 | File "//landbot.py", line 77, in landbot_1 | apartment = create_apartment("827 Cerrito St", "2","500",engine) landbot_1 | File "//landbot.py", line 73, in create_apartment landbot_1 | return db_create_apartment(engine,apartment) landbot_1 | File "//landbot.py", line 67, in db_create_apartment landbot_1 | session.commit() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 1046, in commit landbot_1 | self.transaction.commit() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 504, in commit landbot_1 | self._prepare_impl() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 483, in _prepare_impl landbot_1 | self.session.flush() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 2540, in flush landbot_1 | self._flush(objects) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 2682, in _flush landbot_1 | transaction.rollback(_capture_exception=True) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/langhelpers.py", line 68, in __exit__ landbot_1 | compat.raise_( landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 182, in raise_ landbot_1 | raise exception landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 2642, in _flush landbot_1 | flush_context.execute() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/unitofwork.py", line 422, in execute landbot_1 | rec.execute(self) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/unitofwork.py", line 586, in execute landbot_1 | persistence.save_obj( landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/persistence.py", line 205, in save_obj landbot_1 | for ( landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/persistence.py", line 373, in _organize_states_for_save landbot_1 | for state, dict_, mapper, connection in _connections_for_states( landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/persistence.py", line 1602, in _connections_for_states landbot_1 | connection = uowtransaction.transaction.connection(base_mapper) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 314, in connection landbot_1 | return self._connection_for_bind(bind, execution_options) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 421, in _connection_for_bind landbot_1 | conn = self._parent._connection_for_bind(bind, execution_options) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/orm/session.py", line 433, in _connection_for_bind landbot_1 | conn = bind._contextual_connect() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2302, in _contextual_connect landbot_1 | self._wrap_pool_connect(self.pool.connect, None), landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/engine/base.py", line 2336, in _wrap_pool_connect landbot_1 | return fn() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/pool/base.py", line 364, in connect landbot_1 | return _ConnectionFairy._checkout(self) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/pool/base.py", line 778, in _checkout landbot_1 | fairy = _ConnectionRecord.checkout(pool) landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/pool/base.py", line 495, in checkout landbot_1 | rec = pool._do_get() landbot_1 | File "/usr/local/lib/python3.9/site-packages/sqlalchemy/pool/impl.py", line 128, in _do_get landbot_1 | raise exc.TimeoutError( landbot_1 | sqlalchemy.exc.TimeoutError: QueuePool limit of size 5 overflow 10 reached, connection timed out, timeout 30 (Background on this error at: http://sqlalche.me/e/13/3o7r) gbrown1-m01:datalearning gbrown1$ docker-compose down Stopping datalearning_rentsite_1 ... done Stopping datalearning_connect_1 ... done Stopping datalearning_kafka_1 ... done Stopping datalearning_zookeeper_1 ... done Stopping datalearning_landbot-db_1 ... done Stopping datalearning_rentsite-db_1 ... done Removing datalearning_landbot_1 ... done Removing datalearning_rentsite_1 ... done Removing datalearning_connect_1 ... done Removing datalearning_kafka_1 ... done Removing datalearning_zookeeper_1 ... done Removing datalearning_landbot-db_1 ... done Removing datalearning_rentsite-db_1 ... done Removing network datalearning_default gbrown1-m01:datalearning gbrown1$ gbrown1-m01:datalearning gbrown1$ docker-compose up -d Creating network "datalearning_default" with the default driver Creating datalearning_rentsite-db_1 ... done Creating datalearning_landbot-db_1 ... done Creating datalearning_zookeeper_1 ... done Creating datalearning_kafka_1 ... done Creating datalearning_rentsite_1 ... done Creating datalearning_landbot_1 ... done Creating datalearning_connect_1 ... done gbrown1-m01:datalearning gbrown1$ docker-compose exec rentsite python manage.py makemigrations Migrations for 'aptlist': aptlist/migrations/0012_auto_20210104_1738.py - Alter field date_built on apartment - Alter field lease_expires on lease gbrown1-m01:datalearning gbrown1$ docker-compose exec rentsite python manage.py migrate Operations to perform: Apply all migrations: admin, aptlist, auth, contenttypes, sessions Running migrations: Applying contenttypes.0001_initial... OK Applying auth.0001_initial... OK Applying admin.0001_initial... OK Applying admin.0002_logentry_remove_auto_add... OK Applying admin.0003_logentry_add_action_flag_choices... OK Applying aptlist.0001_initial... OK /usr/local/lib/python3.9/site-packages/django/db/models/fields/__init__.py:1367: RuntimeWarning: DateTimeField Apartment.date_built received a naive datetime (2021-01-04 17:38:07.654137) while time zone support is active. warnings.warn("DateTimeField %s received a naive datetime (%s)" /usr/local/lib/python3.9/site-packages/django/db/models/fields/__init__.py:1367: RuntimeWarning: DateTimeField Apartment.date_built received a naive datetime (2021-01-04 17:38:07.656867) while time zone support is active. warnings.warn("DateTimeField %s received a naive datetime (%s)" /usr/local/lib/python3.9/site-packages/django/db/models/fields/__init__.py:1367: RuntimeWarning: DateTimeField Apartment.date_built received a naive datetime (2021-01-04 17:38:07.657020) while time zone support is active. warnings.warn("DateTimeField %s received a naive datetime (%s)" /usr/local/lib/python3.9/site-packages/django/db/models/fields/__init__.py:1367: RuntimeWarning: DateTimeField Apartment.date_rented received a naive datetime (2021-01-04 17:38:07.660550) while time zone support is active. warnings.warn("DateTimeField %s received a naive datetime (%s)" /usr/local/lib/python3.9/site-packages/django/db/models/fields/__init__.py:1367: RuntimeWarning: DateTimeField Apartment.date_rented received a naive datetime (2021-01-04 17:38:07.661626) while time zone support is active. warnings.warn("DateTimeField %s received a naive datetime (%s)" /usr/local/lib/python3.9/site-packages/django/db/models/fields/__init__.py:1367: RuntimeWarning: DateTimeField Apartment.date_rented received a naive datetime (2021-01-04 17:38:07.661719) while time zone support is active. warnings.warn("DateTimeField %s received a naive datetime (%s)" Applying aptlist.0002_auto_20201218_2315... OK Applying aptlist.0003_auto_20201218_2317... OK Applying aptlist.0004_auto_20201219_0029... OK Applying aptlist.0005_lease_active... OK Applying aptlist.0006_remove_lease_active... OK Applying aptlist.0007_auto_20201221_1817... OK /usr/local/lib/python3.9/site-packages/django/db/models/fields/__init__.py:1367: RuntimeWarning: DateTimeField Lease.lease_expires received a naive datetime (2021-01-04 17:38:07.718849) while time zone support is active. warnings.warn("DateTimeField %s received a naive datetime (%s)" /usr/local/lib/python3.9/site-packages/django/db/models/fields/__init__.py:1367: RuntimeWarning: DateTimeField Lease.lease_expires received a naive datetime (2021-01-04 17:38:07.719017) while time zone support is active. warnings.warn("DateTimeField %s received a naive datetime (%s)" /usr/local/lib/python3.9/site-packages/django/db/models/fields/__init__.py:1367: RuntimeWarning: DateTimeField Lease.lease_expires received a naive datetime (2021-01-04 17:38:07.721996) while time zone support is active. warnings.warn("DateTimeField %s received a naive datetime (%s)" Applying aptlist.0008_auto_20201221_1855... OK Applying aptlist.0009_auto_20201221_1901... OK Applying aptlist.0010_auto_20201221_1901... OK Applying aptlist.0011_auto_20201221_1913... OK Applying aptlist.0012_auto_20210104_1738... OK Applying contenttypes.0002_remove_content_type_name... OK Applying auth.0002_alter_permission_name_max_length... OK Applying auth.0003_alter_user_email_max_length... OK Applying auth.0004_alter_user_username_opts... OK Applying auth.0005_alter_user_last_login_null... OK Applying auth.0006_require_contenttypes_0002... OK Applying auth.0007_alter_validators_add_error_messages... OK Applying auth.0008_alter_user_username_max_length... OK Applying auth.0009_alter_user_last_name_max_length... OK Applying auth.0010_alter_group_name_max_length... OK Applying auth.0011_update_proxy_permissions... OK Applying auth.0012_alter_user_first_name_max_length... OK Applying sessions.0001_initial... OK gbrown1-m01:datalearning gbrown1$ curl -i -X POST -H "Accept:application/json" -H "Content-Type:application/json" localhost:8083/connectors/ -d @debezium.json HTTP/1.1 500 Internal Server Error Date: Mon, 04 Jan 2021 17:38:19 GMT Content-Type: application/json Content-Length: 267 Server: Jetty(9.4.24.v20191120) {"error_code":500,"message":"Unrecognized token 'Here': was expecting (JSON String, Number, Array, Object or token 'null', 'true' or 'false')\n at [Source: (org.glassfish.jersey.message.internal.ReaderInterceptorExecutor$UnCloseableInputStream); line: 1, column: 6]"}gbrown1-m01:datalearning gbrown1$ pwd /Users/gbrown1/datalearning gbrown1-m01:datalearning gbrown1$ curl -i -X POST -H "Accept:application/json" -H "Content-Type:application/json" localhost:8083/connectors/ -d @debezium.json HTTP/1.1 201 Created Date: Mon, 04 Jan 2021 17:38:42 GMT Location: http://localhost:8083/connectors/rentsite-connector Content-Type: application/json Content-Length: 492 Server: Jetty(9.4.24.v20191120) {"name":"rentsite-connector","config":{"connector.class":"io.debezium.connector.postgresql.PostgresConnector","database.hostname":"rentsite-db","database.port":"5432","database.user":"rentsite","database.password":"rentsite","database.dbname":"rentsite","database.server.name":"rentsite-db","table.include.list":"rentsite.aptlist_lease","schema.include.list":"rentsite","publication.autocreate.mode":"filtered","plugin.name":"pgoutput","name":"rentsite-connector"},"tasks":[],"typgbrown1-m01:datalearning gbrown1$ docker-compose logs connect Attaching to datalearning_connect_1 connect_1 | Plugins are loaded from /kafka/connect connect_1 | Using the following environment variables: connect_1 | GROUP_ID=1 connect_1 | CONFIG_STORAGE_TOPIC=my_connect_configs connect_1 | OFFSET_STORAGE_TOPIC=my_connect_offsets connect_1 | STATUS_STORAGE_TOPIC=my_connect_statuses connect_1 | BOOTSTRAP_SERVERS=kafka:9092 connect_1 | REST_HOST_NAME=192.168.128.8 connect_1 | REST_PORT=8083 connect_1 | ADVERTISED_HOST_NAME=connect connect_1 | ADVERTISED_PORT=8083 connect_1 | KEY_CONVERTER=org.apache.kafka.connect.json.JsonConverter connect_1 | VALUE_CONVERTER=org.apache.kafka.connect.json.JsonConverter connect_1 | INTERNAL_KEY_CONVERTER=org.apache.kafka.connect.json.JsonConverter connect_1 | INTERNAL_VALUE_CONVERTER=org.apache.kafka.connect.json.JsonConverter connect_1 | OFFSET_FLUSH_INTERVAL_MS=60000 connect_1 | OFFSET_FLUSH_TIMEOUT_MS=5000 connect_1 | SHUTDOWN_TIMEOUT=10000 connect_1 | --- Setting property from CONNECT_INTERNAL_VALUE_CONVERTER: internal.value.converter=org.apache.kafka.connect.json.JsonConverter connect_1 | --- Setting property from CONNECT_VALUE_CONVERTER: value.converter=org.apache.kafka.connect.json.JsonConverter connect_1 | --- Setting property from CONNECT_REST_ADVERTISED_HOST_NAME: rest.advertised.host.name=connect connect_1 | --- Setting property from CONNECT_OFFSET_FLUSH_INTERVAL_MS: offset.flush.interval.ms=60000 connect_1 | --- Setting property from CONNECT_GROUP_ID: group.id=1 connect_1 | --- Setting property from CONNECT_BOOTSTRAP_SERVERS: bootstrap.servers=kafka:9092 connect_1 | --- Setting property from CONNECT_KEY_CONVERTER: key.converter=org.apache.kafka.connect.json.JsonConverter connect_1 | --- Setting property from CONNECT_TASK_SHUTDOWN_GRACEFUL_TIMEOUT_MS: task.shutdown.graceful.timeout.ms=10000 connect_1 | --- Setting property from CONNECT_REST_HOST_NAME: rest.host.name=192.168.128.8 connect_1 | --- Setting property from CONNECT_PLUGIN_PATH: plugin.path=/kafka/connect connect_1 | --- Setting property from CONNECT_REST_PORT: rest.port=8083 connect_1 | --- Setting property from CONNECT_OFFSET_FLUSH_TIMEOUT_MS: offset.flush.timeout.ms=5000 connect_1 | --- Setting property from CONNECT_STATUS_STORAGE_TOPIC: status.storage.topic=my_connect_statuses connect_1 | --- Setting property from CONNECT_INTERNAL_KEY_CONVERTER: internal.key.converter=org.apache.kafka.connect.json.JsonConverter connect_1 | --- Setting property from CONNECT_CONFIG_STORAGE_TOPIC: config.storage.topic=my_connect_configs connect_1 | --- Setting property from CONNECT_REST_ADVERTISED_PORT: rest.advertised.port=8083 connect_1 | --- Setting property from CONNECT_OFFSET_STORAGE_TOPIC: offset.storage.topic=my_connect_offsets connect_1 | 2021-01-04 17:37:56,762 INFO || WorkerInfo values: connect_1 | jvm.args = -Xms256M, -Xmx2G, -XX:+UseG1GC, -XX:MaxGCPauseMillis=20, -XX:InitiatingHeapOccupancyPercent=35, -XX:+ExplicitGCInvokesConcurrent, -XX:MaxInlineLevel=15, -Djava.awt.headless=true, -Dcom.sun.management.jmxremote, -Dcom.sun.management.jmxremote.authenticate=false, -Dcom.sun.management.jmxremote.ssl=false, -Dkafka.logs.dir=/kafka/bin/../logs, -Dlog4j.configuration=file:/kafka/config/log4j.properties connect_1 | jvm.spec = Oracle Corporation, OpenJDK 64-Bit Server VM, 11.0.8, 11.0.8+10-LTS connect_1 | jvm.classpath = /kafka/bin/../libs/activation-1.1.1.jar:/kafka/bin/../libs/aopalliance-repackaged-2.5.0.jar:/kafka/bin/../libs/argparse4j-0.7.0.jar:/kafka/bin/../libs/audience-annotations-0.5.0.jar:/kafka/bin/../libs/avro-1.9.2.jar:/kafka/bin/../libs/common-config-5.5.1.jar:/kafka/bin/../libs/common-utils-5.5.1.jar:/kafka/bin/../libs/commons-cli-1.4.jar:/kafka/bin/../libs/commons-lang3-3.8.1.jar:/kafka/bin/../libs/connect-api-2.6.0.jar:/kafka/bin/../libs/connect-basic-auth-extension-2.6.0.jar:/kafka/bin/../libs/connect-file-2.6.0.jar:/kafka/bin/../libs/connect-json-2.6.0.jar:/kafka/bin/../libs/connect-mirror-2.6.0.jar:/kafka/bin/../libs/connect-mirror-client-2.6.0.jar:/kafka/bin/../libs/connect-runtime-2.6.0.jar:/kafka/bin/../libs/connect-transforms-2.6.0.jar:/kafka/bin/../libs/hk2-api-2.5.0.jar:/kafka/bin/../libs/hk2-locator-2.5.0.jar:/kafka/bin/../libs/hk2-utils-2.5.0.jar:/kafka/bin/../libs/jackson-annotations-2.10.2.jar:/kafka/bin/../libs/jackson-core-2.10.2.jar:/kafka/bin/../libs/jackson-databind-2.10.2.jar:/kafka/bin/../libs/jackson-dataformat-csv-2.10.2.jar:/kafka/bin/../libs/jackson-datatype-jdk8-2.10.2.jar:/kafka/bin/../libs/jackson-jaxrs-base-2.10.2.jar:/kafka/bin/../libs/jackson-jaxrs-json-provider-2.10.2.jar:/kafka/bin/../libs/jackson-module-jaxb-annotations-2.10.2.jar:/kafka/bin/../libs/jackson-module-paranamer-2.10.2.jar:/kafka/bin/../libs/jackson-module-scala_2.12-2.10.2.jar:/kafka/bin/../libs/jakarta.activation-api-1.2.1.jar:/kafka/bin/../libs/jakarta.annotation-api-1.3.4.jar:/kafka/bin/../libs/jakarta.inject-2.5.0.jar:/kafka/bin/../libs/jakarta.ws.rs-api-2.1.5.jar:/kafka/bin/../libs/jakarta.xml.bind-api-2.3.2.jar:/kafka/bin/../libs/javassist-3.22.0-CR2.jar:/kafka/bin/../libs/javassist-3.26.0-GA.jar:/kafka/bin/../libs/javax.servlet-api-3.1.0.jar:/kafka/bin/../libs/javax.ws.rs-api-2.1.1.jar:/kafka/bin/../libs/jaxb-api-2.3.0.jar:/kafka/bin/../libs/jersey-client-2.28.jar:/kafka/bin/../libs/jersey-common-2.28.jar:/kafka/bin/../libs/jersey-container-servlet-2.28.jar:/kafka/bin/../libs/jersey-container-servlet-core-2.28.jar:/kafka/bin/../libs/jersey-hk2-2.28.jar:/kafka/bin/../libs/jersey-media-jaxb-2.28.jar:/kafka/bin/../libs/jersey-server-2.28.jar:/kafka/bin/../libs/jetty-client-9.4.24.v20191120.jar:/kafka/bin/../libs/jetty-continuation-9.4.24.v20191120.jar:/kafka/bin/../libs/jetty-http-9.4.24.v20191120.jar:/kafka/bin/../libs/jetty-io-9.4.24.v20191120.jar:/kafka/bin/../libs/jetty-security-9.4.24.v20191120.jar:/kafka/bin/../libs/jetty-server-9.4.24.v20191120.jar:/kafka/bin/../libs/jetty-servlet-9.4.24.v20191120.jar:/kafka/bin/../libs/jetty-servlets-9.4.24.v20191120.jar:/kafka/bin/../libs/jetty-util-9.4.24.v20191120.jar:/kafka/bin/../libs/jopt-simple-5.0.4.jar:/kafka/bin/../libs/kafka-avro-serializer-5.5.1.jar:/kafka/bin/../libs/kafka-clients-2.6.0.jar:/kafka/bin/../libs/kafka-connect-avro-converter-5.5.1.jar:/kafka/bin/../libs/kafka-connect-avro-data-5.5.1.jar:/kafka/bin/../libs/kafka-log4j-appender-2.6.0.jar:/kafka/bin/../libs/kafka-schema-registry-client-5.5.1.jar:/kafka/bin/../libs/kafka-schema-serializer-5.5.1.jar:/kafka/bin/../libs/kafka-streams-2.6.0.jar:/kafka/bin/../libs/kafka-streams-examples-2.6.0.jar:/kafka/bin/../libs/kafka-streams-scala_2.12-2.6.0.jar:/kafka/bin/../libs/kafka-streams-test-utils-2.6.0.jar:/kafka/bin/../libs/kafka-tools-2.6.0.jar:/kafka/bin/../libs/kafka_2.12-2.6.0.jar:/kafka/bin/../libs/log4j-1.2.17.jar:/kafka/bin/../libs/lz4-java-1.7.1.jar:/kafka/bin/../libs/maven-artifact-3.6.3.jar:/kafka/bin/../libs/metrics-core-2.2.0.jar:/kafka/bin/../libs/netty-buffer-4.1.50.Final.jar:/kafka/bin/../libs/netty-codec-4.1.50.Final.jar:/kafka/bin/../libs/netty-common-4.1.50.Final.jar:/kafka/bin/../libs/netty-handler-4.1.50.Final.jar:/kafka/bin/../libs/netty-resolver-4.1.50.Final.jar:/kafka/bin/../libs/netty-transport-4.1.50.Final.jar:/kafka/bin/../libs/netty-transport-native-epoll-4.1.50.Final.jar:/kafka/bin/../libs/netty-transport-native-unix-common-4.1.50.Final.jar:/kafka/bin/../libs/osgi-resource-locator-1.0.1.jar:/kafka/bin/../libs/paranamer-2.8.jar:/kafka/bin/../libs/plexus-utils-3.2.1.jar:/kafka/bin/../libs/reflections-0.9.12.jar:/kafka/bin/../libs/rocksdbjni-5.18.4.jar:/kafka/bin/../libs/scala-collection-compat_2.12-2.1.6.jar:/kafka/bin/../libs/scala-java8-compat_2.12-0.9.1.jar:/kafka/bin/../libs/scala-library-2.12.11.jar:/kafka/bin/../libs/scala-logging_2.12-3.9.2.jar:/kafka/bin/../libs/scala-reflect-2.12.11.jar:/kafka/bin/../libs/slf4j-api-1.7.30.jar:/kafka/bin/../libs/slf4j-log4j12-1.7.30.jar:/kafka/bin/../libs/snappy-java-1.1.7.3.jar:/kafka/bin/../libs/validation-api-2.0.1.Final.jar:/kafka/bin/../libs/zookeeper-3.5.8.jar:/kafka/bin/../libs/zookeeper-jute-3.5.8.jar:/kafka/bin/../libs/zstd-jni-1.4.4-7.jar connect_1 | os.spec = Linux, amd64, 5.4.39-linuxkit connect_1 | os.vcpus = 4 connect_1 | [org.apache.kafka.connect.runtime.WorkerInfo] connect_1 | 2021-01-04 17:37:56,780 INFO || Scanning for plugin classes. This might take a moment ... [org.apache.kafka.connect.cli.ConnectDistributed] connect_1 | 2021-01-04 17:37:56,824 INFO || Loading plugin from: /kafka/connect/debezium-connector-oracle [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,529 INFO || Registered loader: PluginClassLoader{pluginLocation=file:/kafka/connect/debezium-connector-oracle/} [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,529 INFO || Added plugin 'io.debezium.connector.oracle.OracleConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,530 INFO || Added plugin 'io.debezium.converters.ByteBufferConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,530 INFO || Added plugin 'io.debezium.converters.CloudEventsConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,531 INFO || Added plugin 'io.debezium.transforms.outbox.EventRouter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,531 INFO || Added plugin 'io.debezium.transforms.ExtractNewRecordState' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,531 INFO || Added plugin 'io.debezium.transforms.ByLogicalTableRouter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,532 INFO || Added plugin 'org.apache.kafka.connect.connector.policy.AllConnectorClientConfigOverridePolicy' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,532 INFO || Added plugin 'org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicy' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,532 INFO || Added plugin 'org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicy' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,534 INFO || Loading plugin from: /kafka/connect/debezium-connector-db2 [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,654 INFO || Registered loader: PluginClassLoader{pluginLocation=file:/kafka/connect/debezium-connector-db2/} [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,654 INFO || Added plugin 'io.debezium.connector.db2.Db2Connector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:57,655 INFO || Loading plugin from: /kafka/connect/debezium-connector-mysql [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,003 INFO || Registered loader: PluginClassLoader{pluginLocation=file:/kafka/connect/debezium-connector-mysql/} [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,003 INFO || Added plugin 'io.debezium.connector.mysql.MySqlConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,018 INFO || Loading plugin from: /kafka/connect/debezium-connector-sqlserver [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,170 INFO || Registered loader: PluginClassLoader{pluginLocation=file:/kafka/connect/debezium-connector-sqlserver/} [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,170 INFO || Added plugin 'io.debezium.connector.sqlserver.SqlServerConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,199 INFO || Loading plugin from: /kafka/connect/debezium-connector-mongodb [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,442 INFO || Registered loader: PluginClassLoader{pluginLocation=file:/kafka/connect/debezium-connector-mongodb/} [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,443 INFO || Added plugin 'io.debezium.connector.mongodb.MongoDbConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,443 INFO || Added plugin 'io.debezium.connector.mongodb.transforms.ExtractNewDocumentState' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,444 INFO || Loading plugin from: /kafka/connect/debezium-connector-postgres [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,700 INFO || Registered loader: PluginClassLoader{pluginLocation=file:/kafka/connect/debezium-connector-postgres/} [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:37:58,700 INFO || Added plugin 'io.debezium.connector.postgresql.PostgresConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,556 INFO || Registered loader: jdk.internal.loader.ClassLoaders$AppClassLoader@3d4eac69 [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,557 INFO || Added plugin 'org.apache.kafka.connect.mirror.MirrorSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,557 INFO || Added plugin 'org.apache.kafka.connect.file.FileStreamSinkConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,557 INFO || Added plugin 'org.apache.kafka.connect.tools.SchemaSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,557 INFO || Added plugin 'org.apache.kafka.connect.tools.MockSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,557 INFO || Added plugin 'org.apache.kafka.connect.file.FileStreamSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,558 INFO || Added plugin 'org.apache.kafka.connect.mirror.MirrorCheckpointConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,558 INFO || Added plugin 'org.apache.kafka.connect.tools.VerifiableSinkConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,558 INFO || Added plugin 'org.apache.kafka.connect.tools.VerifiableSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,558 INFO || Added plugin 'org.apache.kafka.connect.tools.MockSinkConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,558 INFO || Added plugin 'org.apache.kafka.connect.mirror.MirrorHeartbeatConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,559 INFO || Added plugin 'org.apache.kafka.connect.tools.MockConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,559 INFO || Added plugin 'org.apache.kafka.connect.converters.FloatConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,560 INFO || Added plugin 'org.apache.kafka.connect.converters.DoubleConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,560 INFO || Added plugin 'org.apache.kafka.connect.converters.ByteArrayConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,560 INFO || Added plugin 'io.confluent.connect.avro.AvroConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,560 INFO || Added plugin 'org.apache.kafka.connect.converters.LongConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,560 INFO || Added plugin 'org.apache.kafka.connect.converters.IntegerConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,561 INFO || Added plugin 'org.apache.kafka.connect.json.JsonConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,561 INFO || Added plugin 'org.apache.kafka.connect.storage.StringConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,561 INFO || Added plugin 'org.apache.kafka.connect.converters.ShortConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,561 INFO || Added plugin 'org.apache.kafka.connect.storage.SimpleHeaderConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,561 INFO || Added plugin 'org.apache.kafka.connect.transforms.ReplaceField$Value' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,561 INFO || Added plugin 'org.apache.kafka.connect.transforms.SetSchemaMetadata$Value' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,561 INFO || Added plugin 'org.apache.kafka.connect.transforms.ReplaceField$Key' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,562 INFO || Added plugin 'org.apache.kafka.connect.transforms.Filter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,562 INFO || Added plugin 'org.apache.kafka.connect.transforms.InsertField$Value' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,562 INFO || Added plugin 'org.apache.kafka.connect.transforms.TimestampConverter$Key' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,562 INFO || Added plugin 'org.apache.kafka.connect.transforms.MaskField$Value' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,562 INFO || Added plugin 'org.apache.kafka.connect.transforms.TimestampRouter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,562 INFO || Added plugin 'org.apache.kafka.connect.transforms.RegexRouter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,563 INFO || Added plugin 'org.apache.kafka.connect.transforms.HoistField$Value' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,563 INFO || Added plugin 'org.apache.kafka.connect.transforms.ValueToKey' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,563 INFO || Added plugin 'org.apache.kafka.connect.transforms.MaskField$Key' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,563 INFO || Added plugin 'org.apache.kafka.connect.transforms.Cast$Key' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,563 INFO || Added plugin 'org.apache.kafka.connect.transforms.Cast$Value' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,563 INFO || Added plugin 'org.apache.kafka.connect.runtime.PredicatedTransformation' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,563 INFO || Added plugin 'org.apache.kafka.connect.transforms.ExtractField$Key' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,564 INFO || Added plugin 'org.apache.kafka.connect.transforms.Flatten$Value' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,564 INFO || Added plugin 'org.apache.kafka.connect.transforms.InsertField$Key' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,564 INFO || Added plugin 'org.apache.kafka.connect.transforms.Flatten$Key' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,564 INFO || Added plugin 'org.apache.kafka.connect.transforms.SetSchemaMetadata$Key' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,564 INFO || Added plugin 'org.apache.kafka.connect.transforms.ExtractField$Value' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,564 INFO || Added plugin 'org.apache.kafka.connect.transforms.TimestampConverter$Value' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,564 INFO || Added plugin 'org.apache.kafka.connect.transforms.HoistField$Key' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,565 INFO || Added plugin 'org.apache.kafka.connect.transforms.predicates.HasHeaderKey' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,565 INFO || Added plugin 'org.apache.kafka.connect.transforms.predicates.RecordIsTombstone' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,565 INFO || Added plugin 'org.apache.kafka.connect.transforms.predicates.TopicNameMatches' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,565 INFO || Added plugin 'org.apache.kafka.common.config.provider.FileConfigProvider' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,565 INFO || Added plugin 'org.apache.kafka.connect.rest.basic.auth.extension.BasicAuthSecurityRestExtension' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,566 INFO || Added aliases 'Db2Connector' and 'Db2' to plugin 'io.debezium.connector.db2.Db2Connector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,567 INFO || Added aliases 'MongoDbConnector' and 'MongoDb' to plugin 'io.debezium.connector.mongodb.MongoDbConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,567 INFO || Added aliases 'MySqlConnector' and 'MySql' to plugin 'io.debezium.connector.mysql.MySqlConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,567 INFO || Added aliases 'OracleConnector' and 'Oracle' to plugin 'io.debezium.connector.oracle.OracleConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,568 INFO || Added aliases 'PostgresConnector' and 'Postgres' to plugin 'io.debezium.connector.postgresql.PostgresConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,568 INFO || Added aliases 'SqlServerConnector' and 'SqlServer' to plugin 'io.debezium.connector.sqlserver.SqlServerConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,568 INFO || Added aliases 'FileStreamSinkConnector' and 'FileStreamSink' to plugin 'org.apache.kafka.connect.file.FileStreamSinkConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,568 INFO || Added aliases 'FileStreamSourceConnector' and 'FileStreamSource' to plugin 'org.apache.kafka.connect.file.FileStreamSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,569 INFO || Added aliases 'MirrorCheckpointConnector' and 'MirrorCheckpoint' to plugin 'org.apache.kafka.connect.mirror.MirrorCheckpointConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,569 INFO || Added aliases 'MirrorHeartbeatConnector' and 'MirrorHeartbeat' to plugin 'org.apache.kafka.connect.mirror.MirrorHeartbeatConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,569 INFO || Added aliases 'MirrorSourceConnector' and 'MirrorSource' to plugin 'org.apache.kafka.connect.mirror.MirrorSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,569 INFO || Added aliases 'MockConnector' and 'Mock' to plugin 'org.apache.kafka.connect.tools.MockConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,570 INFO || Added aliases 'MockSinkConnector' and 'MockSink' to plugin 'org.apache.kafka.connect.tools.MockSinkConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,570 INFO || Added aliases 'MockSourceConnector' and 'MockSource' to plugin 'org.apache.kafka.connect.tools.MockSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,570 INFO || Added aliases 'SchemaSourceConnector' and 'SchemaSource' to plugin 'org.apache.kafka.connect.tools.SchemaSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,571 INFO || Added aliases 'VerifiableSinkConnector' and 'VerifiableSink' to plugin 'org.apache.kafka.connect.tools.VerifiableSinkConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,571 INFO || Added aliases 'VerifiableSourceConnector' and 'VerifiableSource' to plugin 'org.apache.kafka.connect.tools.VerifiableSourceConnector' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,571 INFO || Added aliases 'AvroConverter' and 'Avro' to plugin 'io.confluent.connect.avro.AvroConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,571 INFO || Added aliases 'ByteBufferConverter' and 'ByteBuffer' to plugin 'io.debezium.converters.ByteBufferConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,571 INFO || Added aliases 'CloudEventsConverter' and 'CloudEvents' to plugin 'io.debezium.converters.CloudEventsConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,572 INFO || Added aliases 'ByteArrayConverter' and 'ByteArray' to plugin 'org.apache.kafka.connect.converters.ByteArrayConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,572 INFO || Added aliases 'DoubleConverter' and 'Double' to plugin 'org.apache.kafka.connect.converters.DoubleConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,572 INFO || Added aliases 'FloatConverter' and 'Float' to plugin 'org.apache.kafka.connect.converters.FloatConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,573 INFO || Added aliases 'IntegerConverter' and 'Integer' to plugin 'org.apache.kafka.connect.converters.IntegerConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,573 INFO || Added aliases 'LongConverter' and 'Long' to plugin 'org.apache.kafka.connect.converters.LongConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,573 INFO || Added aliases 'ShortConverter' and 'Short' to plugin 'org.apache.kafka.connect.converters.ShortConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,573 INFO || Added aliases 'JsonConverter' and 'Json' to plugin 'org.apache.kafka.connect.json.JsonConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,574 INFO || Added aliases 'StringConverter' and 'String' to plugin 'org.apache.kafka.connect.storage.StringConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,574 INFO || Added aliases 'ByteBufferConverter' and 'ByteBuffer' to plugin 'io.debezium.converters.ByteBufferConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,574 INFO || Added aliases 'ByteArrayConverter' and 'ByteArray' to plugin 'org.apache.kafka.connect.converters.ByteArrayConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,574 INFO || Added aliases 'DoubleConverter' and 'Double' to plugin 'org.apache.kafka.connect.converters.DoubleConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,575 INFO || Added aliases 'FloatConverter' and 'Float' to plugin 'org.apache.kafka.connect.converters.FloatConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,575 INFO || Added aliases 'IntegerConverter' and 'Integer' to plugin 'org.apache.kafka.connect.converters.IntegerConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,575 INFO || Added aliases 'LongConverter' and 'Long' to plugin 'org.apache.kafka.connect.converters.LongConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,575 INFO || Added aliases 'ShortConverter' and 'Short' to plugin 'org.apache.kafka.connect.converters.ShortConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,575 INFO || Added aliases 'JsonConverter' and 'Json' to plugin 'org.apache.kafka.connect.json.JsonConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,576 INFO || Added alias 'SimpleHeaderConverter' to plugin 'org.apache.kafka.connect.storage.SimpleHeaderConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,576 INFO || Added aliases 'StringConverter' and 'String' to plugin 'org.apache.kafka.connect.storage.StringConverter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,577 INFO || Added alias 'ExtractNewDocumentState' to plugin 'io.debezium.connector.mongodb.transforms.ExtractNewDocumentState' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,577 INFO || Added alias 'ByLogicalTableRouter' to plugin 'io.debezium.transforms.ByLogicalTableRouter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,578 INFO || Added alias 'ExtractNewRecordState' to plugin 'io.debezium.transforms.ExtractNewRecordState' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,578 INFO || Added alias 'EventRouter' to plugin 'io.debezium.transforms.outbox.EventRouter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,578 INFO || Added aliases 'PredicatedTransformation' and 'Predicated' to plugin 'org.apache.kafka.connect.runtime.PredicatedTransformation' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,579 INFO || Added alias 'Filter' to plugin 'org.apache.kafka.connect.transforms.Filter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,579 INFO || Added alias 'RegexRouter' to plugin 'org.apache.kafka.connect.transforms.RegexRouter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,580 INFO || Added alias 'TimestampRouter' to plugin 'org.apache.kafka.connect.transforms.TimestampRouter' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,580 INFO || Added alias 'ValueToKey' to plugin 'org.apache.kafka.connect.transforms.ValueToKey' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,580 INFO || Added alias 'HasHeaderKey' to plugin 'org.apache.kafka.connect.transforms.predicates.HasHeaderKey' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,580 INFO || Added alias 'RecordIsTombstone' to plugin 'org.apache.kafka.connect.transforms.predicates.RecordIsTombstone' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,581 INFO || Added alias 'TopicNameMatches' to plugin 'org.apache.kafka.connect.transforms.predicates.TopicNameMatches' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,581 INFO || Added alias 'BasicAuthSecurityRestExtension' to plugin 'org.apache.kafka.connect.rest.basic.auth.extension.BasicAuthSecurityRestExtension' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,581 INFO || Added aliases 'AllConnectorClientConfigOverridePolicy' and 'All' to plugin 'org.apache.kafka.connect.connector.policy.AllConnectorClientConfigOverridePolicy' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,581 INFO || Added aliases 'NoneConnectorClientConfigOverridePolicy' and 'None' to plugin 'org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicy' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,581 INFO || Added aliases 'PrincipalConnectorClientConfigOverridePolicy' and 'Principal' to plugin 'org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicy' [org.apache.kafka.connect.runtime.isolation.DelegatingClassLoader] connect_1 | 2021-01-04 17:38:00,703 INFO || DistributedConfig values: connect_1 | access.control.allow.methods = connect_1 | access.control.allow.origin = connect_1 | admin.listeners = null connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | config.providers = [] connect_1 | config.storage.replication.factor = 1 connect_1 | config.storage.topic = my_connect_configs connect_1 | connect.protocol = sessioned connect_1 | connections.max.idle.ms = 540000 connect_1 | connector.client.config.override.policy = None connect_1 | group.id = 1 connect_1 | header.converter = class org.apache.kafka.connect.storage.SimpleHeaderConverter connect_1 | heartbeat.interval.ms = 3000 connect_1 | inter.worker.key.generation.algorithm = HmacSHA256 connect_1 | inter.worker.key.size = null connect_1 | inter.worker.key.ttl.ms = 3600000 connect_1 | inter.worker.signature.algorithm = HmacSHA256 connect_1 | inter.worker.verification.algorithms = [HmacSHA256] connect_1 | internal.key.converter = class org.apache.kafka.connect.json.JsonConverter connect_1 | internal.value.converter = class org.apache.kafka.connect.json.JsonConverter connect_1 | key.converter = class org.apache.kafka.connect.json.JsonConverter connect_1 | listeners = null connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | offset.flush.interval.ms = 60000 connect_1 | offset.flush.timeout.ms = 5000 connect_1 | offset.storage.partitions = 25 connect_1 | offset.storage.replication.factor = 1 connect_1 | offset.storage.topic = my_connect_offsets connect_1 | plugin.path = [/kafka/connect] connect_1 | rebalance.timeout.ms = 60000 connect_1 | receive.buffer.bytes = 32768 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 40000 connect_1 | response.http.headers.config = connect_1 | rest.advertised.host.name = connect connect_1 | rest.advertised.listener = null connect_1 | rest.advertised.port = 8083 connect_1 | rest.extension.classes = [] connect_1 | rest.host.name = 192.168.128.8 connect_1 | rest.port = 8083 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | scheduled.rebalance.max.delay.ms = 300000 connect_1 | security.protocol = PLAINTEXT connect_1 | send.buffer.bytes = 131072 connect_1 | session.timeout.ms = 10000 connect_1 | ssl.cipher.suites = null connect_1 | ssl.client.auth = none connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | status.storage.partitions = 5 connect_1 | status.storage.replication.factor = 1 connect_1 | status.storage.topic = my_connect_statuses connect_1 | task.shutdown.graceful.timeout.ms = 10000 connect_1 | topic.creation.enable = true connect_1 | topic.tracking.allow.reset = true connect_1 | topic.tracking.enable = true connect_1 | value.converter = class org.apache.kafka.connect.json.JsonConverter connect_1 | worker.sync.timeout.ms = 3000 connect_1 | worker.unsync.backoff.ms = 300000 connect_1 | [org.apache.kafka.connect.runtime.distributed.DistributedConfig] connect_1 | 2021-01-04 17:38:00,710 INFO || Worker configuration property 'internal.key.converter' is deprecated and may be removed in an upcoming release. The specified value 'org.apache.kafka.connect.json.JsonConverter' matches the default, so this property can be safely removed from the worker configuration. [org.apache.kafka.connect.runtime.WorkerConfig] connect_1 | 2021-01-04 17:38:00,711 INFO || Worker configuration property 'internal.value.converter' is deprecated and may be removed in an upcoming release. The specified value 'org.apache.kafka.connect.json.JsonConverter' matches the default, so this property can be safely removed from the worker configuration. [org.apache.kafka.connect.runtime.WorkerConfig] connect_1 | 2021-01-04 17:38:00,717 INFO || Creating Kafka admin client [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:00,730 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,896 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,896 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,897 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,897 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,897 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,897 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,897 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,897 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,897 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,898 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,898 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,898 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,898 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,898 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,898 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,898 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,898 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,899 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,899 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,899 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,901 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:00,902 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:00,903 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:00,903 INFO || Kafka startTimeMs: 1609781880902 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:01,552 INFO || Kafka cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:01,581 INFO || Logging initialized @5789ms to org.eclipse.jetty.util.log.Slf4jLog [org.eclipse.jetty.util.log] connect_1 | 2021-01-04 17:38:01,662 INFO || Added connector for http://192.168.128.8:8083 [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:01,663 INFO || Initializing REST server [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:01,673 INFO || jetty-9.4.24.v20191120; built: 2019-11-20T21:37:49.771Z; git: 363d5f2df3a8a28de40604320230664b9c793c16; jvm 11.0.8+10-LTS [org.eclipse.jetty.server.Server] connect_1 | 2021-01-04 17:38:01,710 INFO || Started http_192.168.128.88083@c6634d{HTTP/1.1,[http/1.1]}{192.168.128.8:8083} [org.eclipse.jetty.server.AbstractConnector] connect_1 | 2021-01-04 17:38:01,710 INFO || Started @5919ms [org.eclipse.jetty.server.Server] connect_1 | 2021-01-04 17:38:01,744 INFO || Advertised URI: http://connect:8083/ [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:01,745 INFO || REST server listening at http://192.168.128.8:8083/, advertising URL http://connect:8083/ [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:01,746 INFO || Advertised URI: http://connect:8083/ [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:01,746 INFO || REST admin endpoints at http://connect:8083/ [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:01,746 INFO || Advertised URI: http://connect:8083/ [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:01,748 INFO || Creating Kafka admin client [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:01,749 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,756 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,756 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,757 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,757 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,757 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,758 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,758 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,758 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,759 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,759 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,759 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,760 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,760 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,761 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,761 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,762 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,762 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,763 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,763 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,763 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,764 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,764 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:01,764 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:01,765 INFO || Kafka startTimeMs: 1609781881764 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:01,791 INFO || Kafka cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:01,806 INFO || Setting up None Policy for ConnectorClientConfigOverride. This will disallow any client configuration to be overridden [org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicy] connect_1 | 2021-01-04 17:38:01,815 INFO || Creating Kafka admin client [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:01,816 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,822 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,822 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,822 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,823 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,823 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,823 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,824 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,824 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,824 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,825 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,825 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,825 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,826 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,826 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,826 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,826 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,827 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,827 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,827 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,828 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,828 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:01,829 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:01,829 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:01,829 INFO || Kafka startTimeMs: 1609781881828 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:01,852 INFO || Kafka cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:01,864 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:01,864 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:01,864 INFO || Kafka startTimeMs: 1609781881864 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,044 INFO || JsonConverterConfig values: connect_1 | converter.type = key connect_1 | decimal.format = BASE64 connect_1 | schemas.cache.size = 1000 connect_1 | schemas.enable = false connect_1 | [org.apache.kafka.connect.json.JsonConverterConfig] connect_1 | 2021-01-04 17:38:02,046 INFO || JsonConverterConfig values: connect_1 | converter.type = value connect_1 | decimal.format = BASE64 connect_1 | schemas.cache.size = 1000 connect_1 | schemas.enable = false connect_1 | [org.apache.kafka.connect.json.JsonConverterConfig] connect_1 | 2021-01-04 17:38:02,046 INFO || Creating Kafka admin client [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:02,047 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,053 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,054 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,054 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,054 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,054 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,055 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,055 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,055 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,056 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,056 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,056 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,056 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,056 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,057 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,057 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,057 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,058 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,058 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,058 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,058 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,059 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,059 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,060 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,060 INFO || Kafka startTimeMs: 1609781882059 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,083 INFO || Kafka cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:02,099 INFO || Creating Kafka admin client [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:02,099 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,104 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,105 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,105 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,105 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,105 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,106 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,106 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,106 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,107 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,107 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,107 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,108 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,108 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,108 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,108 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,109 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,109 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,109 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,110 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,110 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,110 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,110 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,110 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,111 INFO || Kafka startTimeMs: 1609781882110 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,134 INFO || Kafka cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:02,143 INFO || Creating Kafka admin client [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:02,143 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,149 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,149 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,149 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,150 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,150 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,150 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,150 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,151 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,151 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,151 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,151 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,152 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,152 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,152 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,153 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,153 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,153 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,154 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,154 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,154 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,155 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,155 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,155 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,156 INFO || Kafka startTimeMs: 1609781882155 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,175 INFO || Kafka cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:02,197 INFO || Creating Kafka admin client [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:02,199 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,203 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,204 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,204 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,204 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,204 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,205 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,205 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,205 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,205 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,205 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,206 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,206 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,206 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,206 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,206 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,207 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,207 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,207 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,207 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,207 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,207 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,208 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,208 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,208 INFO || Kafka startTimeMs: 1609781882208 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,225 INFO || Kafka cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.connect.util.ConnectUtils] connect_1 | 2021-01-04 17:38:02,261 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,262 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,262 INFO || Kafka startTimeMs: 1609781882261 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,266 INFO || Kafka Connect distributed worker initialization took 5486ms [org.apache.kafka.connect.cli.ConnectDistributed] connect_1 | 2021-01-04 17:38:02,267 INFO || Kafka Connect starting [org.apache.kafka.connect.runtime.Connect] connect_1 | 2021-01-04 17:38:02,269 INFO || Initializing REST resources [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:02,270 INFO || [Worker clientId=connect-1, groupId=1] Herder starting [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:02,275 INFO || Worker starting [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:02,275 INFO || Starting KafkaOffsetBackingStore [org.apache.kafka.connect.storage.KafkaOffsetBackingStore] connect_1 | 2021-01-04 17:38:02,275 INFO || Starting KafkaBasedLog with topic my_connect_offsets [org.apache.kafka.connect.util.KafkaBasedLog] connect_1 | 2021-01-04 17:38:02,275 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,280 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,281 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,281 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,281 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,281 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,281 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,282 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,282 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,282 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,282 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,282 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,283 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,283 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,283 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,283 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,284 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,284 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,284 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,285 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,285 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,285 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,285 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,285 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:02,286 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,286 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,287 INFO || Kafka startTimeMs: 1609781882285 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:02,350 INFO || Adding admin resources to main listener [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:02,482 INFO || DefaultSessionIdManager workerName=node0 [org.eclipse.jetty.server.session] connect_1 | 2021-01-04 17:38:02,482 INFO || No SessionScavenger set, using defaults [org.eclipse.jetty.server.session] connect_1 | 2021-01-04 17:38:02,485 INFO || node0 Scavenging every 660000ms [org.eclipse.jetty.server.session] connect_1 | Jan 04, 2021 5:38:03 PM org.glassfish.jersey.internal.inject.Providers checkProviderRuntime connect_1 | WARNING: A provider org.apache.kafka.connect.runtime.rest.resources.LoggingResource registered in SERVER runtime does not implement any provider interfaces applicable in the SERVER runtime. Due to constraint configuration problems the provider org.apache.kafka.connect.runtime.rest.resources.LoggingResource will be ignored. connect_1 | Jan 04, 2021 5:38:03 PM org.glassfish.jersey.internal.inject.Providers checkProviderRuntime connect_1 | WARNING: A provider org.apache.kafka.connect.runtime.rest.resources.ConnectorPluginsResource registered in SERVER runtime does not implement any provider interfaces applicable in the SERVER runtime. Due to constraint configuration problems the provider org.apache.kafka.connect.runtime.rest.resources.ConnectorPluginsResource will be ignored. connect_1 | Jan 04, 2021 5:38:03 PM org.glassfish.jersey.internal.inject.Providers checkProviderRuntime connect_1 | WARNING: A provider org.apache.kafka.connect.runtime.rest.resources.RootResource registered in SERVER runtime does not implement any provider interfaces applicable in the SERVER runtime. Due to constraint configuration problems the provider org.apache.kafka.connect.runtime.rest.resources.RootResource will be ignored. connect_1 | Jan 04, 2021 5:38:03 PM org.glassfish.jersey.internal.inject.Providers checkProviderRuntime connect_1 | WARNING: A provider org.apache.kafka.connect.runtime.rest.resources.ConnectorsResource registered in SERVER runtime does not implement any provider interfaces applicable in the SERVER runtime. Due to constraint configuration problems the provider org.apache.kafka.connect.runtime.rest.resources.ConnectorsResource will be ignored. connect_1 | 2021-01-04 17:38:03,254 INFO || Created topic (name=my_connect_offsets, numPartitions=25, replicationFactor=1, replicasAssignments=null, configs={cleanup.policy=compact}) on brokers at kafka:9092 [org.apache.kafka.connect.util.TopicAdmin] connect_1 | 2021-01-04 17:38:03,268 INFO || ProducerConfig values: connect_1 | acks = -1 connect_1 | batch.size = 16384 connect_1 | bootstrap.servers = [kafka:9092] connect_1 | buffer.memory = 33554432 connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = producer-1 connect_1 | compression.type = none connect_1 | connections.max.idle.ms = 540000 connect_1 | delivery.timeout.ms = 2147483647 connect_1 | enable.idempotence = false connect_1 | interceptor.classes = [] connect_1 | internal.auto.downgrade.txn.commit = false connect_1 | key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer connect_1 | linger.ms = 0 connect_1 | max.block.ms = 60000 connect_1 | max.in.flight.requests.per.connection = 1 connect_1 | max.request.size = 1048576 connect_1 | metadata.max.age.ms = 300000 connect_1 | metadata.max.idle.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner connect_1 | receive.buffer.bytes = 32768 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | transaction.timeout.ms = 60000 connect_1 | transactional.id = null connect_1 | value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer connect_1 | [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,294 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,294 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,294 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,294 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,295 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,296 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,296 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,297 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,297 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,297 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,297 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,297 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,297 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,298 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,298 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,298 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,298 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,298 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,298 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,298 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,299 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,299 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,299 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,301 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,301 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,301 INFO || Kafka startTimeMs: 1609781883300 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,310 INFO || ConsumerConfig values: connect_1 | allow.auto.create.topics = true connect_1 | auto.commit.interval.ms = 5000 connect_1 | auto.offset.reset = earliest connect_1 | bootstrap.servers = [kafka:9092] connect_1 | check.crcs = true connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = consumer-1-1 connect_1 | client.rack = connect_1 | connections.max.idle.ms = 540000 connect_1 | default.api.timeout.ms = 60000 connect_1 | enable.auto.commit = false connect_1 | exclude.internal.topics = true connect_1 | fetch.max.bytes = 52428800 connect_1 | fetch.max.wait.ms = 500 connect_1 | fetch.min.bytes = 1 connect_1 | group.id = 1 connect_1 | group.instance.id = null connect_1 | heartbeat.interval.ms = 3000 connect_1 | interceptor.classes = [] connect_1 | internal.leave.group.on.close = true connect_1 | internal.throw.on.fetch.stable.offset.unsupported = false connect_1 | isolation.level = read_uncommitted connect_1 | key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer connect_1 | max.partition.fetch.bytes = 1048576 connect_1 | max.poll.interval.ms = 300000 connect_1 | max.poll.records = 500 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor] connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | session.timeout.ms = 10000 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer connect_1 | [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,318 INFO || [Producer clientId=producer-1] Cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.clients.Metadata] connect_1 | 2021-01-04 17:38:03,354 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,355 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,355 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,355 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,356 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,356 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,356 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,356 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,356 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,356 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,356 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,357 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,357 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,357 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,357 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,357 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,357 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,358 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,358 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,358 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,358 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,358 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,358 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,359 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,359 INFO || Kafka startTimeMs: 1609781883358 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,370 INFO || [Consumer clientId=consumer-1-1, groupId=1] Cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.clients.Metadata] connect_1 | Jan 04, 2021 5:38:03 PM org.glassfish.jersey.internal.Errors logErrors connect_1 | WARNING: The following warnings have been detected: WARNING: The (sub)resource method listLoggers in org.apache.kafka.connect.runtime.rest.resources.LoggingResource contains empty path annotation. connect_1 | WARNING: The (sub)resource method createConnector in org.apache.kafka.connect.runtime.rest.resources.ConnectorsResource contains empty path annotation. connect_1 | WARNING: The (sub)resource method listConnectors in org.apache.kafka.connect.runtime.rest.resources.ConnectorsResource contains empty path annotation. connect_1 | WARNING: The (sub)resource method listConnectorPlugins in org.apache.kafka.connect.runtime.rest.resources.ConnectorPluginsResource contains empty path annotation. connect_1 | WARNING: The (sub)resource method serverInfo in org.apache.kafka.connect.runtime.rest.resources.RootResource contains empty path annotation. connect_1 | connect_1 | 2021-01-04 17:38:03,385 INFO || Started o.e.j.s.ServletContextHandler@75d982d3{/,null,AVAILABLE} [org.eclipse.jetty.server.handler.ContextHandler] connect_1 | 2021-01-04 17:38:03,385 INFO || REST resources initialized; server is started and ready to handle requests [org.apache.kafka.connect.runtime.rest.RestServer] connect_1 | 2021-01-04 17:38:03,385 INFO || Kafka Connect started [org.apache.kafka.connect.runtime.Connect] connect_1 | 2021-01-04 17:38:03,421 INFO || [Consumer clientId=consumer-1-1, groupId=1] Subscribed to partition(s): my_connect_offsets-0, my_connect_offsets-5, my_connect_offsets-10, my_connect_offsets-20, my_connect_offsets-15, my_connect_offsets-9, my_connect_offsets-11, my_connect_offsets-4, my_connect_offsets-16, my_connect_offsets-17, my_connect_offsets-3, my_connect_offsets-24, my_connect_offsets-23, my_connect_offsets-13, my_connect_offsets-18, my_connect_offsets-22, my_connect_offsets-8, my_connect_offsets-2, my_connect_offsets-12, my_connect_offsets-19, my_connect_offsets-14, my_connect_offsets-1, my_connect_offsets-6, my_connect_offsets-7, my_connect_offsets-21 [org.apache.kafka.clients.consumer.KafkaConsumer] connect_1 | 2021-01-04 17:38:03,426 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-0 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,427 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-5 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,427 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-10 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,427 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-20 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,428 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-15 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,428 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-9 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,428 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-11 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,428 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-4 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,428 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-16 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,428 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-17 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-3 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-24 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-23 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-13 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-18 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-22 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-8 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-2 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-12 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-19 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,429 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-14 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,430 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-1 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,430 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-6 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,430 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-7 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,430 INFO || [Consumer clientId=consumer-1-1, groupId=1] Seeking to EARLIEST offset of partition my_connect_offsets-21 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,488 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-2 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,490 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-4 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,490 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-6 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,490 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-8 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,490 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-0 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,490 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-18 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,491 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-20 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,491 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-22 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,491 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-24 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,491 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-10 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,491 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-12 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,491 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-14 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,492 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-16 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,492 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-3 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,492 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-5 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,492 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-7 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,492 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-9 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,493 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-1 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,493 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-19 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,493 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-21 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,493 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-23 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,493 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-11 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,494 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-13 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,494 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-15 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,494 INFO || [Consumer clientId=consumer-1-1, groupId=1] Resetting offset for partition my_connect_offsets-17 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,496 INFO || Finished reading KafkaBasedLog for topic my_connect_offsets [org.apache.kafka.connect.util.KafkaBasedLog] connect_1 | 2021-01-04 17:38:03,496 INFO || Started KafkaBasedLog for topic my_connect_offsets [org.apache.kafka.connect.util.KafkaBasedLog] connect_1 | 2021-01-04 17:38:03,496 INFO || Finished reading offsets topic and starting KafkaOffsetBackingStore [org.apache.kafka.connect.storage.KafkaOffsetBackingStore] connect_1 | 2021-01-04 17:38:03,503 INFO || Worker started [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:03,503 INFO || Starting KafkaBasedLog with topic my_connect_statuses [org.apache.kafka.connect.util.KafkaBasedLog] connect_1 | 2021-01-04 17:38:03,508 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,514 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,514 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,514 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,514 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,514 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,514 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,514 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,515 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,515 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,515 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,515 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,515 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,515 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,515 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,515 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,516 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,516 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,516 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,516 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,516 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,516 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,516 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,516 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,517 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,517 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,517 INFO || Kafka startTimeMs: 1609781883517 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,676 INFO || Created topic (name=my_connect_statuses, numPartitions=5, replicationFactor=1, replicasAssignments=null, configs={cleanup.policy=compact}) on brokers at kafka:9092 [org.apache.kafka.connect.util.TopicAdmin] connect_1 | 2021-01-04 17:38:03,680 INFO || ProducerConfig values: connect_1 | acks = -1 connect_1 | batch.size = 16384 connect_1 | bootstrap.servers = [kafka:9092] connect_1 | buffer.memory = 33554432 connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = producer-2 connect_1 | compression.type = none connect_1 | connections.max.idle.ms = 540000 connect_1 | delivery.timeout.ms = 120000 connect_1 | enable.idempotence = false connect_1 | interceptor.classes = [] connect_1 | internal.auto.downgrade.txn.commit = false connect_1 | key.serializer = class org.apache.kafka.common.serialization.StringSerializer connect_1 | linger.ms = 0 connect_1 | max.block.ms = 60000 connect_1 | max.in.flight.requests.per.connection = 1 connect_1 | max.request.size = 1048576 connect_1 | metadata.max.age.ms = 300000 connect_1 | metadata.max.idle.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner connect_1 | receive.buffer.bytes = 32768 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 0 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | transaction.timeout.ms = 60000 connect_1 | transactional.id = null connect_1 | value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer connect_1 | [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,688 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,688 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,688 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,688 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,688 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,688 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,688 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,689 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,689 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,689 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,689 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,689 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,690 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,690 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,690 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,690 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,690 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,691 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,693 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,693 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,693 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,693 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,693 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,693 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,694 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,694 INFO || Kafka startTimeMs: 1609781883693 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,696 INFO || ConsumerConfig values: connect_1 | allow.auto.create.topics = true connect_1 | auto.commit.interval.ms = 5000 connect_1 | auto.offset.reset = earliest connect_1 | bootstrap.servers = [kafka:9092] connect_1 | check.crcs = true connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = consumer-1-2 connect_1 | client.rack = connect_1 | connections.max.idle.ms = 540000 connect_1 | default.api.timeout.ms = 60000 connect_1 | enable.auto.commit = false connect_1 | exclude.internal.topics = true connect_1 | fetch.max.bytes = 52428800 connect_1 | fetch.max.wait.ms = 500 connect_1 | fetch.min.bytes = 1 connect_1 | group.id = 1 connect_1 | group.instance.id = null connect_1 | heartbeat.interval.ms = 3000 connect_1 | interceptor.classes = [] connect_1 | internal.leave.group.on.close = true connect_1 | internal.throw.on.fetch.stable.offset.unsupported = false connect_1 | isolation.level = read_uncommitted connect_1 | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer connect_1 | max.partition.fetch.bytes = 1048576 connect_1 | max.poll.interval.ms = 300000 connect_1 | max.poll.records = 500 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor] connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | session.timeout.ms = 10000 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer connect_1 | [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,702 INFO || [Producer clientId=producer-2] Cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.clients.Metadata] connect_1 | 2021-01-04 17:38:03,708 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,709 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,709 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,709 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,709 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,709 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,710 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,710 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,710 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,710 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,710 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,710 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,710 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,710 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,710 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,711 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,711 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,711 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,711 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,711 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,711 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,712 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,712 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,712 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,712 INFO || Kafka startTimeMs: 1609781883712 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,721 INFO || [Consumer clientId=consumer-1-2, groupId=1] Cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.clients.Metadata] connect_1 | 2021-01-04 17:38:03,739 INFO || [Consumer clientId=consumer-1-2, groupId=1] Subscribed to partition(s): my_connect_statuses-0, my_connect_statuses-4, my_connect_statuses-1, my_connect_statuses-2, my_connect_statuses-3 [org.apache.kafka.clients.consumer.KafkaConsumer] connect_1 | 2021-01-04 17:38:03,740 INFO || [Consumer clientId=consumer-1-2, groupId=1] Seeking to EARLIEST offset of partition my_connect_statuses-0 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,740 INFO || [Consumer clientId=consumer-1-2, groupId=1] Seeking to EARLIEST offset of partition my_connect_statuses-4 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,740 INFO || [Consumer clientId=consumer-1-2, groupId=1] Seeking to EARLIEST offset of partition my_connect_statuses-1 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,740 INFO || [Consumer clientId=consumer-1-2, groupId=1] Seeking to EARLIEST offset of partition my_connect_statuses-2 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,740 INFO || [Consumer clientId=consumer-1-2, groupId=1] Seeking to EARLIEST offset of partition my_connect_statuses-3 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,763 INFO || [Consumer clientId=consumer-1-2, groupId=1] Resetting offset for partition my_connect_statuses-0 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,763 INFO || [Consumer clientId=consumer-1-2, groupId=1] Resetting offset for partition my_connect_statuses-1 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,763 INFO || [Consumer clientId=consumer-1-2, groupId=1] Resetting offset for partition my_connect_statuses-2 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,763 INFO || [Consumer clientId=consumer-1-2, groupId=1] Resetting offset for partition my_connect_statuses-3 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,764 INFO || [Consumer clientId=consumer-1-2, groupId=1] Resetting offset for partition my_connect_statuses-4 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,764 INFO || Finished reading KafkaBasedLog for topic my_connect_statuses [org.apache.kafka.connect.util.KafkaBasedLog] connect_1 | 2021-01-04 17:38:03,765 INFO || Started KafkaBasedLog for topic my_connect_statuses [org.apache.kafka.connect.util.KafkaBasedLog] connect_1 | 2021-01-04 17:38:03,765 INFO || Starting KafkaConfigBackingStore [org.apache.kafka.connect.storage.KafkaConfigBackingStore] connect_1 | 2021-01-04 17:38:03,765 INFO || Starting KafkaBasedLog with topic my_connect_configs [org.apache.kafka.connect.util.KafkaBasedLog] connect_1 | 2021-01-04 17:38:03,766 INFO || AdminClientConfig values: connect_1 | bootstrap.servers = [kafka:9092] connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connect_1 | connections.max.idle.ms = 300000 connect_1 | default.api.timeout.ms = 60000 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,771 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,771 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,771 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,771 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,771 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,774 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,774 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,774 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,774 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,775 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,775 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,775 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,775 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,775 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,776 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,776 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,776 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,776 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,776 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,776 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,776 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,777 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,777 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.admin.AdminClientConfig] connect_1 | 2021-01-04 17:38:03,777 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,778 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,780 INFO || Kafka startTimeMs: 1609781883777 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,842 INFO || Created topic (name=my_connect_configs, numPartitions=1, replicationFactor=1, replicasAssignments=null, configs={cleanup.policy=compact}) on brokers at kafka:9092 [org.apache.kafka.connect.util.TopicAdmin] connect_1 | 2021-01-04 17:38:03,845 INFO || ProducerConfig values: connect_1 | acks = -1 connect_1 | batch.size = 16384 connect_1 | bootstrap.servers = [kafka:9092] connect_1 | buffer.memory = 33554432 connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = producer-3 connect_1 | compression.type = none connect_1 | connections.max.idle.ms = 540000 connect_1 | delivery.timeout.ms = 2147483647 connect_1 | enable.idempotence = false connect_1 | interceptor.classes = [] connect_1 | internal.auto.downgrade.txn.commit = false connect_1 | key.serializer = class org.apache.kafka.common.serialization.StringSerializer connect_1 | linger.ms = 0 connect_1 | max.block.ms = 60000 connect_1 | max.in.flight.requests.per.connection = 1 connect_1 | max.request.size = 1048576 connect_1 | metadata.max.age.ms = 300000 connect_1 | metadata.max.idle.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner connect_1 | receive.buffer.bytes = 32768 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | transaction.timeout.ms = 60000 connect_1 | transactional.id = null connect_1 | value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer connect_1 | [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,852 WARN || The configuration 'group.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,853 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,853 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,853 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,854 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,854 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,854 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,854 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,854 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,855 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,855 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,856 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,856 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,856 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,856 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,857 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,857 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,857 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,858 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,858 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,858 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,859 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,859 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:03,859 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,860 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,860 INFO || Kafka startTimeMs: 1609781883859 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,861 INFO || ConsumerConfig values: connect_1 | allow.auto.create.topics = true connect_1 | auto.commit.interval.ms = 5000 connect_1 | auto.offset.reset = earliest connect_1 | bootstrap.servers = [kafka:9092] connect_1 | check.crcs = true connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = consumer-1-3 connect_1 | client.rack = connect_1 | connections.max.idle.ms = 540000 connect_1 | default.api.timeout.ms = 60000 connect_1 | enable.auto.commit = false connect_1 | exclude.internal.topics = true connect_1 | fetch.max.bytes = 52428800 connect_1 | fetch.max.wait.ms = 500 connect_1 | fetch.min.bytes = 1 connect_1 | group.id = 1 connect_1 | group.instance.id = null connect_1 | heartbeat.interval.ms = 3000 connect_1 | interceptor.classes = [] connect_1 | internal.leave.group.on.close = true connect_1 | internal.throw.on.fetch.stable.offset.unsupported = false connect_1 | isolation.level = read_uncommitted connect_1 | key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer connect_1 | max.partition.fetch.bytes = 1048576 connect_1 | max.poll.interval.ms = 300000 connect_1 | max.poll.records = 500 connect_1 | metadata.max.age.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor] connect_1 | receive.buffer.bytes = 65536 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 30000 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | session.timeout.ms = 10000 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer connect_1 | [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,865 INFO || [Producer clientId=producer-3] Cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.clients.Metadata] connect_1 | 2021-01-04 17:38:03,869 WARN || The configuration 'rest.advertised.port' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,870 WARN || The configuration 'task.shutdown.graceful.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,870 WARN || The configuration 'plugin.path' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,870 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,871 WARN || The configuration 'status.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,871 WARN || The configuration 'offset.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,871 WARN || The configuration 'value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,871 WARN || The configuration 'key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,871 WARN || The configuration 'config.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,871 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,871 WARN || The configuration 'rest.advertised.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,871 WARN || The configuration 'status.storage.topic' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,871 WARN || The configuration 'rest.host.name' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 WARN || The configuration 'offset.flush.timeout.ms' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 WARN || The configuration 'config.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 WARN || The configuration 'offset.flush.interval.ms' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 WARN || The configuration 'rest.port' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 WARN || The configuration 'key.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 WARN || The configuration 'internal.key.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 WARN || The configuration 'value.converter.schemas.enable' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 WARN || The configuration 'internal.value.converter' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 WARN || The configuration 'offset.storage.replication.factor' was supplied but isn't a known config. [org.apache.kafka.clients.consumer.ConsumerConfig] connect_1 | 2021-01-04 17:38:03,872 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,872 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,872 INFO || Kafka startTimeMs: 1609781883872 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:03,878 INFO || [Consumer clientId=consumer-1-3, groupId=1] Cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.clients.Metadata] connect_1 | 2021-01-04 17:38:03,891 INFO || [Consumer clientId=consumer-1-3, groupId=1] Subscribed to partition(s): my_connect_configs-0 [org.apache.kafka.clients.consumer.KafkaConsumer] connect_1 | 2021-01-04 17:38:03,891 INFO || [Consumer clientId=consumer-1-3, groupId=1] Seeking to EARLIEST offset of partition my_connect_configs-0 [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,906 INFO || [Consumer clientId=consumer-1-3, groupId=1] Resetting offset for partition my_connect_configs-0 to offset 0. [org.apache.kafka.clients.consumer.internals.SubscriptionState] connect_1 | 2021-01-04 17:38:03,906 INFO || Finished reading KafkaBasedLog for topic my_connect_configs [org.apache.kafka.connect.util.KafkaBasedLog] connect_1 | 2021-01-04 17:38:03,906 INFO || Started KafkaBasedLog for topic my_connect_configs [org.apache.kafka.connect.util.KafkaBasedLog] connect_1 | 2021-01-04 17:38:03,906 INFO || Started KafkaConfigBackingStore [org.apache.kafka.connect.storage.KafkaConfigBackingStore] connect_1 | 2021-01-04 17:38:03,906 INFO || [Worker clientId=connect-1, groupId=1] Herder started [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:03,923 INFO || [Worker clientId=connect-1, groupId=1] Cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.clients.Metadata] connect_1 | 2021-01-04 17:38:04,763 INFO || [Worker clientId=connect-1, groupId=1] Discovered group coordinator 192.168.128.6:9092 (id: 2147483646 rack: null) [org.apache.kafka.clients.consumer.internals.AbstractCoordinator] connect_1 | 2021-01-04 17:38:04,767 INFO || [Worker clientId=connect-1, groupId=1] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] connect_1 | 2021-01-04 17:38:04,767 INFO || [Worker clientId=connect-1, groupId=1] (Re-)joining group [org.apache.kafka.clients.consumer.internals.AbstractCoordinator] connect_1 | 2021-01-04 17:38:04,803 INFO || [Worker clientId=connect-1, groupId=1] Join group failed with org.apache.kafka.common.errors.MemberIdRequiredException: The group member needs to have a valid member id before actually entering a consumer group. [org.apache.kafka.clients.consumer.internals.AbstractCoordinator] connect_1 | 2021-01-04 17:38:04,803 INFO || [Worker clientId=connect-1, groupId=1] (Re-)joining group [org.apache.kafka.clients.consumer.internals.AbstractCoordinator] connect_1 | 2021-01-04 17:38:04,972 INFO || [Worker clientId=connect-1, groupId=1] Successfully joined group with generation 1 [org.apache.kafka.clients.consumer.internals.AbstractCoordinator] connect_1 | 2021-01-04 17:38:04,972 INFO || [Worker clientId=connect-1, groupId=1] Joined group at generation 1 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-b0909f7d-06f8-4461-afe3-411392e96f0c', leaderUrl='http://connect:8083/', offset=-1, connectorIds=[], taskIds=[], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:04,974 INFO || [Worker clientId=connect-1, groupId=1] Starting connectors and tasks using config offset -1 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:04,975 INFO || [Worker clientId=connect-1, groupId=1] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:05,075 INFO || [Worker clientId=connect-1, groupId=1] Session key updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:19,307 ERROR || Uncaught exception in REST call to /connectors/ [org.apache.kafka.connect.runtime.rest.errors.ConnectExceptionMapper] connect_1 | com.fasterxml.jackson.core.JsonParseException: Unrecognized token 'Here': was expecting (JSON String, Number, Array, Object or token 'null', 'true' or 'false') connect_1 | at [Source: (org.glassfish.jersey.message.internal.ReaderInterceptorExecutor$UnCloseableInputStream); line: 1, column: 6] connect_1 | at com.fasterxml.jackson.core.JsonParser._constructError(JsonParser.java:1840) connect_1 | at com.fasterxml.jackson.core.base.ParserMinimalBase._reportError(ParserMinimalBase.java:722) connect_1 | at com.fasterxml.jackson.core.json.UTF8StreamJsonParser._reportInvalidToken(UTF8StreamJsonParser.java:3556) connect_1 | at com.fasterxml.jackson.core.json.UTF8StreamJsonParser._handleUnexpectedValue(UTF8StreamJsonParser.java:2651) connect_1 | at com.fasterxml.jackson.core.json.UTF8StreamJsonParser._nextTokenNotInObject(UTF8StreamJsonParser.java:856) connect_1 | at com.fasterxml.jackson.core.json.UTF8StreamJsonParser.nextToken(UTF8StreamJsonParser.java:753) connect_1 | at com.fasterxml.jackson.jaxrs.base.ProviderBase.readFrom(ProviderBase.java:773) connect_1 | at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor$TerminalReaderInterceptor.invokeReadFrom(ReaderInterceptorExecutor.java:233) connect_1 | at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor$TerminalReaderInterceptor.aroundReadFrom(ReaderInterceptorExecutor.java:212) connect_1 | at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor.proceed(ReaderInterceptorExecutor.java:132) connect_1 | at org.glassfish.jersey.server.internal.MappableExceptionWrapperInterceptor.aroundReadFrom(MappableExceptionWrapperInterceptor.java:49) connect_1 | at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor.proceed(ReaderInterceptorExecutor.java:132) connect_1 | at org.glassfish.jersey.message.internal.MessageBodyFactory.readFrom(MessageBodyFactory.java:1067) connect_1 | at org.glassfish.jersey.message.internal.InboundMessageContext.readEntity(InboundMessageContext.java:850) connect_1 | at org.glassfish.jersey.server.ContainerRequest.readEntity(ContainerRequest.java:247) connect_1 | at org.glassfish.jersey.server.internal.inject.EntityParamValueParamProvider$EntityValueSupplier.apply(EntityParamValueParamProvider.java:73) connect_1 | at org.glassfish.jersey.server.internal.inject.EntityParamValueParamProvider$EntityValueSupplier.apply(EntityParamValueParamProvider.java:56) connect_1 | at org.glassfish.jersey.server.spi.internal.ParamValueFactoryWithSource.apply(ParamValueFactoryWithSource.java:50) connect_1 | at org.glassfish.jersey.server.spi.internal.ParameterValueHelper.getParameterValues(ParameterValueHelper.java:68) connect_1 | at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$AbstractMethodParamInvoker.getParamValues(JavaResourceMethodDispatcherProvider.java:109) connect_1 | at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$ResponseOutInvoker.doDispatch(JavaResourceMethodDispatcherProvider.java:176) connect_1 | at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.dispatch(AbstractJavaResourceMethodDispatcher.java:79) connect_1 | at org.glassfish.jersey.server.model.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:469) connect_1 | at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:391) connect_1 | at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:80) connect_1 | at org.glassfish.jersey.server.ServerRuntime$1.run(ServerRuntime.java:253) connect_1 | at org.glassfish.jersey.internal.Errors$1.call(Errors.java:248) connect_1 | at org.glassfish.jersey.internal.Errors$1.call(Errors.java:244) connect_1 | at org.glassfish.jersey.internal.Errors.process(Errors.java:292) connect_1 | at org.glassfish.jersey.internal.Errors.process(Errors.java:274) connect_1 | at org.glassfish.jersey.internal.Errors.process(Errors.java:244) connect_1 | at org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:265) connect_1 | at org.glassfish.jersey.server.ServerRuntime.process(ServerRuntime.java:232) connect_1 | at org.glassfish.jersey.server.ApplicationHandler.handle(ApplicationHandler.java:679) connect_1 | at org.glassfish.jersey.servlet.WebComponent.serviceImpl(WebComponent.java:392) connect_1 | at org.glassfish.jersey.servlet.WebComponent.service(WebComponent.java:346) connect_1 | at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:365) connect_1 | at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:318) connect_1 | at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:205) connect_1 | at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:760) connect_1 | at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:547) connect_1 | at org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:233) connect_1 | at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:1607) connect_1 | at org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:233) connect_1 | at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1297) connect_1 | at org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:188) connect_1 | at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:485) connect_1 | at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:1577) connect_1 | at org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:186) connect_1 | at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1212) connect_1 | at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) connect_1 | at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:221) connect_1 | at org.eclipse.jetty.server.handler.StatisticsHandler.handle(StatisticsHandler.java:173) connect_1 | at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127) connect_1 | at org.eclipse.jetty.server.Server.handle(Server.java:500) connect_1 | at org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:383) connect_1 | at org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:547) connect_1 | at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:375) connect_1 | at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:270) connect_1 | at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311) connect_1 | at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:103) connect_1 | at org.eclipse.jetty.io.ChannelEndPoint$2.run(ChannelEndPoint.java:117) connect_1 | at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:806) connect_1 | at org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:938) connect_1 | at java.base/java.lang.Thread.run(Thread.java:834) connect_1 | 2021-01-04 17:38:42,250 WARN || Using configuration property "schema.blacklist" is deprecated and will be removed in future versions. Please use "schema.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,251 WARN || Using configuration property "table.blacklist" is deprecated and will be removed in future versions. Please use "table.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,252 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,265 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,265 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,265 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,266 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,266 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,266 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,266 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,266 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,266 WARN || Using configuration property "table.blacklist" is deprecated and will be removed in future versions. Please use "table.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,266 WARN || Using configuration property "table.blacklist" is deprecated and will be removed in future versions. Please use "table.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,267 WARN || Using configuration property "schema.blacklist" is deprecated and will be removed in future versions. Please use "schema.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,267 WARN || Using configuration property "schema.blacklist" is deprecated and will be removed in future versions. Please use "schema.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:42,401 INFO || Successfully tested connection for jdbc:postgresql://rentsite-db:5432/rentsite with user 'rentsite' [io.debezium.connector.postgresql.PostgresConnector] connect_1 | 2021-01-04 17:38:42,407 INFO || AbstractConfig values: connect_1 | [org.apache.kafka.common.config.AbstractConfig] connect_1 | 2021-01-04 17:38:42,421 INFO || [Worker clientId=connect-1, groupId=1] Connector rentsite-connector config updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:42,930 INFO || [Worker clientId=connect-1, groupId=1] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] connect_1 | 2021-01-04 17:38:42,930 INFO || [Worker clientId=connect-1, groupId=1] (Re-)joining group [org.apache.kafka.clients.consumer.internals.AbstractCoordinator] connect_1 | 2021-01-04 17:38:42,952 INFO || [Worker clientId=connect-1, groupId=1] Successfully joined group with generation 2 [org.apache.kafka.clients.consumer.internals.AbstractCoordinator] connect_1 | 2021-01-04 17:38:42,954 INFO || [Worker clientId=connect-1, groupId=1] Joined group at generation 2 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-b0909f7d-06f8-4461-afe3-411392e96f0c', leaderUrl='http://connect:8083/', offset=2, connectorIds=[rentsite-connector], taskIds=[], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:42,955 INFO || [Worker clientId=connect-1, groupId=1] Starting connectors and tasks using config offset 2 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:42,967 INFO || [Worker clientId=connect-1, groupId=1] Starting connector rentsite-connector [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:42,973 INFO || Creating connector rentsite-connector of type io.debezium.connector.postgresql.PostgresConnector [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:42,982 INFO || SourceConnectorConfig values: connect_1 | config.action.reload = restart connect_1 | connector.class = io.debezium.connector.postgresql.PostgresConnector connect_1 | errors.log.enable = false connect_1 | errors.log.include.messages = false connect_1 | errors.retry.delay.max.ms = 60000 connect_1 | errors.retry.timeout = 0 connect_1 | errors.tolerance = none connect_1 | header.converter = null connect_1 | key.converter = null connect_1 | name = rentsite-connector connect_1 | predicates = [] connect_1 | tasks.max = 1 connect_1 | topic.creation.groups = [] connect_1 | transforms = [] connect_1 | value.converter = null connect_1 | [org.apache.kafka.connect.runtime.SourceConnectorConfig] connect_1 | 2021-01-04 17:38:42,983 INFO || EnrichedConnectorConfig values: connect_1 | config.action.reload = restart connect_1 | connector.class = io.debezium.connector.postgresql.PostgresConnector connect_1 | errors.log.enable = false connect_1 | errors.log.include.messages = false connect_1 | errors.retry.delay.max.ms = 60000 connect_1 | errors.retry.timeout = 0 connect_1 | errors.tolerance = none connect_1 | header.converter = null connect_1 | key.converter = null connect_1 | name = rentsite-connector connect_1 | predicates = [] connect_1 | tasks.max = 1 connect_1 | topic.creation.groups = [] connect_1 | transforms = [] connect_1 | value.converter = null connect_1 | [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] connect_1 | 2021-01-04 17:38:42,990 INFO || Instantiated connector rentsite-connector with version 1.3.1.Final of type class io.debezium.connector.postgresql.PostgresConnector [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:42,991 INFO || Finished creating connector rentsite-connector [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:42,994 INFO || [Worker clientId=connect-1, groupId=1] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:43,013 INFO || SourceConnectorConfig values: connect_1 | config.action.reload = restart connect_1 | connector.class = io.debezium.connector.postgresql.PostgresConnector connect_1 | errors.log.enable = false connect_1 | errors.log.include.messages = false connect_1 | errors.retry.delay.max.ms = 60000 connect_1 | errors.retry.timeout = 0 connect_1 | errors.tolerance = none connect_1 | header.converter = null connect_1 | key.converter = null connect_1 | name = rentsite-connector connect_1 | predicates = [] connect_1 | tasks.max = 1 connect_1 | topic.creation.groups = [] connect_1 | transforms = [] connect_1 | value.converter = null connect_1 | [org.apache.kafka.connect.runtime.SourceConnectorConfig] connect_1 | 2021-01-04 17:38:43,015 INFO || EnrichedConnectorConfig values: connect_1 | config.action.reload = restart connect_1 | connector.class = io.debezium.connector.postgresql.PostgresConnector connect_1 | errors.log.enable = false connect_1 | errors.log.include.messages = false connect_1 | errors.retry.delay.max.ms = 60000 connect_1 | errors.retry.timeout = 0 connect_1 | errors.tolerance = none connect_1 | header.converter = null connect_1 | key.converter = null connect_1 | name = rentsite-connector connect_1 | predicates = [] connect_1 | tasks.max = 1 connect_1 | topic.creation.groups = [] connect_1 | transforms = [] connect_1 | value.converter = null connect_1 | [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] connect_1 | 2021-01-04 17:38:43,957 INFO || [Worker clientId=connect-1, groupId=1] Tasks [rentsite-connector-0] configs updated [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:43,959 INFO || [Worker clientId=connect-1, groupId=1] Handling task config update by restarting tasks [] [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:43,960 INFO || [Worker clientId=connect-1, groupId=1] Rebalance started [org.apache.kafka.connect.runtime.distributed.WorkerCoordinator] connect_1 | 2021-01-04 17:38:43,960 INFO || [Worker clientId=connect-1, groupId=1] (Re-)joining group [org.apache.kafka.clients.consumer.internals.AbstractCoordinator] connect_1 | 2021-01-04 17:38:43,969 INFO || [Worker clientId=connect-1, groupId=1] Successfully joined group with generation 3 [org.apache.kafka.clients.consumer.internals.AbstractCoordinator] connect_1 | 2021-01-04 17:38:43,970 INFO || [Worker clientId=connect-1, groupId=1] Joined group at generation 3 with protocol version 2 and got assignment: Assignment{error=0, leader='connect-1-b0909f7d-06f8-4461-afe3-411392e96f0c', leaderUrl='http://connect:8083/', offset=4, connectorIds=[rentsite-connector], taskIds=[rentsite-connector-0], revokedConnectorIds=[], revokedTaskIds=[], delay=0} with rebalance delay: 0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:43,972 INFO || [Worker clientId=connect-1, groupId=1] Starting connectors and tasks using config offset 4 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:43,973 INFO || [Worker clientId=connect-1, groupId=1] Starting task rentsite-connector-0 [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:43,974 INFO || Creating task rentsite-connector-0 [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:43,976 INFO || ConnectorConfig values: connect_1 | config.action.reload = restart connect_1 | connector.class = io.debezium.connector.postgresql.PostgresConnector connect_1 | errors.log.enable = false connect_1 | errors.log.include.messages = false connect_1 | errors.retry.delay.max.ms = 60000 connect_1 | errors.retry.timeout = 0 connect_1 | errors.tolerance = none connect_1 | header.converter = null connect_1 | key.converter = null connect_1 | name = rentsite-connector connect_1 | predicates = [] connect_1 | tasks.max = 1 connect_1 | transforms = [] connect_1 | value.converter = null connect_1 | [org.apache.kafka.connect.runtime.ConnectorConfig] connect_1 | 2021-01-04 17:38:43,976 INFO || EnrichedConnectorConfig values: connect_1 | config.action.reload = restart connect_1 | connector.class = io.debezium.connector.postgresql.PostgresConnector connect_1 | errors.log.enable = false connect_1 | errors.log.include.messages = false connect_1 | errors.retry.delay.max.ms = 60000 connect_1 | errors.retry.timeout = 0 connect_1 | errors.tolerance = none connect_1 | header.converter = null connect_1 | key.converter = null connect_1 | name = rentsite-connector connect_1 | predicates = [] connect_1 | tasks.max = 1 connect_1 | transforms = [] connect_1 | value.converter = null connect_1 | [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] connect_1 | 2021-01-04 17:38:43,980 INFO || TaskConfig values: connect_1 | task.class = class io.debezium.connector.postgresql.PostgresConnectorTask connect_1 | [org.apache.kafka.connect.runtime.TaskConfig] connect_1 | 2021-01-04 17:38:43,981 INFO || Instantiated task rentsite-connector-0 with version 1.3.1.Final of type io.debezium.connector.postgresql.PostgresConnectorTask [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:43,983 INFO || JsonConverterConfig values: connect_1 | converter.type = key connect_1 | decimal.format = BASE64 connect_1 | schemas.cache.size = 1000 connect_1 | schemas.enable = true connect_1 | [org.apache.kafka.connect.json.JsonConverterConfig] connect_1 | 2021-01-04 17:38:43,984 INFO || Set up the key converter class org.apache.kafka.connect.json.JsonConverter for task rentsite-connector-0 using the worker config [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:43,984 INFO || JsonConverterConfig values: connect_1 | converter.type = value connect_1 | decimal.format = BASE64 connect_1 | schemas.cache.size = 1000 connect_1 | schemas.enable = true connect_1 | [org.apache.kafka.connect.json.JsonConverterConfig] connect_1 | 2021-01-04 17:38:43,984 INFO || Set up the value converter class org.apache.kafka.connect.json.JsonConverter for task rentsite-connector-0 using the worker config [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:43,985 INFO || Set up the header converter class org.apache.kafka.connect.storage.SimpleHeaderConverter for task rentsite-connector-0 using the worker config [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:43,990 INFO || SourceConnectorConfig values: connect_1 | config.action.reload = restart connect_1 | connector.class = io.debezium.connector.postgresql.PostgresConnector connect_1 | errors.log.enable = false connect_1 | errors.log.include.messages = false connect_1 | errors.retry.delay.max.ms = 60000 connect_1 | errors.retry.timeout = 0 connect_1 | errors.tolerance = none connect_1 | header.converter = null connect_1 | key.converter = null connect_1 | name = rentsite-connector connect_1 | predicates = [] connect_1 | tasks.max = 1 connect_1 | topic.creation.groups = [] connect_1 | transforms = [] connect_1 | value.converter = null connect_1 | [org.apache.kafka.connect.runtime.SourceConnectorConfig] connect_1 | 2021-01-04 17:38:43,990 INFO || EnrichedConnectorConfig values: connect_1 | config.action.reload = restart connect_1 | connector.class = io.debezium.connector.postgresql.PostgresConnector connect_1 | errors.log.enable = false connect_1 | errors.log.include.messages = false connect_1 | errors.retry.delay.max.ms = 60000 connect_1 | errors.retry.timeout = 0 connect_1 | errors.tolerance = none connect_1 | header.converter = null connect_1 | key.converter = null connect_1 | name = rentsite-connector connect_1 | predicates = [] connect_1 | tasks.max = 1 connect_1 | topic.creation.groups = [] connect_1 | transforms = [] connect_1 | value.converter = null connect_1 | [org.apache.kafka.connect.runtime.ConnectorConfig$EnrichedConnectorConfig] connect_1 | 2021-01-04 17:38:43,992 INFO || Initializing: org.apache.kafka.connect.runtime.TransformationChain{} [org.apache.kafka.connect.runtime.Worker] connect_1 | 2021-01-04 17:38:43,994 INFO || ProducerConfig values: connect_1 | acks = -1 connect_1 | batch.size = 16384 connect_1 | bootstrap.servers = [kafka:9092] connect_1 | buffer.memory = 33554432 connect_1 | client.dns.lookup = use_all_dns_ips connect_1 | client.id = connector-producer-rentsite-connector-0 connect_1 | compression.type = none connect_1 | connections.max.idle.ms = 540000 connect_1 | delivery.timeout.ms = 2147483647 connect_1 | enable.idempotence = false connect_1 | interceptor.classes = [] connect_1 | internal.auto.downgrade.txn.commit = false connect_1 | key.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer connect_1 | linger.ms = 0 connect_1 | max.block.ms = 9223372036854775807 connect_1 | max.in.flight.requests.per.connection = 1 connect_1 | max.request.size = 1048576 connect_1 | metadata.max.age.ms = 300000 connect_1 | metadata.max.idle.ms = 300000 connect_1 | metric.reporters = [] connect_1 | metrics.num.samples = 2 connect_1 | metrics.recording.level = INFO connect_1 | metrics.sample.window.ms = 30000 connect_1 | partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner connect_1 | receive.buffer.bytes = 32768 connect_1 | reconnect.backoff.max.ms = 1000 connect_1 | reconnect.backoff.ms = 50 connect_1 | request.timeout.ms = 2147483647 connect_1 | retries = 2147483647 connect_1 | retry.backoff.ms = 100 connect_1 | sasl.client.callback.handler.class = null connect_1 | sasl.jaas.config = null connect_1 | sasl.kerberos.kinit.cmd = /usr/bin/kinit connect_1 | sasl.kerberos.min.time.before.relogin = 60000 connect_1 | sasl.kerberos.service.name = null connect_1 | sasl.kerberos.ticket.renew.jitter = 0.05 connect_1 | sasl.kerberos.ticket.renew.window.factor = 0.8 connect_1 | sasl.login.callback.handler.class = null connect_1 | sasl.login.class = null connect_1 | sasl.login.refresh.buffer.seconds = 300 connect_1 | sasl.login.refresh.min.period.seconds = 60 connect_1 | sasl.login.refresh.window.factor = 0.8 connect_1 | sasl.login.refresh.window.jitter = 0.05 connect_1 | sasl.mechanism = GSSAPI connect_1 | security.protocol = PLAINTEXT connect_1 | security.providers = null connect_1 | send.buffer.bytes = 131072 connect_1 | ssl.cipher.suites = null connect_1 | ssl.enabled.protocols = [TLSv1.2, TLSv1.3] connect_1 | ssl.endpoint.identification.algorithm = https connect_1 | ssl.engine.factory.class = null connect_1 | ssl.key.password = null connect_1 | ssl.keymanager.algorithm = SunX509 connect_1 | ssl.keystore.location = null connect_1 | ssl.keystore.password = null connect_1 | ssl.keystore.type = JKS connect_1 | ssl.protocol = TLSv1.3 connect_1 | ssl.provider = null connect_1 | ssl.secure.random.implementation = null connect_1 | ssl.trustmanager.algorithm = PKIX connect_1 | ssl.truststore.location = null connect_1 | ssl.truststore.password = null connect_1 | ssl.truststore.type = JKS connect_1 | transaction.timeout.ms = 60000 connect_1 | transactional.id = null connect_1 | value.serializer = class org.apache.kafka.common.serialization.ByteArraySerializer connect_1 | [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:44,003 WARN || The configuration 'metrics.context.connect.kafka.cluster.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:44,004 WARN || The configuration 'metrics.context.connect.group.id' was supplied but isn't a known config. [org.apache.kafka.clients.producer.ProducerConfig] connect_1 | 2021-01-04 17:38:44,005 INFO || Kafka version: 2.6.0 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:44,005 INFO || Kafka commitId: 62abe01bee039651 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:44,005 INFO || Kafka startTimeMs: 1609781924005 [org.apache.kafka.common.utils.AppInfoParser] connect_1 | 2021-01-04 17:38:44,018 INFO || [Producer clientId=connector-producer-rentsite-connector-0] Cluster ID: XotioX1AQwiOXSaNKjB0IA [org.apache.kafka.clients.Metadata] connect_1 | 2021-01-04 17:38:44,019 INFO || [Worker clientId=connect-1, groupId=1] Finished starting connectors and tasks [org.apache.kafka.connect.runtime.distributed.DistributedHerder] connect_1 | 2021-01-04 17:38:44,024 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,025 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,025 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,026 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,026 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,026 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,026 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,027 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,027 WARN || Using configuration property "table.blacklist" is deprecated and will be removed in future versions. Please use "table.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,027 WARN || Using configuration property "table.blacklist" is deprecated and will be removed in future versions. Please use "table.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,027 WARN || Using configuration property "schema.blacklist" is deprecated and will be removed in future versions. Please use "schema.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,027 WARN || Using configuration property "schema.blacklist" is deprecated and will be removed in future versions. Please use "schema.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,028 INFO || Starting PostgresConnectorTask with configuration: [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || connector.class = io.debezium.connector.postgresql.PostgresConnector [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || publication.autocreate.mode = filtered [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || database.user = rentsite [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || database.dbname = rentsite [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || database.server.name = rentsite-db [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || schema.include.list = rentsite [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || database.port = 5432 [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || plugin.name = pgoutput [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || task.class = io.debezium.connector.postgresql.PostgresConnectorTask [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || database.hostname = rentsite-db [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || database.password = ******** [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || name = rentsite-connector [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,033 INFO || table.include.list = rentsite.aptlist_lease [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,034 WARN || Using configuration property "schema.blacklist" is deprecated and will be removed in future versions. Please use "schema.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,034 WARN || Using configuration property "table.blacklist" is deprecated and will be removed in future versions. Please use "table.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,034 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,099 WARN || Using configuration property "schema.blacklist" is deprecated and will be removed in future versions. Please use "schema.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,099 WARN || Using configuration property "table.blacklist" is deprecated and will be removed in future versions. Please use "table.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,101 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,101 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,104 WARN || Using configuration property "schema.blacklist" is deprecated and will be removed in future versions. Please use "schema.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,104 WARN || Using configuration property "table.blacklist" is deprecated and will be removed in future versions. Please use "table.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,105 WARN || Using configuration property "column.whitelist" is deprecated and will be removed in future versions. Please use "column.include.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,105 WARN || Using configuration property "column.blacklist" is deprecated and will be removed in future versions. Please use "column.exclude.list" instead. [io.debezium.config.Configuration] connect_1 | 2021-01-04 17:38:44,207 INFO Postgres|rentsite-db|postgres-connector-task user 'rentsite' connected to database 'rentsite' on PostgreSQL 12.5 on x86_64-pc-linux-musl, compiled by gcc (Alpine 9.3.0) 9.3.0, 64-bit with roles: connect_1 | role 'pg_read_all_settings' [superuser: false, replication: false, inherit: true, create role: false, create db: false, can log in: false] connect_1 | role 'pg_stat_scan_tables' [superuser: false, replication: false, inherit: true, create role: false, create db: false, can log in: false] connect_1 | role 'pg_write_server_files' [superuser: false, replication: false, inherit: true, create role: false, create db: false, can log in: false] connect_1 | role 'pg_monitor' [superuser: false, replication: false, inherit: true, create role: false, create db: false, can log in: false] connect_1 | role 'pg_read_server_files' [superuser: false, replication: false, inherit: true, create role: false, create db: false, can log in: false] connect_1 | role 'rentsite' [superuser: true, replication: true, inherit: true, create role: true, create db: true, can log in: true] connect_1 | role 'pg_execute_server_program' [superuser: false, replication: false, inherit: true, create role: false, create db: false, can log in: false] connect_1 | role 'pg_read_all_stats' [superuser: false, replication: false, inherit: true, create role: false, create db: false, can log in: false] connect_1 | role 'pg_signal_backend' [superuser: false, replication: false, inherit: true, create role: false, create db: false, can log in: false] [io.debezium.connector.postgresql.PostgresConnectorTask] connect_1 | 2021-01-04 17:38:44,214 INFO Postgres|rentsite-db|postgres-connector-task Obtained valid replication slot ReplicationSlot [active=false, latestFlushedLsn=null, catalogXmin=null] [io.debezium.connector.postgresql.connection.PostgresConnection] connect_1 | 2021-01-04 17:38:44,214 INFO Postgres|rentsite-db|postgres-connector-task No previous offset found [io.debezium.connector.postgresql.PostgresConnectorTask] connect_1 | 2021-01-04 17:38:44,214 INFO Postgres|rentsite-db|postgres-connector-task Taking initial snapshot for new datasource [io.debezium.connector.postgresql.snapshot.InitialSnapshotter] connect_1 | 2021-01-04 17:38:44,226 INFO Postgres|rentsite-db|postgres-connector-task Initializing PgOutput logical decoder publication [io.debezium.connector.postgresql.connection.PostgresReplicationConnection] connect_1 | 2021-01-04 17:38:44,228 INFO Postgres|rentsite-db|postgres-connector-task Creating new publication 'dbz_publication' for plugin 'PGOUTPUT' [io.debezium.connector.postgresql.connection.PostgresReplicationConnection] connect_1 | 2021-01-04 17:38:44,234 INFO Postgres|rentsite-db|postgres-connector-task Creating Publication with statement 'CREATE PUBLICATION dbz_publication FOR TABLE ;' [io.debezium.connector.postgresql.connection.PostgresReplicationConnection] connect_1 | 2021-01-04 17:38:44,242 INFO || WorkerSourceTask{id=rentsite-connector-0} Committing offsets [org.apache.kafka.connect.runtime.WorkerSourceTask] connect_1 | 2021-01-04 17:38:44,242 INFO || WorkerSourceTask{id=rentsite-connector-0} flushing 0 outstanding messages for offset commit [org.apache.kafka.connect.runtime.WorkerSourceTask] connect_1 | 2021-01-04 17:38:44,242 ERROR || WorkerSourceTask{id=rentsite-connector-0} Task threw an uncaught and unrecoverable exception [org.apache.kafka.connect.runtime.WorkerTask] connect_1 | org.apache.kafka.connect.errors.ConnectException: Unable to create filtered publication dbz_publication for connect_1 | at io.debezium.connector.postgresql.connection.PostgresReplicationConnection.initPublication(PostgresReplicationConnection.java:164) connect_1 | at io.debezium.connector.postgresql.connection.PostgresReplicationConnection.createReplicationSlot(PostgresReplicationConnection.java:343) connect_1 | at io.debezium.connector.postgresql.PostgresConnectorTask.start(PostgresConnectorTask.java:125) connect_1 | at io.debezium.connector.common.BaseSourceTask.start(BaseSourceTask.java:106) connect_1 | at org.apache.kafka.connect.runtime.WorkerSourceTask.execute(WorkerSourceTask.java:232) connect_1 | at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:185) connect_1 | at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:235) connect_1 | at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) connect_1 | at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) connect_1 | at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) connect_1 | at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) connect_1 | at java.base/java.lang.Thread.run(Thread.java:834) connect_1 | Caused by: org.postgresql.util.PSQLException: ERROR: syntax error at or near ";" connect_1 | Position: 46 connect_1 | at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2532) connect_1 | at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2267) connect_1 | at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:312) connect_1 | at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:448) connect_1 | at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:369) connect_1 | at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:310) connect_1 | at org.postgresql.jdbc.PgStatement.executeCachedSql(PgStatement.java:296) connect_1 | at org.postgresql.jdbc.PgStatement.executeWithFlags(PgStatement.java:273) connect_1 | at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:268) connect_1 | at io.debezium.connector.postgresql.connection.PostgresReplicationConnection.initPublication(PostgresReplicationConnection.java:161) connect_1 | ... 11 more connect_1 | 2021-01-04 17:38:44,242 ERROR || WorkerSourceTask{id=rentsite-connector-0} Task is being killed and will not recover until manually restarted [org.apache.kafka.connect.runtime.WorkerTask] connect_1 | 2021-01-04 17:38:44,243 INFO || Stopping down connector [io.debezium.connector.common.BaseSourceTask] connect_1 | 2021-01-04 17:38:44,243 INFO || [Producer clientId=connector-producer-rentsite-connector-0] Closing the Kafka producer with timeoutMillis = 30000 ms. [org.apache.kafka.clients.producer.KafkaProducer]