btullis@deploy1002:~$ kubectl logs -f datahub-main-system-update-job-kw8xv datahub-system-update-job -f ERROR StatusLogger Log4j2 could not find a logging implementation. Please add log4j-core to the classpath. Using SimpleLogger to log to the console... . ____ _ __ _ _ /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \ ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \ \\/ ___)| |_)| | | | | || (_| | ) ) ) ) ' |____| .__|_| |_|_| |_\__, | / / / / =========|_|==============|___/=/_/_/_/ :: Spring Boot :: (v2.5.12) 2023-06-26 13:49:08.542 INFO 1 --- [ main] io.ebean.EbeanVersion : ebean version: 11.33.3 2023-06-26 13:49:08.640 INFO 1 --- [ main] io.ebean.config.properties.LoadContext : loaded properties from [application.yml] 2023-06-26 13:49:08.952 INFO 1 --- [ main] io.ebean.datasource.pool.ConnectionPool : DataSourcePool [gmsEbeanServiceConfig] autoCommit[false] transIsolation[READ_COMMITTED] min[2] max[50] 2023-06-26 13:49:10.735 INFO 1 --- [ main] io.ebean.internal.DefaultContainer : DatabasePlatform name:gmsEbeanServiceConfig platform:mysql 2023-06-26 13:49:13.551 INFO 1 --- [ main] c.l.g.f.k.s.KafkaSchemaRegistryFactory : creating schema registry config using url: http://localhost:8081 2023-06-26 13:49:13.753 INFO 1 --- [ main] o.a.k.clients.producer.ProducerConfig : ProducerConfig values: acks = 1 batch.size = 16384 bootstrap.servers = [kafka-test1006.eqiad.wmnet:9092] buffer.memory = 33554432 client.dns.lookup = default client.id = producer-1 compression.type = none connections.max.idle.ms = 540000 delivery.timeout.ms = 30000 enable.idempotence = false interceptor.classes = [] key.serializer = class org.apache.kafka.common.serialization.StringSerializer linger.ms = 0 max.block.ms = 60000 max.in.flight.requests.per.connection = 5 max.request.size = 1048576 metadata.max.age.ms = 300000 metadata.max.idle.ms = 300000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner receive.buffer.bytes = 32768 reconnect.backoff.max.ms = 1000 reconnect.backoff.ms = 50 request.timeout.ms = 3000 retries = 3 retry.backoff.ms = 500 sasl.client.callback.handler.class = null sasl.jaas.config = null sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.login.callback.handler.class = null sasl.login.class = null sasl.login.refresh.buffer.seconds = 300 sasl.login.refresh.min.period.seconds = 60 sasl.login.refresh.window.factor = 0.8 sasl.login.refresh.window.jitter = 0.05 sasl.mechanism = GSSAPI security.protocol = PLAINTEXT security.providers = null send.buffer.bytes = 131072 ssl.cipher.suites = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = https ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS transaction.timeout.ms = 60000 transactional.id = null value.serializer = class io.confluent.kafka.serializers.KafkaAvroSerializer 2023-06-26 13:49:13.771 INFO 1 --- [ main] i.c.k.s.KafkaAvroSerializerConfig : KafkaAvroSerializerConfig values: bearer.auth.token = [hidden] proxy.port = -1 schema.reflection = false auto.register.schemas = true max.schemas.per.subject = 1000 basic.auth.credentials.source = URL value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy schema.registry.url = [http://localhost:8081] basic.auth.user.info = [hidden] proxy.host = use.latest.version = false schema.registry.basic.auth.user.info = [hidden] bearer.auth.credentials.source = STATIC_TOKEN key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy 2023-06-26 13:49:14.134 WARN 1 --- [ main] o.a.k.clients.producer.ProducerConfig : The configuration 'schema.registry.security.protocol' was supplied but isn't a known config. 2023-06-26 13:49:14.138 INFO 1 --- [ main] o.a.kafka.common.utils.AppInfoParser : Kafka version: 5.5.1-ccs 2023-06-26 13:49:14.138 INFO 1 --- [ main] o.a.kafka.common.utils.AppInfoParser : Kafka commitId: cb1873c1fdf5f5f9 2023-06-26 13:49:14.138 INFO 1 --- [ main] o.a.kafka.common.utils.AppInfoParser : Kafka startTimeMs: 1687787354135 2023-06-26 13:49:14.542 INFO 1 --- [ main] c.l.m.m.registry.ConfigEntityRegistry : Loading bare config entity registry file at /datahub/datahub-gms/resources/entity-registry.yml 2023-06-26 13:49:16.833 INFO 1 --- [ad | producer-1] org.apache.kafka.clients.Metadata : [Producer clientId=producer-1] Cluster ID: mLKVObxtRjeI8fkVzjClCA 2023-06-26 13:49:20.150 INFO 1 --- [ main] org.reflections.Reflections : Reflections took 5093 ms to scan 46 urls, producing 836 keys and 8276 values 2023-06-26 13:49:20.939 INFO 1 --- [ main] org.reflections.Reflections : Reflections took 69 ms to scan 8 urls, producing 34 keys and 72 values 2023-06-26 13:49:27.536 WARN 1 --- [ main] c.l.m.m.r.PluginEntityRegistryLoader : /etc/datahub/plugins/models directory does not exist or is not a directory. Plugin scanning will be disabled. 2023-06-26 13:49:28.057 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataHubPolicyKey schema is compatible with previous schema due to 2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataHubPolicyInfo schema is compatible with previous schema due to 2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpGroupInfo schema is compatible with previous schema due to 2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpGroupKey schema is compatible with previous schema due to 2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : globalTags schema is compatible with previous schema due to 2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : status schema is compatible with previous schema due to 2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlFeatureTableKey schema is compatible with previous schema due to 2023-06-26 13:49:28.059 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlModelGroupKey schema is compatible with previous schema due to 2023-06-26 13:49:28.059 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataJobKey schema is compatible with previous schema due to 2023-06-26 13:49:28.132 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : ownership schema is compatible with previous schema due to 2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : tagKey schema is compatible with previous schema due to 2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : tagProperties schema is compatible with previous schema due to 2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : schemaFieldKey schema is compatible with previous schema due to 2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : glossaryTermInfo schema is compatible with previous schema due to 2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : ownership schema is compatible with previous schema due to 2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : glossaryTermKey schema is compatible with previous schema due to 2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlPrimaryKeyKey schema is compatible with previous schema due to 2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dashboardKey schema is compatible with previous schema due to 2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataPlatformKey schema is compatible with previous schema due to 2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataPlatformInfo schema is compatible with previous schema due to 2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataHubRetentionConfig schema is compatible with previous schema due to 2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataHubRetentionKey schema is compatible with previous schema due to 2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpUserKey schema is compatible with previous schema due to 2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpUserEditableInfo schema is compatible with previous schema due to 2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpUserInfo schema is compatible with previous schema due to 2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpUserStatus schema is compatible with previous schema due to 2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : groupMembership schema is compatible with previous schema due to 2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : globalTags schema is compatible with previous schema due to 2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : status schema is compatible with previous schema due to 2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlModelKey schema is compatible with previous schema due to 2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataFlowKey schema is compatible with previous schema due to 2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : glossaryNodeInfo schema is compatible with previous schema due to 2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : glossaryNodeKey schema is compatible with previous schema due to 2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : ownership schema is compatible with previous schema due to 2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : status schema is compatible with previous schema due to 2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlFeatureKey schema is compatible with previous schema due to 2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : datasetKey schema is compatible with previous schema due to 2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : viewProperties schema is compatible with previous schema due to 2023-06-26 13:49:28.137 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : schemaMetadata schema is compatible with previous schema due to 2023-06-26 13:49:28.138 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : status schema is compatible with previous schema due to 2023-06-26 13:49:28.138 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : chartKey schema is compatible with previous schema due to 2023-06-26 13:49:29.044 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No scheduled executor is provided to HttpClientFactory, using it's own scheduled executor. 2023-06-26 13:49:29.045 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No callback executor is provided to HttpClientFactory, using it's own call back executor. 2023-06-26 13:49:29.045 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No Compression executor is provided to HttpClientFactory, using it's own compression executor. 2023-06-26 13:49:29.052 INFO 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : The service 'null' has been assigned to the ChannelPoolManager with key 'noSpecifiedNamePrefix 1138266797 ', http.protocolVersion=HTTP_1_1, usePipelineV2=false, requestTimeout=10000ms, streamingTimeout=-1ms 2023-06-26 13:49:36.240 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No scheduled executor is provided to HttpClientFactory, using it's own scheduled executor. 2023-06-26 13:49:36.240 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No callback executor is provided to HttpClientFactory, using it's own call back executor. 2023-06-26 13:49:36.240 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No Compression executor is provided to HttpClientFactory, using it's own compression executor. 2023-06-26 13:49:36.241 INFO 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : The service 'null' has been assigned to the ChannelPoolManager with key 'noSpecifiedNamePrefix 1138266797 ', http.protocolVersion=HTTP_1_1, usePipelineV2=false, requestTimeout=10000ms, streamingTimeout=-1ms 2023-06-26 13:49:36.749 INFO 1 --- [ main] c.l.g.f.k.s.AwsGlueSchemaRegistryFactory : Creating AWS Glue registry 2023-06-26 13:49:36.836 INFO 1 --- [ main] c.l.g.f.k.SimpleKafkaConsumerFactory : Simple KafkaListenerContainerFactory built successfully 2023-06-26 13:49:36.851 INFO 1 --- [ main] c.l.g.f.kafka.KafkaEventConsumerFactory : Event-based KafkaListenerContainerFactory built successfully. Consumers = 1 2023-06-26 13:49:41.548 WARN 1 --- [ main] c.d.p.configuration.ConfigProvider : Configuration config.yml file not found at location /etc/datahub/plugins/auth 2023-06-26 13:49:41.548 INFO 1 --- [ main] c.l.g.f.auth.AuthorizerChainFactory : Default DataHubAuthorizer is enabled. Appending it to the authorization chain. Starting upgrade with id SystemUpdate... Executing Step 1/5: BuildIndicesPreStep... 2023-06-26 13:49:45.338 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's diff between new mappings (left) and old mappings (right): {ldap=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {fields={delimited={analyzer=word_delimited, search_analyzer=query_word_delimited, search_quote_analyzer=quote_analyzer, type=text}, keyword={type=keyword}, ngram={analyzer=partial, doc_values=false, max_shingle_size=4, type=search_as_you_type}}, normalizer=keyword_normalizer, type=keyword}), urn=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {fields={delimited={analyzer=urn_component, search_analyzer=query_urn_component, search_quote_analyzer=quote_analyzer, type=text}, ngram={analyzer=partial_urn_component, doc_values=false, max_shingle_size=4, type=search_as_you_type}}, type=keyword})} 2023-06-26 13:49:45.534 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's an update to `analysis` settings that requires reindexing. Target: {"index":{"max_ngram_diff":17,"refresh_interval":"1s","analysis":{"filter":{"autocomplete_custom_delimiter":{"type":"word_delimiter","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"sticky_delimiter_graph":{"type":"word_delimiter_graph","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"generate_number_parts":false,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"datahub_stop_words":{"type":"stop","ignore_case":"true","stopwords":["urn","li","data","table"]},"min_length":{"type":"length","min":"3"},"stem_override":{"type":"stemmer_override","rules":["customers, customer => customer","staging => staging","production => production","urn:li:dataplatform:hive => urn:li:dataplatform:hive","hive => hive","bigquery => bigquery","big query => big query","query => query"]},"alpha_num_space":{"type":"pattern_capture","patterns":["([a-z0-9 _-]{2,})","([a-z0-9 ]{2,})","\\\"([^\\\"]*)\\\""]},"remove_quotes":{"type":"pattern_replace","pattern":"['\"]","replacement":""},"multifilter":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph,flatten_graph","lowercase,alpha_num_space,default_syn_graph,flatten_graph"]},"multifilter_graph":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph","lowercase,alpha_num_space,default_syn_graph"]},"default_syn_graph":{"type":"synonym_graph","lenient":"false","synonyms":["cac, customer acquisition cost => cac, customer, acquisition, cost","stg, staging","dev, development","prod, production","glue, athena","s3, s_3","data platform, dataplatform","bigquery, big query => bigquery, big, query"]}},"tokenizer":{"slash_tokenizer":{"type":"pattern","pattern":"[/]"},"main_tokenizer":{"type":"pattern","pattern":"[(),./:]"}},"normalizer":{"keyword_normalizer":{"filter":["lowercase","asciifolding"]}},"analyzer":{"slash_pattern":{"tokenizer":"slash_tokenizer","filter":["lowercase"]},"browse_path_hierarchy":{"tokenizer":"path_hierarchy"},"custom_keyword":{"tokenizer":"keyword","filter":["trim","lowercase","asciifolding","snowball"]},"quote_analyzer":{"tokenizer":"keyword","filter":["asciifolding","lowercase","remove_quotes","datahub_stop_words","stop","min_length"]},"word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"partial":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]},"partial_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]}}},"number_of_shards":1,"number_of_replicas":1}} Current: {"index.creation_date":"1683106114246","index.number_of_replicas":"1","index.number_of_shards":"1","index.provided_name":"staging_corpuserindex_v2","index.uuid":"3naEY6-uRnibJV9p-5HA_A","index.version.created":"135238227"} 2023-06-26 13:49:45.537 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's an update to settings that requires reindexing. Target: {"index":{"max_ngram_diff":17,"refresh_interval":"1s","analysis":{"filter":{"autocomplete_custom_delimiter":{"type":"word_delimiter","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"sticky_delimiter_graph":{"type":"word_delimiter_graph","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"generate_number_parts":false,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"datahub_stop_words":{"type":"stop","ignore_case":"true","stopwords":["urn","li","data","table"]},"min_length":{"type":"length","min":"3"},"stem_override":{"type":"stemmer_override","rules":["customers, customer => customer","staging => staging","production => production","urn:li:dataplatform:hive => urn:li:dataplatform:hive","hive => hive","bigquery => bigquery","big query => big query","query => query"]},"alpha_num_space":{"type":"pattern_capture","patterns":["([a-z0-9 _-]{2,})","([a-z0-9 ]{2,})","\\\"([^\\\"]*)\\\""]},"remove_quotes":{"type":"pattern_replace","pattern":"['\"]","replacement":""},"multifilter":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph,flatten_graph","lowercase,alpha_num_space,default_syn_graph,flatten_graph"]},"multifilter_graph":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph","lowercase,alpha_num_space,default_syn_graph"]},"default_syn_graph":{"type":"synonym_graph","lenient":"false","synonyms":["cac, customer acquisition cost => cac, customer, acquisition, cost","stg, staging","dev, development","prod, production","glue, athena","s3, s_3","data platform, dataplatform","bigquery, big query => bigquery, big, query"]}},"tokenizer":{"slash_tokenizer":{"type":"pattern","pattern":"[/]"},"main_tokenizer":{"type":"pattern","pattern":"[(),./:]"}},"normalizer":{"keyword_normalizer":{"filter":["lowercase","asciifolding"]}},"analyzer":{"slash_pattern":{"tokenizer":"slash_tokenizer","filter":["lowercase"]},"browse_path_hierarchy":{"tokenizer":"path_hierarchy"},"custom_keyword":{"tokenizer":"keyword","filter":["trim","lowercase","asciifolding","snowball"]},"quote_analyzer":{"tokenizer":"keyword","filter":["asciifolding","lowercase","remove_quotes","datahub_stop_words","stop","min_length"]},"word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"partial":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]},"partial_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]}}},"number_of_shards":1,"number_of_replicas":1}} Current: {"index.creation_date":"1683106114246","index.number_of_replicas":"1","index.number_of_shards":"1","index.provided_name":"staging_corpuserindex_v2","index.uuid":"3naEY6-uRnibJV9p-5HA_A","index.version.created":"135238227"} 2023-06-26 13:49:45.735 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_system_metadata_service_v1 - There's diff between new mappings (left) and old mappings (right): {aspect=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword}), runId=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword}), urn=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword})} 2023-06-26 13:49:46.032 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPreStep : Updated index staging_corpuserindex_v2 with new settings. Settings: {index.blocks.write=true}, Acknowledged: true 2023-06-26 13:49:46.573 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPreStep : Cloned index staging_corpuserindex_v2 into staging_corpuserindex_v2_clone_1687787386034, Acknowledged: true 2023-06-26 13:49:46.722 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPreStep : Updated index staging_system_metadata_service_v1 with new settings. Settings: {index.blocks.write=true}, Acknowledged: true 2023-06-26 13:49:47.128 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPreStep : Cloned index staging_system_metadata_service_v1 into staging_system_metadata_service_v1_clone_1687787386722, Acknowledged: true Completed Step 1/5: BuildIndicesPreStep successfully. Executing Step 2/5: BuildIndicesStep... 2023-06-26 13:49:47.130 INFO 1 --- [ main] c.l.m.g.e.ElasticSearchGraphService : Setting up elastic graph index 2023-06-26 13:49:47.132 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_graph_service_v1 does not exist. Creating 2023-06-26 13:49:47.447 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_graph_service_v1 2023-06-26 13:49:47.738 ERROR 1 --- [pool-9-thread-1] c.l.m.s.elasticsearch.query.ESSearchDAO : Search query failed org.elasticsearch.ElasticsearchStatusException: Elasticsearch exception [type=index_not_found_exception, reason=no such index [staging_datahubpolicyindex_v2]] at org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:187) ~[elasticsearch-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.parseEntity(RestHighLevelClient.java:1911) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.parseResponseException(RestHighLevelClient.java:1888) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1645) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.performRequest(RestHighLevelClient.java:1602) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.performRequestAndParseEntity(RestHighLevelClient.java:1572) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.search(RestHighLevelClient.java:1088) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.executeAndExtract(ESSearchDAO.java:57) ~[metadata-io.jar!/:na] at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.search(ESSearchDAO.java:90) ~[metadata-io.jar!/:na] at com.linkedin.metadata.search.elasticsearch.ElasticSearchService.fullTextSearch(ElasticSearchService.java:111) ~[metadata-io.jar!/:na] at com.linkedin.metadata.client.JavaEntityClient.search(JavaEntityClient.java:312) ~[metadata-io.jar!/:na] at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:50) ~[auth-impl.jar!/:na] at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:42) ~[auth-impl.jar!/:na] at com.datahub.authorization.DataHubAuthorizer$PolicyRefreshRunnable.run(DataHubAuthorizer.java:223) ~[auth-impl.jar!/:na] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) ~[na:na] at java.base/java.util.concurrent.FutureTask.runAndReset(FutureTask.java:305) ~[na:na] at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:305) ~[na:na] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[na:na] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[na:na] at java.base/java.lang.Thread.run(Thread.java:829) ~[na:na] Suppressed: org.elasticsearch.client.ResponseException: method [POST], host [http://datahubsearch.svc.eqiad.wmnet:9200], URI [/staging_datahubpolicyindex_v2/_search?typed_keys=true&max_concurrent_shard_requests=5&ignore_unavailable=false&expand_wildcards=open&allow_no_indices=true&ignore_throttled=true&search_type=query_then_fetch&batched_reduce_size=512&ccs_minimize_roundtrips=true], status line [HTTP/1.1 404 Not Found] {"error":{"root_cause":[{"type":"index_not_found_exception","reason":"no such index [staging_datahubpolicyindex_v2]","index":"staging_datahubpolicyindex_v2","resource.id":"staging_datahubpolicyindex_v2","resource.type":"index_or_alias","index_uuid":"_na_"}],"type":"index_not_found_exception","reason":"no such index [staging_datahubpolicyindex_v2]","index":"staging_datahubpolicyindex_v2","resource.id":"staging_datahubpolicyindex_v2","resource.type":"index_or_alias","index_uuid":"_na_"},"status":404} at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:326) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestClient.performRequest(RestClient.java:296) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestClient.performRequest(RestClient.java:270) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1632) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] ... 16 common frames omitted 2023-06-26 13:49:47.738 ERROR 1 --- [pool-9-thread-1] c.d.authorization.DataHubAuthorizer : Failed to retrieve policy urns! Skipping updating policy cache until next refresh. start: 0, count: 30 com.datahub.util.exception.ESQueryException: Search query failed: at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.executeAndExtract(ESSearchDAO.java:62) ~[metadata-io.jar!/:na] at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.search(ESSearchDAO.java:90) ~[metadata-io.jar!/:na] at com.linkedin.metadata.search.elasticsearch.ElasticSearchService.fullTextSearch(ElasticSearchService.java:111) ~[metadata-io.jar!/:na] at com.linkedin.metadata.client.JavaEntityClient.search(JavaEntityClient.java:312) ~[metadata-io.jar!/:na] at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:50) ~[auth-impl.jar!/:na] at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:42) ~[auth-impl.jar!/:na] at com.datahub.authorization.DataHubAuthorizer$PolicyRefreshRunnable.run(DataHubAuthorizer.java:223) ~[auth-impl.jar!/:na] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) ~[na:na] at java.base/java.util.concurrent.FutureTask.runAndReset(FutureTask.java:305) ~[na:na] at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:305) ~[na:na] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[na:na] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[na:na] at java.base/java.lang.Thread.run(Thread.java:829) ~[na:na] Caused by: org.elasticsearch.ElasticsearchStatusException: Elasticsearch exception [type=index_not_found_exception, reason=no such index [staging_datahubpolicyindex_v2]] at org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:187) ~[elasticsearch-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.parseEntity(RestHighLevelClient.java:1911) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.parseResponseException(RestHighLevelClient.java:1888) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1645) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.performRequest(RestHighLevelClient.java:1602) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.performRequestAndParseEntity(RestHighLevelClient.java:1572) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.search(RestHighLevelClient.java:1088) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.executeAndExtract(ESSearchDAO.java:57) ~[metadata-io.jar!/:na] ... 12 common frames omitted Suppressed: org.elasticsearch.client.ResponseException: method [POST], host [http://datahubsearch.svc.eqiad.wmnet:9200], URI [/staging_datahubpolicyindex_v2/_search?typed_keys=true&max_concurrent_shard_requests=5&ignore_unavailable=false&expand_wildcards=open&allow_no_indices=true&ignore_throttled=true&search_type=query_then_fetch&batched_reduce_size=512&ccs_minimize_roundtrips=true], status line [HTTP/1.1 404 Not Found] {"error":{"root_cause":[{"type":"index_not_found_exception","reason":"no such index [staging_datahubpolicyindex_v2]","index":"staging_datahubpolicyindex_v2","resource.id":"staging_datahubpolicyindex_v2","resource.type":"index_or_alias","index_uuid":"_na_"}],"type":"index_not_found_exception","reason":"no such index [staging_datahubpolicyindex_v2]","index":"staging_datahubpolicyindex_v2","resource.id":"staging_datahubpolicyindex_v2","resource.type":"index_or_alias","index_uuid":"_na_"},"status":404} at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:326) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestClient.performRequest(RestClient.java:296) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestClient.performRequest(RestClient.java:270) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2] at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1632) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2] ... 16 common frames omitted 2023-06-26 13:49:47.749 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's diff between new mappings (left) and old mappings (right): {ldap=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {fields={delimited={analyzer=word_delimited, search_analyzer=query_word_delimited, search_quote_analyzer=quote_analyzer, type=text}, keyword={type=keyword}, ngram={analyzer=partial, doc_values=false, max_shingle_size=4, type=search_as_you_type}}, normalizer=keyword_normalizer, type=keyword}), urn=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {fields={delimited={analyzer=urn_component, search_analyzer=query_urn_component, search_quote_analyzer=quote_analyzer, type=text}, ngram={analyzer=partial_urn_component, doc_values=false, max_shingle_size=4, type=search_as_you_type}}, type=keyword})} 2023-06-26 13:49:47.750 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's an update to `analysis` settings that requires reindexing. Target: {"index":{"max_ngram_diff":17,"refresh_interval":"1s","analysis":{"filter":{"autocomplete_custom_delimiter":{"type":"word_delimiter","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"sticky_delimiter_graph":{"type":"word_delimiter_graph","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"generate_number_parts":false,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"datahub_stop_words":{"type":"stop","ignore_case":"true","stopwords":["urn","li","data","table"]},"min_length":{"type":"length","min":"3"},"stem_override":{"type":"stemmer_override","rules":["customers, customer => customer","staging => staging","production => production","urn:li:dataplatform:hive => urn:li:dataplatform:hive","hive => hive","bigquery => bigquery","big query => big query","query => query"]},"alpha_num_space":{"type":"pattern_capture","patterns":["([a-z0-9 _-]{2,})","([a-z0-9 ]{2,})","\\\"([^\\\"]*)\\\""]},"remove_quotes":{"type":"pattern_replace","pattern":"['\"]","replacement":""},"multifilter":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph,flatten_graph","lowercase,alpha_num_space,default_syn_graph,flatten_graph"]},"multifilter_graph":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph","lowercase,alpha_num_space,default_syn_graph"]},"default_syn_graph":{"type":"synonym_graph","lenient":"false","synonyms":["cac, customer acquisition cost => cac, customer, acquisition, cost","stg, staging","dev, development","prod, production","glue, athena","s3, s_3","data platform, dataplatform","bigquery, big query => bigquery, big, query"]}},"tokenizer":{"slash_tokenizer":{"type":"pattern","pattern":"[/]"},"main_tokenizer":{"type":"pattern","pattern":"[(),./:]"}},"normalizer":{"keyword_normalizer":{"filter":["lowercase","asciifolding"]}},"analyzer":{"slash_pattern":{"tokenizer":"slash_tokenizer","filter":["lowercase"]},"browse_path_hierarchy":{"tokenizer":"path_hierarchy"},"custom_keyword":{"tokenizer":"keyword","filter":["trim","lowercase","asciifolding","snowball"]},"quote_analyzer":{"tokenizer":"keyword","filter":["asciifolding","lowercase","remove_quotes","datahub_stop_words","stop","min_length"]},"word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"partial":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]},"partial_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]}}},"number_of_shards":1,"number_of_replicas":1}} Current: {"index.blocks.write":"true","index.creation_date":"1683106114246","index.number_of_replicas":"1","index.number_of_shards":"1","index.provided_name":"staging_corpuserindex_v2","index.uuid":"3naEY6-uRnibJV9p-5HA_A","index.version.created":"135238227"} 2023-06-26 13:49:47.750 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's an update to settings that requires reindexing. Target: {"index":{"max_ngram_diff":17,"refresh_interval":"1s","analysis":{"filter":{"autocomplete_custom_delimiter":{"type":"word_delimiter","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"sticky_delimiter_graph":{"type":"word_delimiter_graph","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"generate_number_parts":false,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"datahub_stop_words":{"type":"stop","ignore_case":"true","stopwords":["urn","li","data","table"]},"min_length":{"type":"length","min":"3"},"stem_override":{"type":"stemmer_override","rules":["customers, customer => customer","staging => staging","production => production","urn:li:dataplatform:hive => urn:li:dataplatform:hive","hive => hive","bigquery => bigquery","big query => big query","query => query"]},"alpha_num_space":{"type":"pattern_capture","patterns":["([a-z0-9 _-]{2,})","([a-z0-9 ]{2,})","\\\"([^\\\"]*)\\\""]},"remove_quotes":{"type":"pattern_replace","pattern":"['\"]","replacement":""},"multifilter":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph,flatten_graph","lowercase,alpha_num_space,default_syn_graph,flatten_graph"]},"multifilter_graph":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph","lowercase,alpha_num_space,default_syn_graph"]},"default_syn_graph":{"type":"synonym_graph","lenient":"false","synonyms":["cac, customer acquisition cost => cac, customer, acquisition, cost","stg, staging","dev, development","prod, production","glue, athena","s3, s_3","data platform, dataplatform","bigquery, big query => bigquery, big, query"]}},"tokenizer":{"slash_tokenizer":{"type":"pattern","pattern":"[/]"},"main_tokenizer":{"type":"pattern","pattern":"[(),./:]"}},"normalizer":{"keyword_normalizer":{"filter":["lowercase","asciifolding"]}},"analyzer":{"slash_pattern":{"tokenizer":"slash_tokenizer","filter":["lowercase"]},"browse_path_hierarchy":{"tokenizer":"path_hierarchy"},"custom_keyword":{"tokenizer":"keyword","filter":["trim","lowercase","asciifolding","snowball"]},"quote_analyzer":{"tokenizer":"keyword","filter":["asciifolding","lowercase","remove_quotes","datahub_stop_words","stop","min_length"]},"word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"partial":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]},"partial_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]}}},"number_of_shards":1,"number_of_replicas":1}} Current: {"index.blocks.write":"true","index.creation_date":"1683106114246","index.number_of_replicas":"1","index.number_of_shards":"1","index.provided_name":"staging_corpuserindex_v2","index.uuid":"3naEY6-uRnibJV9p-5HA_A","index.version.created":"135238227"} 2023-06-26 13:49:47.855 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_containerindex_v2 does not exist. Creating 2023-06-26 13:49:48.268 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_containerindex_v2 2023-06-26 13:49:48.268 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubpolicyindex_v2 does not exist. Creating 2023-06-26 13:49:48.716 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubpolicyindex_v2 2023-06-26 13:49:48.716 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubaccesstokenindex_v2 does not exist. Creating 2023-06-26 13:49:49.079 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubaccesstokenindex_v2 2023-06-26 13:49:49.079 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubupgradeindex_v2 does not exist. Creating 2023-06-26 13:49:49.443 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubupgradeindex_v2 2023-06-26 13:49:49.444 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubroleindex_v2 does not exist. Creating 2023-06-26 13:49:49.790 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubroleindex_v2 2023-06-26 13:49:49.791 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_corpgroupindex_v2 does not exist. Creating 2023-06-26 13:49:50.118 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_corpgroupindex_v2 2023-06-26 13:49:50.119 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubstepstateindex_v2 does not exist. Creating 2023-06-26 13:49:50.451 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubstepstateindex_v2 2023-06-26 13:49:50.451 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_globalsettingsindex_v2 does not exist. Creating 2023-06-26 13:49:50.853 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_globalsettingsindex_v2 2023-06-26 13:49:50.853 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataprocessindex_v2 does not exist. Creating 2023-06-26 13:49:51.169 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataprocessindex_v2 2023-06-26 13:49:51.169 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlfeaturetableindex_v2 does not exist. Creating 2023-06-26 13:49:51.513 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlfeaturetableindex_v2 2023-06-26 13:49:51.513 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlmodelgroupindex_v2 does not exist. Creating 2023-06-26 13:49:51.824 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlmodelgroupindex_v2 2023-06-26 13:49:51.825 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_postindex_v2 does not exist. Creating 2023-06-26 13:49:52.130 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_postindex_v2 2023-06-26 13:49:52.131 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubexecutionrequestindex_v2 does not exist. Creating 2023-06-26 13:49:52.459 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubexecutionrequestindex_v2 2023-06-26 13:49:52.459 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_invitetokenindex_v2 does not exist. Creating 2023-06-26 13:49:52.826 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_invitetokenindex_v2 2023-06-26 13:49:52.826 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datajobindex_v2 does not exist. Creating 2023-06-26 13:49:53.185 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datajobindex_v2 2023-06-26 13:49:53.185 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_assertionindex_v2 does not exist. Creating 2023-06-26 13:49:53.575 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_assertionindex_v2 2023-06-26 13:49:53.575 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataplatforminstanceindex_v2 does not exist. Creating 2023-06-26 13:49:53.934 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataplatforminstanceindex_v2 2023-06-26 13:49:53.934 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_tagindex_v2 does not exist. Creating 2023-06-26 13:49:54.294 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_tagindex_v2 2023-06-26 13:49:54.295 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_schemafieldindex_v2 does not exist. Creating 2023-06-26 13:49:54.598 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_schemafieldindex_v2 2023-06-26 13:49:54.598 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_glossarytermindex_v2 does not exist. Creating 2023-06-26 13:49:54.964 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_glossarytermindex_v2 2023-06-26 13:49:54.964 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlprimarykeyindex_v2 does not exist. Creating 2023-06-26 13:49:55.270 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlprimarykeyindex_v2 2023-06-26 13:49:55.270 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dashboardindex_v2 does not exist. Creating 2023-06-26 13:49:55.579 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dashboardindex_v2 2023-06-26 13:49:55.579 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_notebookindex_v2 does not exist. Creating 2023-06-26 13:49:55.931 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_notebookindex_v2 2023-06-26 13:49:55.931 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlmodeldeploymentindex_v2 does not exist. Creating 2023-06-26 13:49:56.276 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlmodeldeploymentindex_v2 2023-06-26 13:49:56.276 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataplatformindex_v2 does not exist. Creating 2023-06-26 13:49:56.653 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataplatformindex_v2 2023-06-26 13:49:56.653 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubretentionindex_v2 does not exist. Creating 2023-06-26 13:49:56.970 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubretentionindex_v2 2023-06-26 13:49:57.011 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_corpuserindex_v2_1687787396970 does not exist. Creating 2023-06-26 13:49:57.290 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_corpuserindex_v2_1687787396970 2023-06-26 13:50:17.351 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Task: 50uKScxYREyFVlCeM4NwpQ:131447489 - Reindexing from staging_corpuserindex_v2 to staging_corpuserindex_v2_1687787396970 in progress... 2023-06-26 13:50:17.361 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Task: 50uKScxYREyFVlCeM4NwpQ:131447489 - Reindexing staging_corpuserindex_v2 to staging_corpuserindex_v2_1687787396970 task was successful 2023-06-26 13:50:17.361 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Reindex from staging_corpuserindex_v2 to staging_corpuserindex_v2_1687787396970 succeeded 2023-06-26 13:50:17.606 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Finished setting up staging_corpuserindex_v2 2023-06-26 13:50:17.606 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_testindex_v2 does not exist. Creating 2023-06-26 13:50:17.824 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_testindex_v2 2023-06-26 13:50:17.824 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlmodelindex_v2 does not exist. Creating 2023-06-26 13:50:18.238 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlmodelindex_v2 2023-06-26 13:50:18.238 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataflowindex_v2 does not exist. Creating 2023-06-26 13:50:18.687 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataflowindex_v2 2023-06-26 13:50:18.687 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlfeatureindex_v2 does not exist. Creating 2023-06-26 13:50:19.021 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlfeatureindex_v2 2023-06-26 13:50:19.021 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_glossarynodeindex_v2 does not exist. Creating 2023-06-26 13:50:19.382 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_glossarynodeindex_v2 2023-06-26 13:50:19.383 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubingestionsourceindex_v2 does not exist. Creating 2023-06-26 13:50:19.736 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubingestionsourceindex_v2 2023-06-26 13:50:19.736 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_domainindex_v2 does not exist. Creating 2023-06-26 13:50:20.094 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_domainindex_v2 2023-06-26 13:50:20.094 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubviewindex_v2 does not exist. Creating 2023-06-26 13:50:20.429 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubviewindex_v2 2023-06-26 13:50:20.429 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_telemetryindex_v2 does not exist. Creating 2023-06-26 13:50:20.853 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_telemetryindex_v2 2023-06-26 13:50:20.853 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubsecretindex_v2 does not exist. Creating 2023-06-26 13:50:21.131 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubsecretindex_v2 2023-06-26 13:50:21.131 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_chartindex_v2 does not exist. Creating 2023-06-26 13:50:21.430 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_chartindex_v2 2023-06-26 13:50:21.430 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datasetindex_v2 does not exist. Creating 2023-06-26 13:50:21.735 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datasetindex_v2 2023-06-26 13:50:21.735 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataprocessinstanceindex_v2 does not exist. Creating 2023-06-26 13:50:22.089 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataprocessinstanceindex_v2 2023-06-26 13:50:22.090 INFO 1 --- [ main] l.m.s.ElasticSearchSystemMetadataService : Setting up system metadata index 2023-06-26 13:50:22.098 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_system_metadata_service_v1 - There's diff between new mappings (left) and old mappings (right): {aspect=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword}), runId=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword}), urn=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword})} 2023-06-26 13:50:22.104 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_system_metadata_service_v1_1687787422099 does not exist. Creating 2023-06-26 13:50:22.414 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_system_metadata_service_v1_1687787422099 2023-06-26 13:50:42.443 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Task: 50uKScxYREyFVlCeM4NwpQ:131448073 - Reindexing from staging_system_metadata_service_v1 to staging_system_metadata_service_v1_1687787422099 in progress... 2023-06-26 13:50:42.454 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Task: 50uKScxYREyFVlCeM4NwpQ:131448073 - Reindexing staging_system_metadata_service_v1 to staging_system_metadata_service_v1_1687787422099 task was successful 2023-06-26 13:50:42.454 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Reindex from staging_system_metadata_service_v1 to staging_system_metadata_service_v1_1687787422099 succeeded 2023-06-26 13:50:42.647 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Finished setting up staging_system_metadata_service_v1 2023-06-26 13:50:42.662 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datajob_datahubingestionrunsummaryaspect_v1 does not exist. Creating 2023-06-26 13:50:42.929 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datajob_datahubingestionrunsummaryaspect_v1 2023-06-26 13:50:42.929 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datajob_datahubingestioncheckpointaspect_v1 does not exist. Creating 2023-06-26 13:50:43.242 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datajob_datahubingestioncheckpointaspect_v1 2023-06-26 13:50:43.242 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_assertion_assertionruneventaspect_v1 does not exist. Creating 2023-06-26 13:50:43.528 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_assertion_assertionruneventaspect_v1 2023-06-26 13:50:43.528 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dashboard_dashboardusagestatisticsaspect_v1 does not exist. Creating 2023-06-26 13:50:43.804 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dashboard_dashboardusagestatisticsaspect_v1 2023-06-26 13:50:43.804 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_chart_chartusagestatisticsaspect_v1 does not exist. Creating 2023-06-26 13:50:44.114 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_chart_chartusagestatisticsaspect_v1 2023-06-26 13:50:44.115 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataset_datasetprofileaspect_v1 does not exist. Creating 2023-06-26 13:50:44.427 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataset_datasetprofileaspect_v1 2023-06-26 13:50:44.427 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataset_datasetusagestatisticsaspect_v1 does not exist. Creating 2023-06-26 13:50:44.725 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataset_datasetusagestatisticsaspect_v1 2023-06-26 13:50:44.725 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataset_operationaspect_v1 does not exist. Creating 2023-06-26 13:50:45.032 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataset_operationaspect_v1 2023-06-26 13:50:45.032 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataprocessinstance_dataprocessinstanceruneventaspect_v1 does not exist. Creating 2023-06-26 13:50:45.342 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataprocessinstance_dataprocessinstanceruneventaspect_v1 Completed Step 2/5: BuildIndicesStep successfully. Executing Step 3/5: BuildIndicesPostStep... 2023-06-26 13:50:45.489 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPostStep : Updated index staging_corpuserindex_v2 with new settings. Settings: {index.blocks.write=false}, Acknowledged: true 2023-06-26 13:50:45.600 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPostStep : Updated index staging_system_metadata_service_v1 with new settings. Settings: {index.blocks.write=false}, Acknowledged: true Completed Step 3/5: BuildIndicesPostStep successfully. Executing Step 4/5: DataHubStartupStep... 2023-06-26 13:50:46.744 ERROR 1 --- [ main] i.c.k.s.client.rest.RestService : Failed to send HTTP request to endpoint: http://localhost:8081/subjects/DataHubUpgradeHistory_v1-value/versions java.net.ConnectException: Connection refused (Connection refused) at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na] at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na] at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na] at java.base/sun.net.www.http.HttpClient.(HttpClient.java:275) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na] at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na] at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na] at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na] at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na] at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na] at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12] at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na] at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na] at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na] 2023-06-26 13:50:46.745 ERROR 1 --- [ main] c.l.d.u.s.e.steps.DataHubStartupStep : DataHubStartupStep failed. org.apache.kafka.common.errors.SerializationException: Error serializing Avro message Caused by: java.net.ConnectException: Connection refused (Connection refused) at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na] at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na] at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na] at java.base/sun.net.www.http.HttpClient.(HttpClient.java:275) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na] at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na] at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na] at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na] at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na] at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na] at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12] at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na] at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na] at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na] 2023-06-26 13:50:46.747 ERROR 1 --- [ main] i.c.k.s.client.rest.RestService : Failed to send HTTP request to endpoint: http://localhost:8081/subjects/DataHubUpgradeHistory_v1-value/versions java.net.ConnectException: Connection refused (Connection refused) at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na] at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na] at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na] at java.base/sun.net.www.http.HttpClient.(HttpClient.java:275) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na] at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na] at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na] at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na] at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na] at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na] at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12] at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na] at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na] at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na] 2023-06-26 13:50:46.748 ERROR 1 --- [ main] c.l.d.u.s.e.steps.DataHubStartupStep : DataHubStartupStep failed. org.apache.kafka.common.errors.SerializationException: Error serializing Avro message Caused by: java.net.ConnectException: Connection refused (Connection refused) at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na] at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na] at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na] at java.base/sun.net.www.http.HttpClient.(HttpClient.java:275) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na] at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na] at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na] at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na] at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na] at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na] at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12] at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na] at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na] at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na] 2023-06-26 13:50:46.749 ERROR 1 --- [ main] i.c.k.s.client.rest.RestService : Failed to send HTTP request to endpoint: http://localhost:8081/subjects/DataHubUpgradeHistory_v1-value/versions java.net.ConnectException: Connection refused (Connection refused) at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na] at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na] at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na] at java.base/sun.net.www.http.HttpClient.(HttpClient.java:275) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na] at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na] at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na] at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na] at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na] at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na] at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12] at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na] at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na] at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na] 2023-06-26 13:50:46.750 ERROR 1 --- [ main] c.l.d.u.s.e.steps.DataHubStartupStep : DataHubStartupStep failed. org.apache.kafka.common.errors.SerializationException: Error serializing Avro message Caused by: java.net.ConnectException: Connection refused (Connection refused) at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na] at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na] at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na] at java.base/sun.net.www.http.HttpClient.(HttpClient.java:275) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na] at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na] at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na] at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na] at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na] at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na] at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12] at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na] at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na] at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na] 2023-06-26 13:50:46.751 ERROR 1 --- [ main] i.c.k.s.client.rest.RestService : Failed to send HTTP request to endpoint: http://localhost:8081/subjects/DataHubUpgradeHistory_v1-value/versions java.net.ConnectException: Connection refused (Connection refused) at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na] at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na] at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na] at java.base/sun.net.www.http.HttpClient.(HttpClient.java:275) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na] at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na] at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na] at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na] at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na] at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na] at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12] at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na] at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na] at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na] 2023-06-26 13:50:46.752 ERROR 1 --- [ main] c.l.d.u.s.e.steps.DataHubStartupStep : DataHubStartupStep failed. org.apache.kafka.common.errors.SerializationException: Error serializing Avro message Caused by: java.net.ConnectException: Connection refused (Connection refused) at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na] at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na] at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na] at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na] at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na] at java.base/sun.net.www.http.HttpClient.(HttpClient.java:275) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na] at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na] at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na] at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na] at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na] at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na] at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na] at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na] at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na] at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na] at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na] at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12] at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12] at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na] at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na] at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na] at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na] at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na] at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na] Failed Step 4/5: DataHubStartupStep. Failed after 3 retries. Exiting upgrade SystemUpdate with failure. Upgrade SystemUpdate completed with result FAILED. Exiting... 2023-06-26 13:50:46.842 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutdown requested 2023-06-26 13:50:46.842 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutting down 2023-06-26 13:50:46.845 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools 2023-06-26 13:50:46.847 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown 2023-06-26 13:50:46.847 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels 2023-06-26 13:50:46.942 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools 2023-06-26 13:50:46.942 INFO 1 --- [ Event Loop-3-1] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete 2023-06-26 13:50:46.942 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown 2023-06-26 13:50:46.942 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels 2023-06-26 13:50:46.942 INFO 1 --- [ Event Loop-3-2] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete 2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutdown requested 2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutting down 2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools 2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown 2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels 2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools 2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown 2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels 2023-06-26 13:50:46.943 INFO 1 --- [ Event Loop-3-1] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete 2023-06-26 13:50:46.943 INFO 1 --- [ Event Loop-3-2] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete 2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutdown requested 2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutting down 2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools 2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown 2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels 2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools 2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown 2023-06-26 13:50:47.247 INFO 1 --- [ Event Loop-1-1] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete 2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels 2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutdown requested 2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutting down 2023-06-26 13:50:47.247 INFO 1 --- [ Event Loop-1-2] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete 2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools 2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown 2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels 2023-06-26 13:50:47.248 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools 2023-06-26 13:50:47.248 INFO 1 --- [ Event Loop-1-1] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete 2023-06-26 13:50:47.248 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown 2023-06-26 13:50:47.248 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels 2023-06-26 13:50:47.248 INFO 1 --- [ Event Loop-1-2] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete 2023-06-26 13:50:47.248 INFO 1 --- [ionShutdownHook] o.a.k.clients.producer.KafkaProducer : [Producer clientId=producer-1] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. ANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.7.2ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.7.2ANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.7.2ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.7.2