Page MenuHomePhabricator
Paste P49480

datahub-upgrade container output
ActivePublic

Authored by BTullis on Jun 26 2023, 1:54 PM.
btullis@deploy1002:~$ kubectl logs -f datahub-main-system-update-job-kw8xv datahub-system-update-job -f
ERROR StatusLogger Log4j2 could not find a logging implementation. Please add log4j-core to the classpath. Using SimpleLogger to log to the console...
. ____ _ __ _ _
/\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \
( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \
\\/ ___)| |_)| | | | | || (_| | ) ) ) )
' |____| .__|_| |_|_| |_\__, | / / / /
=========|_|==============|___/=/_/_/_/
:: Spring Boot :: (v2.5.12)
2023-06-26 13:49:08.542 INFO 1 --- [ main] io.ebean.EbeanVersion : ebean version: 11.33.3
2023-06-26 13:49:08.640 INFO 1 --- [ main] io.ebean.config.properties.LoadContext : loaded properties from [application.yml]
2023-06-26 13:49:08.952 INFO 1 --- [ main] io.ebean.datasource.pool.ConnectionPool : DataSourcePool [gmsEbeanServiceConfig] autoCommit[false] transIsolation[READ_COMMITTED] min[2] max[50]
2023-06-26 13:49:10.735 INFO 1 --- [ main] io.ebean.internal.DefaultContainer : DatabasePlatform name:gmsEbeanServiceConfig platform:mysql
2023-06-26 13:49:13.551 INFO 1 --- [ main] c.l.g.f.k.s.KafkaSchemaRegistryFactory : creating schema registry config using url: http://localhost:8081
2023-06-26 13:49:13.753 INFO 1 --- [ main] o.a.k.clients.producer.ProducerConfig : ProducerConfig values:
acks = 1
batch.size = 16384
bootstrap.servers = [kafka-test1006.eqiad.wmnet:9092]
buffer.memory = 33554432
client.dns.lookup = default
client.id = producer-1
compression.type = none
connections.max.idle.ms = 540000
delivery.timeout.ms = 30000
enable.idempotence = false
interceptor.classes = []
key.serializer = class org.apache.kafka.common.serialization.StringSerializer
linger.ms = 0
max.block.ms = 60000
max.in.flight.requests.per.connection = 5
max.request.size = 1048576
metadata.max.age.ms = 300000
metadata.max.idle.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner
receive.buffer.bytes = 32768
reconnect.backoff.max.ms = 1000
reconnect.backoff.ms = 50
request.timeout.ms = 3000
retries = 3
retry.backoff.ms = 500
sasl.client.callback.handler.class = null
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.login.callback.handler.class = null
sasl.login.class = null
sasl.login.refresh.buffer.seconds = 300
sasl.login.refresh.min.period.seconds = 60
sasl.login.refresh.window.factor = 0.8
sasl.login.refresh.window.jitter = 0.05
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
security.providers = null
send.buffer.bytes = 131072
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
ssl.endpoint.identification.algorithm = https
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLS
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
transaction.timeout.ms = 60000
transactional.id = null
value.serializer = class io.confluent.kafka.serializers.KafkaAvroSerializer
2023-06-26 13:49:13.771 INFO 1 --- [ main] i.c.k.s.KafkaAvroSerializerConfig : KafkaAvroSerializerConfig values:
bearer.auth.token = [hidden]
proxy.port = -1
schema.reflection = false
auto.register.schemas = true
max.schemas.per.subject = 1000
basic.auth.credentials.source = URL
value.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy
schema.registry.url = [http://localhost:8081]
basic.auth.user.info = [hidden]
proxy.host =
use.latest.version = false
schema.registry.basic.auth.user.info = [hidden]
bearer.auth.credentials.source = STATIC_TOKEN
key.subject.name.strategy = class io.confluent.kafka.serializers.subject.TopicNameStrategy
2023-06-26 13:49:14.134 WARN 1 --- [ main] o.a.k.clients.producer.ProducerConfig : The configuration 'schema.registry.security.protocol' was supplied but isn't a known config.
2023-06-26 13:49:14.138 INFO 1 --- [ main] o.a.kafka.common.utils.AppInfoParser : Kafka version: 5.5.1-ccs
2023-06-26 13:49:14.138 INFO 1 --- [ main] o.a.kafka.common.utils.AppInfoParser : Kafka commitId: cb1873c1fdf5f5f9
2023-06-26 13:49:14.138 INFO 1 --- [ main] o.a.kafka.common.utils.AppInfoParser : Kafka startTimeMs: 1687787354135
2023-06-26 13:49:14.542 INFO 1 --- [ main] c.l.m.m.registry.ConfigEntityRegistry : Loading bare config entity registry file at /datahub/datahub-gms/resources/entity-registry.yml
2023-06-26 13:49:16.833 INFO 1 --- [ad | producer-1] org.apache.kafka.clients.Metadata : [Producer clientId=producer-1] Cluster ID: mLKVObxtRjeI8fkVzjClCA
2023-06-26 13:49:20.150 INFO 1 --- [ main] org.reflections.Reflections : Reflections took 5093 ms to scan 46 urls, producing 836 keys and 8276 values
2023-06-26 13:49:20.939 INFO 1 --- [ main] org.reflections.Reflections : Reflections took 69 ms to scan 8 urls, producing 34 keys and 72 values
2023-06-26 13:49:27.536 WARN 1 --- [ main] c.l.m.m.r.PluginEntityRegistryLoader : /etc/datahub/plugins/models directory does not exist or is not a directory. Plugin scanning will be disabled.
2023-06-26 13:49:28.057 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataHubPolicyKey schema is compatible with previous schema due to
2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataHubPolicyInfo schema is compatible with previous schema due to
2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpGroupInfo schema is compatible with previous schema due to
2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpGroupKey schema is compatible with previous schema due to
2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : globalTags schema is compatible with previous schema due to
2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : status schema is compatible with previous schema due to
2023-06-26 13:49:28.058 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlFeatureTableKey schema is compatible with previous schema due to
2023-06-26 13:49:28.059 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlModelGroupKey schema is compatible with previous schema due to
2023-06-26 13:49:28.059 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataJobKey schema is compatible with previous schema due to
2023-06-26 13:49:28.132 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : ownership schema is compatible with previous schema due to
2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : tagKey schema is compatible with previous schema due to
2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : tagProperties schema is compatible with previous schema due to
2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : schemaFieldKey schema is compatible with previous schema due to
2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : glossaryTermInfo schema is compatible with previous schema due to
2023-06-26 13:49:28.133 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : ownership schema is compatible with previous schema due to
2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : glossaryTermKey schema is compatible with previous schema due to
2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlPrimaryKeyKey schema is compatible with previous schema due to
2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dashboardKey schema is compatible with previous schema due to
2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataPlatformKey schema is compatible with previous schema due to
2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataPlatformInfo schema is compatible with previous schema due to
2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataHubRetentionConfig schema is compatible with previous schema due to
2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataHubRetentionKey schema is compatible with previous schema due to
2023-06-26 13:49:28.134 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpUserKey schema is compatible with previous schema due to
2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpUserEditableInfo schema is compatible with previous schema due to
2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpUserInfo schema is compatible with previous schema due to
2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : corpUserStatus schema is compatible with previous schema due to
2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : groupMembership schema is compatible with previous schema due to
2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : globalTags schema is compatible with previous schema due to
2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : status schema is compatible with previous schema due to
2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlModelKey schema is compatible with previous schema due to
2023-06-26 13:49:28.135 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : dataFlowKey schema is compatible with previous schema due to
2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : glossaryNodeInfo schema is compatible with previous schema due to
2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : glossaryNodeKey schema is compatible with previous schema due to
2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : ownership schema is compatible with previous schema due to
2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : status schema is compatible with previous schema due to
2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : mlFeatureKey schema is compatible with previous schema due to
2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : datasetKey schema is compatible with previous schema due to
2023-06-26 13:49:28.136 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : viewProperties schema is compatible with previous schema due to
2023-06-26 13:49:28.137 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : schemaMetadata schema is compatible with previous schema due to
2023-06-26 13:49:28.138 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : status schema is compatible with previous schema due to
2023-06-26 13:49:28.138 INFO 1 --- [ main] c.l.m.m.registry.MergedEntityRegistry : chartKey schema is compatible with previous schema due to
2023-06-26 13:49:29.044 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No scheduled executor is provided to HttpClientFactory, using it's own scheduled executor.
2023-06-26 13:49:29.045 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No callback executor is provided to HttpClientFactory, using it's own call back executor.
2023-06-26 13:49:29.045 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No Compression executor is provided to HttpClientFactory, using it's own compression executor.
2023-06-26 13:49:29.052 INFO 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : The service 'null' has been assigned to the ChannelPoolManager with key 'noSpecifiedNamePrefix 1138266797 ', http.protocolVersion=HTTP_1_1, usePipelineV2=false, requestTimeout=10000ms, streamingTimeout=-1ms
2023-06-26 13:49:36.240 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No scheduled executor is provided to HttpClientFactory, using it's own scheduled executor.
2023-06-26 13:49:36.240 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No callback executor is provided to HttpClientFactory, using it's own call back executor.
2023-06-26 13:49:36.240 WARN 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : No Compression executor is provided to HttpClientFactory, using it's own compression executor.
2023-06-26 13:49:36.241 INFO 1 --- [ main] c.l.r.t.http.client.HttpClientFactory : The service 'null' has been assigned to the ChannelPoolManager with key 'noSpecifiedNamePrefix 1138266797 ', http.protocolVersion=HTTP_1_1, usePipelineV2=false, requestTimeout=10000ms, streamingTimeout=-1ms
2023-06-26 13:49:36.749 INFO 1 --- [ main] c.l.g.f.k.s.AwsGlueSchemaRegistryFactory : Creating AWS Glue registry
2023-06-26 13:49:36.836 INFO 1 --- [ main] c.l.g.f.k.SimpleKafkaConsumerFactory : Simple KafkaListenerContainerFactory built successfully
2023-06-26 13:49:36.851 INFO 1 --- [ main] c.l.g.f.kafka.KafkaEventConsumerFactory : Event-based KafkaListenerContainerFactory built successfully. Consumers = 1
2023-06-26 13:49:41.548 WARN 1 --- [ main] c.d.p.configuration.ConfigProvider : Configuration config.yml file not found at location /etc/datahub/plugins/auth
2023-06-26 13:49:41.548 INFO 1 --- [ main] c.l.g.f.auth.AuthorizerChainFactory : Default DataHubAuthorizer is enabled. Appending it to the authorization chain.
Starting upgrade with id SystemUpdate...
Executing Step 1/5: BuildIndicesPreStep...
2023-06-26 13:49:45.338 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's diff between new mappings (left) and old mappings (right): {ldap=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {fields={delimited={analyzer=word_delimited, search_analyzer=query_word_delimited, search_quote_analyzer=quote_analyzer, type=text}, keyword={type=keyword}, ngram={analyzer=partial, doc_values=false, max_shingle_size=4, type=search_as_you_type}}, normalizer=keyword_normalizer, type=keyword}), urn=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {fields={delimited={analyzer=urn_component, search_analyzer=query_urn_component, search_quote_analyzer=quote_analyzer, type=text}, ngram={analyzer=partial_urn_component, doc_values=false, max_shingle_size=4, type=search_as_you_type}}, type=keyword})}
2023-06-26 13:49:45.534 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's an update to `analysis` settings that requires reindexing. Target: {"index":{"max_ngram_diff":17,"refresh_interval":"1s","analysis":{"filter":{"autocomplete_custom_delimiter":{"type":"word_delimiter","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"sticky_delimiter_graph":{"type":"word_delimiter_graph","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"generate_number_parts":false,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"datahub_stop_words":{"type":"stop","ignore_case":"true","stopwords":["urn","li","data","table"]},"min_length":{"type":"length","min":"3"},"stem_override":{"type":"stemmer_override","rules":["customers, customer => customer","staging => staging","production => production","urn:li:dataplatform:hive => urn:li:dataplatform:hive","hive => hive","bigquery => bigquery","big query => big query","query => query"]},"alpha_num_space":{"type":"pattern_capture","patterns":["([a-z0-9 _-]{2,})","([a-z0-9 ]{2,})","\\\"([^\\\"]*)\\\""]},"remove_quotes":{"type":"pattern_replace","pattern":"['\"]","replacement":""},"multifilter":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph,flatten_graph","lowercase,alpha_num_space,default_syn_graph,flatten_graph"]},"multifilter_graph":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph","lowercase,alpha_num_space,default_syn_graph"]},"default_syn_graph":{"type":"synonym_graph","lenient":"false","synonyms":["cac, customer acquisition cost => cac, customer, acquisition, cost","stg, staging","dev, development","prod, production","glue, athena","s3, s_3","data platform, dataplatform","bigquery, big query => bigquery, big, query"]}},"tokenizer":{"slash_tokenizer":{"type":"pattern","pattern":"[/]"},"main_tokenizer":{"type":"pattern","pattern":"[(),./:]"}},"normalizer":{"keyword_normalizer":{"filter":["lowercase","asciifolding"]}},"analyzer":{"slash_pattern":{"tokenizer":"slash_tokenizer","filter":["lowercase"]},"browse_path_hierarchy":{"tokenizer":"path_hierarchy"},"custom_keyword":{"tokenizer":"keyword","filter":["trim","lowercase","asciifolding","snowball"]},"quote_analyzer":{"tokenizer":"keyword","filter":["asciifolding","lowercase","remove_quotes","datahub_stop_words","stop","min_length"]},"word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"partial":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]},"partial_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]}}},"number_of_shards":1,"number_of_replicas":1}} Current: {"index.creation_date":"1683106114246","index.number_of_replicas":"1","index.number_of_shards":"1","index.provided_name":"staging_corpuserindex_v2","index.uuid":"3naEY6-uRnibJV9p-5HA_A","index.version.created":"135238227"}
2023-06-26 13:49:45.537 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's an update to settings that requires reindexing. Target: {"index":{"max_ngram_diff":17,"refresh_interval":"1s","analysis":{"filter":{"autocomplete_custom_delimiter":{"type":"word_delimiter","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"sticky_delimiter_graph":{"type":"word_delimiter_graph","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"generate_number_parts":false,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"datahub_stop_words":{"type":"stop","ignore_case":"true","stopwords":["urn","li","data","table"]},"min_length":{"type":"length","min":"3"},"stem_override":{"type":"stemmer_override","rules":["customers, customer => customer","staging => staging","production => production","urn:li:dataplatform:hive => urn:li:dataplatform:hive","hive => hive","bigquery => bigquery","big query => big query","query => query"]},"alpha_num_space":{"type":"pattern_capture","patterns":["([a-z0-9 _-]{2,})","([a-z0-9 ]{2,})","\\\"([^\\\"]*)\\\""]},"remove_quotes":{"type":"pattern_replace","pattern":"['\"]","replacement":""},"multifilter":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph,flatten_graph","lowercase,alpha_num_space,default_syn_graph,flatten_graph"]},"multifilter_graph":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph","lowercase,alpha_num_space,default_syn_graph"]},"default_syn_graph":{"type":"synonym_graph","lenient":"false","synonyms":["cac, customer acquisition cost => cac, customer, acquisition, cost","stg, staging","dev, development","prod, production","glue, athena","s3, s_3","data platform, dataplatform","bigquery, big query => bigquery, big, query"]}},"tokenizer":{"slash_tokenizer":{"type":"pattern","pattern":"[/]"},"main_tokenizer":{"type":"pattern","pattern":"[(),./:]"}},"normalizer":{"keyword_normalizer":{"filter":["lowercase","asciifolding"]}},"analyzer":{"slash_pattern":{"tokenizer":"slash_tokenizer","filter":["lowercase"]},"browse_path_hierarchy":{"tokenizer":"path_hierarchy"},"custom_keyword":{"tokenizer":"keyword","filter":["trim","lowercase","asciifolding","snowball"]},"quote_analyzer":{"tokenizer":"keyword","filter":["asciifolding","lowercase","remove_quotes","datahub_stop_words","stop","min_length"]},"word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"partial":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]},"partial_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]}}},"number_of_shards":1,"number_of_replicas":1}} Current: {"index.creation_date":"1683106114246","index.number_of_replicas":"1","index.number_of_shards":"1","index.provided_name":"staging_corpuserindex_v2","index.uuid":"3naEY6-uRnibJV9p-5HA_A","index.version.created":"135238227"}
2023-06-26 13:49:45.735 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_system_metadata_service_v1 - There's diff between new mappings (left) and old mappings (right): {aspect=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword}), runId=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword}), urn=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword})}
2023-06-26 13:49:46.032 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPreStep : Updated index staging_corpuserindex_v2 with new settings. Settings: {index.blocks.write=true}, Acknowledged: true
2023-06-26 13:49:46.573 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPreStep : Cloned index staging_corpuserindex_v2 into staging_corpuserindex_v2_clone_1687787386034, Acknowledged: true
2023-06-26 13:49:46.722 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPreStep : Updated index staging_system_metadata_service_v1 with new settings. Settings: {index.blocks.write=true}, Acknowledged: true
2023-06-26 13:49:47.128 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPreStep : Cloned index staging_system_metadata_service_v1 into staging_system_metadata_service_v1_clone_1687787386722, Acknowledged: true
Completed Step 1/5: BuildIndicesPreStep successfully.
Executing Step 2/5: BuildIndicesStep...
2023-06-26 13:49:47.130 INFO 1 --- [ main] c.l.m.g.e.ElasticSearchGraphService : Setting up elastic graph index
2023-06-26 13:49:47.132 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_graph_service_v1 does not exist. Creating
2023-06-26 13:49:47.447 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_graph_service_v1
2023-06-26 13:49:47.738 ERROR 1 --- [pool-9-thread-1] c.l.m.s.elasticsearch.query.ESSearchDAO : Search query failed
org.elasticsearch.ElasticsearchStatusException: Elasticsearch exception [type=index_not_found_exception, reason=no such index [staging_datahubpolicyindex_v2]]
at org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:187) ~[elasticsearch-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.parseEntity(RestHighLevelClient.java:1911) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.parseResponseException(RestHighLevelClient.java:1888) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1645) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.performRequest(RestHighLevelClient.java:1602) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.performRequestAndParseEntity(RestHighLevelClient.java:1572) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.search(RestHighLevelClient.java:1088) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.executeAndExtract(ESSearchDAO.java:57) ~[metadata-io.jar!/:na]
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.search(ESSearchDAO.java:90) ~[metadata-io.jar!/:na]
at com.linkedin.metadata.search.elasticsearch.ElasticSearchService.fullTextSearch(ElasticSearchService.java:111) ~[metadata-io.jar!/:na]
at com.linkedin.metadata.client.JavaEntityClient.search(JavaEntityClient.java:312) ~[metadata-io.jar!/:na]
at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:50) ~[auth-impl.jar!/:na]
at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:42) ~[auth-impl.jar!/:na]
at com.datahub.authorization.DataHubAuthorizer$PolicyRefreshRunnable.run(DataHubAuthorizer.java:223) ~[auth-impl.jar!/:na]
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) ~[na:na]
at java.base/java.util.concurrent.FutureTask.runAndReset(FutureTask.java:305) ~[na:na]
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:305) ~[na:na]
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[na:na]
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[na:na]
at java.base/java.lang.Thread.run(Thread.java:829) ~[na:na]
Suppressed: org.elasticsearch.client.ResponseException: method [POST], host [http://datahubsearch.svc.eqiad.wmnet:9200], URI [/staging_datahubpolicyindex_v2/_search?typed_keys=true&max_concurrent_shard_requests=5&ignore_unavailable=false&expand_wildcards=open&allow_no_indices=true&ignore_throttled=true&search_type=query_then_fetch&batched_reduce_size=512&ccs_minimize_roundtrips=true], status line [HTTP/1.1 404 Not Found]
{"error":{"root_cause":[{"type":"index_not_found_exception","reason":"no such index [staging_datahubpolicyindex_v2]","index":"staging_datahubpolicyindex_v2","resource.id":"staging_datahubpolicyindex_v2","resource.type":"index_or_alias","index_uuid":"_na_"}],"type":"index_not_found_exception","reason":"no such index [staging_datahubpolicyindex_v2]","index":"staging_datahubpolicyindex_v2","resource.id":"staging_datahubpolicyindex_v2","resource.type":"index_or_alias","index_uuid":"_na_"},"status":404}
at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:326) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:296) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:270) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1632) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
... 16 common frames omitted
2023-06-26 13:49:47.738 ERROR 1 --- [pool-9-thread-1] c.d.authorization.DataHubAuthorizer : Failed to retrieve policy urns! Skipping updating policy cache until next refresh. start: 0, count: 30
com.datahub.util.exception.ESQueryException: Search query failed:
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.executeAndExtract(ESSearchDAO.java:62) ~[metadata-io.jar!/:na]
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.search(ESSearchDAO.java:90) ~[metadata-io.jar!/:na]
at com.linkedin.metadata.search.elasticsearch.ElasticSearchService.fullTextSearch(ElasticSearchService.java:111) ~[metadata-io.jar!/:na]
at com.linkedin.metadata.client.JavaEntityClient.search(JavaEntityClient.java:312) ~[metadata-io.jar!/:na]
at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:50) ~[auth-impl.jar!/:na]
at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:42) ~[auth-impl.jar!/:na]
at com.datahub.authorization.DataHubAuthorizer$PolicyRefreshRunnable.run(DataHubAuthorizer.java:223) ~[auth-impl.jar!/:na]
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) ~[na:na]
at java.base/java.util.concurrent.FutureTask.runAndReset(FutureTask.java:305) ~[na:na]
at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:305) ~[na:na]
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[na:na]
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[na:na]
at java.base/java.lang.Thread.run(Thread.java:829) ~[na:na]
Caused by: org.elasticsearch.ElasticsearchStatusException: Elasticsearch exception [type=index_not_found_exception, reason=no such index [staging_datahubpolicyindex_v2]]
at org.elasticsearch.rest.BytesRestResponse.errorFromXContent(BytesRestResponse.java:187) ~[elasticsearch-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.parseEntity(RestHighLevelClient.java:1911) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.parseResponseException(RestHighLevelClient.java:1888) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1645) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.performRequest(RestHighLevelClient.java:1602) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.performRequestAndParseEntity(RestHighLevelClient.java:1572) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.search(RestHighLevelClient.java:1088) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.executeAndExtract(ESSearchDAO.java:57) ~[metadata-io.jar!/:na]
... 12 common frames omitted
Suppressed: org.elasticsearch.client.ResponseException: method [POST], host [http://datahubsearch.svc.eqiad.wmnet:9200], URI [/staging_datahubpolicyindex_v2/_search?typed_keys=true&max_concurrent_shard_requests=5&ignore_unavailable=false&expand_wildcards=open&allow_no_indices=true&ignore_throttled=true&search_type=query_then_fetch&batched_reduce_size=512&ccs_minimize_roundtrips=true], status line [HTTP/1.1 404 Not Found]
{"error":{"root_cause":[{"type":"index_not_found_exception","reason":"no such index [staging_datahubpolicyindex_v2]","index":"staging_datahubpolicyindex_v2","resource.id":"staging_datahubpolicyindex_v2","resource.type":"index_or_alias","index_uuid":"_na_"}],"type":"index_not_found_exception","reason":"no such index [staging_datahubpolicyindex_v2]","index":"staging_datahubpolicyindex_v2","resource.id":"staging_datahubpolicyindex_v2","resource.type":"index_or_alias","index_uuid":"_na_"},"status":404}
at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:326) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:296) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:270) ~[elasticsearch-rest-client-7.10.2.jar!/:7.10.2]
at org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1632) ~[elasticsearch-rest-high-level-client-7.10.2.jar!/:7.10.2]
... 16 common frames omitted
2023-06-26 13:49:47.749 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's diff between new mappings (left) and old mappings (right): {ldap=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {fields={delimited={analyzer=word_delimited, search_analyzer=query_word_delimited, search_quote_analyzer=quote_analyzer, type=text}, keyword={type=keyword}, ngram={analyzer=partial, doc_values=false, max_shingle_size=4, type=search_as_you_type}}, normalizer=keyword_normalizer, type=keyword}), urn=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {fields={delimited={analyzer=urn_component, search_analyzer=query_urn_component, search_quote_analyzer=quote_analyzer, type=text}, ngram={analyzer=partial_urn_component, doc_values=false, max_shingle_size=4, type=search_as_you_type}}, type=keyword})}
2023-06-26 13:49:47.750 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's an update to `analysis` settings that requires reindexing. Target: {"index":{"max_ngram_diff":17,"refresh_interval":"1s","analysis":{"filter":{"autocomplete_custom_delimiter":{"type":"word_delimiter","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"sticky_delimiter_graph":{"type":"word_delimiter_graph","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"generate_number_parts":false,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"datahub_stop_words":{"type":"stop","ignore_case":"true","stopwords":["urn","li","data","table"]},"min_length":{"type":"length","min":"3"},"stem_override":{"type":"stemmer_override","rules":["customers, customer => customer","staging => staging","production => production","urn:li:dataplatform:hive => urn:li:dataplatform:hive","hive => hive","bigquery => bigquery","big query => big query","query => query"]},"alpha_num_space":{"type":"pattern_capture","patterns":["([a-z0-9 _-]{2,})","([a-z0-9 ]{2,})","\\\"([^\\\"]*)\\\""]},"remove_quotes":{"type":"pattern_replace","pattern":"['\"]","replacement":""},"multifilter":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph,flatten_graph","lowercase,alpha_num_space,default_syn_graph,flatten_graph"]},"multifilter_graph":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph","lowercase,alpha_num_space,default_syn_graph"]},"default_syn_graph":{"type":"synonym_graph","lenient":"false","synonyms":["cac, customer acquisition cost => cac, customer, acquisition, cost","stg, staging","dev, development","prod, production","glue, athena","s3, s_3","data platform, dataplatform","bigquery, big query => bigquery, big, query"]}},"tokenizer":{"slash_tokenizer":{"type":"pattern","pattern":"[/]"},"main_tokenizer":{"type":"pattern","pattern":"[(),./:]"}},"normalizer":{"keyword_normalizer":{"filter":["lowercase","asciifolding"]}},"analyzer":{"slash_pattern":{"tokenizer":"slash_tokenizer","filter":["lowercase"]},"browse_path_hierarchy":{"tokenizer":"path_hierarchy"},"custom_keyword":{"tokenizer":"keyword","filter":["trim","lowercase","asciifolding","snowball"]},"quote_analyzer":{"tokenizer":"keyword","filter":["asciifolding","lowercase","remove_quotes","datahub_stop_words","stop","min_length"]},"word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"partial":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]},"partial_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]}}},"number_of_shards":1,"number_of_replicas":1}} Current: {"index.blocks.write":"true","index.creation_date":"1683106114246","index.number_of_replicas":"1","index.number_of_shards":"1","index.provided_name":"staging_corpuserindex_v2","index.uuid":"3naEY6-uRnibJV9p-5HA_A","index.version.created":"135238227"}
2023-06-26 13:49:47.750 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_corpuserindex_v2 - There's an update to settings that requires reindexing. Target: {"index":{"max_ngram_diff":17,"refresh_interval":"1s","analysis":{"filter":{"autocomplete_custom_delimiter":{"type":"word_delimiter","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"sticky_delimiter_graph":{"type":"word_delimiter_graph","split_on_numerics":false,"split_on_case_change":false,"preserve_original":true,"generate_number_parts":false,"type_table":[": => SUBWORD_DELIM","_ => ALPHANUM","- => ALPHA"]},"datahub_stop_words":{"type":"stop","ignore_case":"true","stopwords":["urn","li","data","table"]},"min_length":{"type":"length","min":"3"},"stem_override":{"type":"stemmer_override","rules":["customers, customer => customer","staging => staging","production => production","urn:li:dataplatform:hive => urn:li:dataplatform:hive","hive => hive","bigquery => bigquery","big query => big query","query => query"]},"alpha_num_space":{"type":"pattern_capture","patterns":["([a-z0-9 _-]{2,})","([a-z0-9 ]{2,})","\\\"([^\\\"]*)\\\""]},"remove_quotes":{"type":"pattern_replace","pattern":"['\"]","replacement":""},"multifilter":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph,flatten_graph","lowercase,alpha_num_space,default_syn_graph,flatten_graph"]},"multifilter_graph":{"type":"multiplexer","filters":["lowercase,sticky_delimiter_graph","lowercase,alpha_num_space,default_syn_graph"]},"default_syn_graph":{"type":"synonym_graph","lenient":"false","synonyms":["cac, customer acquisition cost => cac, customer, acquisition, cost","stg, staging","dev, development","prod, production","glue, athena","s3, s_3","data platform, dataplatform","bigquery, big query => bigquery, big, query"]}},"tokenizer":{"slash_tokenizer":{"type":"pattern","pattern":"[/]"},"main_tokenizer":{"type":"pattern","pattern":"[(),./:]"}},"normalizer":{"keyword_normalizer":{"filter":["lowercase","asciifolding"]}},"analyzer":{"slash_pattern":{"tokenizer":"slash_tokenizer","filter":["lowercase"]},"browse_path_hierarchy":{"tokenizer":"path_hierarchy"},"custom_keyword":{"tokenizer":"keyword","filter":["trim","lowercase","asciifolding","snowball"]},"quote_analyzer":{"tokenizer":"keyword","filter":["asciifolding","lowercase","remove_quotes","datahub_stop_words","stop","min_length"]},"word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_word_delimited":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"query_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","multifilter_graph","lowercase","datahub_stop_words","stop","stem_override","snowball","remove_quotes","unique","min_length"]},"partial":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]},"partial_urn_component":{"tokenizer":"main_tokenizer","filter":["asciifolding","autocomplete_custom_delimiter","lowercase","datahub_stop_words"]}}},"number_of_shards":1,"number_of_replicas":1}} Current: {"index.blocks.write":"true","index.creation_date":"1683106114246","index.number_of_replicas":"1","index.number_of_shards":"1","index.provided_name":"staging_corpuserindex_v2","index.uuid":"3naEY6-uRnibJV9p-5HA_A","index.version.created":"135238227"}
2023-06-26 13:49:47.855 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_containerindex_v2 does not exist. Creating
2023-06-26 13:49:48.268 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_containerindex_v2
2023-06-26 13:49:48.268 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubpolicyindex_v2 does not exist. Creating
2023-06-26 13:49:48.716 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubpolicyindex_v2
2023-06-26 13:49:48.716 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubaccesstokenindex_v2 does not exist. Creating
2023-06-26 13:49:49.079 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubaccesstokenindex_v2
2023-06-26 13:49:49.079 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubupgradeindex_v2 does not exist. Creating
2023-06-26 13:49:49.443 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubupgradeindex_v2
2023-06-26 13:49:49.444 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubroleindex_v2 does not exist. Creating
2023-06-26 13:49:49.790 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubroleindex_v2
2023-06-26 13:49:49.791 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_corpgroupindex_v2 does not exist. Creating
2023-06-26 13:49:50.118 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_corpgroupindex_v2
2023-06-26 13:49:50.119 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubstepstateindex_v2 does not exist. Creating
2023-06-26 13:49:50.451 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubstepstateindex_v2
2023-06-26 13:49:50.451 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_globalsettingsindex_v2 does not exist. Creating
2023-06-26 13:49:50.853 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_globalsettingsindex_v2
2023-06-26 13:49:50.853 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataprocessindex_v2 does not exist. Creating
2023-06-26 13:49:51.169 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataprocessindex_v2
2023-06-26 13:49:51.169 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlfeaturetableindex_v2 does not exist. Creating
2023-06-26 13:49:51.513 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlfeaturetableindex_v2
2023-06-26 13:49:51.513 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlmodelgroupindex_v2 does not exist. Creating
2023-06-26 13:49:51.824 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlmodelgroupindex_v2
2023-06-26 13:49:51.825 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_postindex_v2 does not exist. Creating
2023-06-26 13:49:52.130 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_postindex_v2
2023-06-26 13:49:52.131 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubexecutionrequestindex_v2 does not exist. Creating
2023-06-26 13:49:52.459 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubexecutionrequestindex_v2
2023-06-26 13:49:52.459 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_invitetokenindex_v2 does not exist. Creating
2023-06-26 13:49:52.826 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_invitetokenindex_v2
2023-06-26 13:49:52.826 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datajobindex_v2 does not exist. Creating
2023-06-26 13:49:53.185 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datajobindex_v2
2023-06-26 13:49:53.185 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_assertionindex_v2 does not exist. Creating
2023-06-26 13:49:53.575 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_assertionindex_v2
2023-06-26 13:49:53.575 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataplatforminstanceindex_v2 does not exist. Creating
2023-06-26 13:49:53.934 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataplatforminstanceindex_v2
2023-06-26 13:49:53.934 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_tagindex_v2 does not exist. Creating
2023-06-26 13:49:54.294 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_tagindex_v2
2023-06-26 13:49:54.295 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_schemafieldindex_v2 does not exist. Creating
2023-06-26 13:49:54.598 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_schemafieldindex_v2
2023-06-26 13:49:54.598 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_glossarytermindex_v2 does not exist. Creating
2023-06-26 13:49:54.964 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_glossarytermindex_v2
2023-06-26 13:49:54.964 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlprimarykeyindex_v2 does not exist. Creating
2023-06-26 13:49:55.270 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlprimarykeyindex_v2
2023-06-26 13:49:55.270 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dashboardindex_v2 does not exist. Creating
2023-06-26 13:49:55.579 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dashboardindex_v2
2023-06-26 13:49:55.579 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_notebookindex_v2 does not exist. Creating
2023-06-26 13:49:55.931 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_notebookindex_v2
2023-06-26 13:49:55.931 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlmodeldeploymentindex_v2 does not exist. Creating
2023-06-26 13:49:56.276 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlmodeldeploymentindex_v2
2023-06-26 13:49:56.276 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataplatformindex_v2 does not exist. Creating
2023-06-26 13:49:56.653 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataplatformindex_v2
2023-06-26 13:49:56.653 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubretentionindex_v2 does not exist. Creating
2023-06-26 13:49:56.970 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubretentionindex_v2
2023-06-26 13:49:57.011 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_corpuserindex_v2_1687787396970 does not exist. Creating
2023-06-26 13:49:57.290 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_corpuserindex_v2_1687787396970
2023-06-26 13:50:17.351 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Task: 50uKScxYREyFVlCeM4NwpQ:131447489 - Reindexing from staging_corpuserindex_v2 to staging_corpuserindex_v2_1687787396970 in progress...
2023-06-26 13:50:17.361 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Task: 50uKScxYREyFVlCeM4NwpQ:131447489 - Reindexing staging_corpuserindex_v2 to staging_corpuserindex_v2_1687787396970 task was successful
2023-06-26 13:50:17.361 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Reindex from staging_corpuserindex_v2 to staging_corpuserindex_v2_1687787396970 succeeded
2023-06-26 13:50:17.606 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Finished setting up staging_corpuserindex_v2
2023-06-26 13:50:17.606 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_testindex_v2 does not exist. Creating
2023-06-26 13:50:17.824 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_testindex_v2
2023-06-26 13:50:17.824 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlmodelindex_v2 does not exist. Creating
2023-06-26 13:50:18.238 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlmodelindex_v2
2023-06-26 13:50:18.238 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataflowindex_v2 does not exist. Creating
2023-06-26 13:50:18.687 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataflowindex_v2
2023-06-26 13:50:18.687 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_mlfeatureindex_v2 does not exist. Creating
2023-06-26 13:50:19.021 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_mlfeatureindex_v2
2023-06-26 13:50:19.021 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_glossarynodeindex_v2 does not exist. Creating
2023-06-26 13:50:19.382 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_glossarynodeindex_v2
2023-06-26 13:50:19.383 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubingestionsourceindex_v2 does not exist. Creating
2023-06-26 13:50:19.736 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubingestionsourceindex_v2
2023-06-26 13:50:19.736 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_domainindex_v2 does not exist. Creating
2023-06-26 13:50:20.094 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_domainindex_v2
2023-06-26 13:50:20.094 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubviewindex_v2 does not exist. Creating
2023-06-26 13:50:20.429 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubviewindex_v2
2023-06-26 13:50:20.429 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_telemetryindex_v2 does not exist. Creating
2023-06-26 13:50:20.853 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_telemetryindex_v2
2023-06-26 13:50:20.853 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datahubsecretindex_v2 does not exist. Creating
2023-06-26 13:50:21.131 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datahubsecretindex_v2
2023-06-26 13:50:21.131 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_chartindex_v2 does not exist. Creating
2023-06-26 13:50:21.430 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_chartindex_v2
2023-06-26 13:50:21.430 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datasetindex_v2 does not exist. Creating
2023-06-26 13:50:21.735 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datasetindex_v2
2023-06-26 13:50:21.735 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataprocessinstanceindex_v2 does not exist. Creating
2023-06-26 13:50:22.089 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataprocessinstanceindex_v2
2023-06-26 13:50:22.090 INFO 1 --- [ main] l.m.s.ElasticSearchSystemMetadataService : Setting up system metadata index
2023-06-26 13:50:22.098 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ReindexConfig : Index: staging_system_metadata_service_v1 - There's diff between new mappings (left) and old mappings (right): {aspect=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword}), runId=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword}), urn=({fields={keyword={ignore_above=256, type=keyword}}, type=text}, {type=keyword})}
2023-06-26 13:50:22.104 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_system_metadata_service_v1_1687787422099 does not exist. Creating
2023-06-26 13:50:22.414 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_system_metadata_service_v1_1687787422099
2023-06-26 13:50:42.443 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Task: 50uKScxYREyFVlCeM4NwpQ:131448073 - Reindexing from staging_system_metadata_service_v1 to staging_system_metadata_service_v1_1687787422099 in progress...
2023-06-26 13:50:42.454 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Task: 50uKScxYREyFVlCeM4NwpQ:131448073 - Reindexing staging_system_metadata_service_v1 to staging_system_metadata_service_v1_1687787422099 task was successful
2023-06-26 13:50:42.454 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Reindex from staging_system_metadata_service_v1 to staging_system_metadata_service_v1_1687787422099 succeeded
2023-06-26 13:50:42.647 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Finished setting up staging_system_metadata_service_v1
2023-06-26 13:50:42.662 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datajob_datahubingestionrunsummaryaspect_v1 does not exist. Creating
2023-06-26 13:50:42.929 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datajob_datahubingestionrunsummaryaspect_v1
2023-06-26 13:50:42.929 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_datajob_datahubingestioncheckpointaspect_v1 does not exist. Creating
2023-06-26 13:50:43.242 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_datajob_datahubingestioncheckpointaspect_v1
2023-06-26 13:50:43.242 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_assertion_assertionruneventaspect_v1 does not exist. Creating
2023-06-26 13:50:43.528 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_assertion_assertionruneventaspect_v1
2023-06-26 13:50:43.528 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dashboard_dashboardusagestatisticsaspect_v1 does not exist. Creating
2023-06-26 13:50:43.804 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dashboard_dashboardusagestatisticsaspect_v1
2023-06-26 13:50:43.804 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_chart_chartusagestatisticsaspect_v1 does not exist. Creating
2023-06-26 13:50:44.114 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_chart_chartusagestatisticsaspect_v1
2023-06-26 13:50:44.115 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataset_datasetprofileaspect_v1 does not exist. Creating
2023-06-26 13:50:44.427 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataset_datasetprofileaspect_v1
2023-06-26 13:50:44.427 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataset_datasetusagestatisticsaspect_v1 does not exist. Creating
2023-06-26 13:50:44.725 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataset_datasetusagestatisticsaspect_v1
2023-06-26 13:50:44.725 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataset_operationaspect_v1 does not exist. Creating
2023-06-26 13:50:45.032 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataset_operationaspect_v1
2023-06-26 13:50:45.032 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Index staging_dataprocessinstance_dataprocessinstanceruneventaspect_v1 does not exist. Creating
2023-06-26 13:50:45.342 INFO 1 --- [ main] c.l.m.s.e.indexbuilder.ESIndexBuilder : Created index staging_dataprocessinstance_dataprocessinstanceruneventaspect_v1
Completed Step 2/5: BuildIndicesStep successfully.
Executing Step 3/5: BuildIndicesPostStep...
2023-06-26 13:50:45.489 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPostStep : Updated index staging_corpuserindex_v2 with new settings. Settings: {index.blocks.write=false}, Acknowledged: true
2023-06-26 13:50:45.600 INFO 1 --- [ main] c.l.d.u.s.e.steps.BuildIndicesPostStep : Updated index staging_system_metadata_service_v1 with new settings. Settings: {index.blocks.write=false}, Acknowledged: true
Completed Step 3/5: BuildIndicesPostStep successfully.
Executing Step 4/5: DataHubStartupStep...
2023-06-26 13:50:46.744 ERROR 1 --- [ main] i.c.k.s.client.rest.RestService : Failed to send HTTP request to endpoint: http://localhost:8081/subjects/DataHubUpgradeHistory_v1-value/versions
java.net.ConnectException: Connection refused (Connection refused)
at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na]
at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na]
at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na]
at java.base/sun.net.www.http.HttpClient.<init>(HttpClient.java:275) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na]
at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na]
at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na]
at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12]
at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na]
2023-06-26 13:50:46.745 ERROR 1 --- [ main] c.l.d.u.s.e.steps.DataHubStartupStep : DataHubStartupStep failed.
org.apache.kafka.common.errors.SerializationException: Error serializing Avro message
Caused by: java.net.ConnectException: Connection refused (Connection refused)
at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na]
at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na]
at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na]
at java.base/sun.net.www.http.HttpClient.<init>(HttpClient.java:275) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na]
at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na]
at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na]
at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12]
at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na]
2023-06-26 13:50:46.747 ERROR 1 --- [ main] i.c.k.s.client.rest.RestService : Failed to send HTTP request to endpoint: http://localhost:8081/subjects/DataHubUpgradeHistory_v1-value/versions
java.net.ConnectException: Connection refused (Connection refused)
at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na]
at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na]
at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na]
at java.base/sun.net.www.http.HttpClient.<init>(HttpClient.java:275) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na]
at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na]
at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na]
at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12]
at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na]
2023-06-26 13:50:46.748 ERROR 1 --- [ main] c.l.d.u.s.e.steps.DataHubStartupStep : DataHubStartupStep failed.
org.apache.kafka.common.errors.SerializationException: Error serializing Avro message
Caused by: java.net.ConnectException: Connection refused (Connection refused)
at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na]
at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na]
at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na]
at java.base/sun.net.www.http.HttpClient.<init>(HttpClient.java:275) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na]
at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na]
at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na]
at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12]
at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na]
2023-06-26 13:50:46.749 ERROR 1 --- [ main] i.c.k.s.client.rest.RestService : Failed to send HTTP request to endpoint: http://localhost:8081/subjects/DataHubUpgradeHistory_v1-value/versions
java.net.ConnectException: Connection refused (Connection refused)
at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na]
at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na]
at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na]
at java.base/sun.net.www.http.HttpClient.<init>(HttpClient.java:275) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na]
at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na]
at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na]
at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12]
at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na]
2023-06-26 13:50:46.750 ERROR 1 --- [ main] c.l.d.u.s.e.steps.DataHubStartupStep : DataHubStartupStep failed.
org.apache.kafka.common.errors.SerializationException: Error serializing Avro message
Caused by: java.net.ConnectException: Connection refused (Connection refused)
at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na]
at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na]
at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na]
at java.base/sun.net.www.http.HttpClient.<init>(HttpClient.java:275) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na]
at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na]
at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na]
at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12]
at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na]
2023-06-26 13:50:46.751 ERROR 1 --- [ main] i.c.k.s.client.rest.RestService : Failed to send HTTP request to endpoint: http://localhost:8081/subjects/DataHubUpgradeHistory_v1-value/versions
java.net.ConnectException: Connection refused (Connection refused)
at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na]
at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na]
at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na]
at java.base/sun.net.www.http.HttpClient.<init>(HttpClient.java:275) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na]
at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na]
at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na]
at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12]
at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na]
2023-06-26 13:50:46.752 ERROR 1 --- [ main] c.l.d.u.s.e.steps.DataHubStartupStep : DataHubStartupStep failed.
org.apache.kafka.common.errors.SerializationException: Error serializing Avro message
Caused by: java.net.ConnectException: Connection refused (Connection refused)
at java.base/java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:412) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:255) ~[na:na]
at java.base/java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:237) ~[na:na]
at java.base/java.net.Socket.connect(Socket.java:609) ~[na:na]
at java.base/sun.net.NetworkClient.doConnect(NetworkClient.java:177) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:507) ~[na:na]
at java.base/sun.net.www.http.HttpClient.openServer(HttpClient.java:602) ~[na:na]
at java.base/sun.net.www.http.HttpClient.<init>(HttpClient.java:275) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:374) ~[na:na]
at java.base/sun.net.www.http.HttpClient.New(HttpClient.java:395) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1253) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1187) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1081) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:1015) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367) ~[na:na]
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342) ~[na:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244) ~[kafka-schema-registry-client-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59) ~[kafka-avro-serializer-5.5.1.jar!/:na]
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862) ~[kafka-clients-5.5.1-ccs.jar!/:na]
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceDataHubUpgradeHistoryEvent(KafkaEventProducer.java:170) ~[kafka-producer.jar!/:na]
at com.linkedin.datahub.upgrade.system.elasticsearch.steps.DataHubStartupStep.lambda$executable$0(DataHubStartupStep.java:37) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39) ~[classes!/:na]
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30) ~[classes!/:na]
at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80) ~[classes!/:na]
at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:812) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:796) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.SpringApplication.run(SpringApplication.java:346) ~[spring-boot-2.5.12.jar!/:2.5.12]
at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:143) ~[spring-boot-2.5.12.jar!/:2.5.12]
at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23) ~[classes!/:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:na]
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:na]
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:na]
at java.base/java.lang.reflect.Method.invoke(Method.java:566) ~[na:na]
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) ~[datahub-upgrade.jar:na]
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:88) ~[datahub-upgrade.jar:na]
Failed Step 4/5: DataHubStartupStep. Failed after 3 retries.
Exiting upgrade SystemUpdate with failure.
Upgrade SystemUpdate completed with result FAILED. Exiting...
2023-06-26 13:50:46.842 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutdown requested
2023-06-26 13:50:46.842 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutting down
2023-06-26 13:50:46.845 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools
2023-06-26 13:50:46.847 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown
2023-06-26 13:50:46.847 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels
2023-06-26 13:50:46.942 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools
2023-06-26 13:50:46.942 INFO 1 --- [ Event Loop-3-1] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete
2023-06-26 13:50:46.942 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown
2023-06-26 13:50:46.942 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels
2023-06-26 13:50:46.942 INFO 1 --- [ Event Loop-3-2] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete
2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutdown requested
2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutting down
2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools
2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown
2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels
2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools
2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown
2023-06-26 13:50:46.943 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels
2023-06-26 13:50:46.943 INFO 1 --- [ Event Loop-3-1] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete
2023-06-26 13:50:46.943 INFO 1 --- [ Event Loop-3-2] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete
2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutdown requested
2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutting down
2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools
2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown
2023-06-26 13:50:47.246 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels
2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools
2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown
2023-06-26 13:50:47.247 INFO 1 --- [ Event Loop-1-1] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete
2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels
2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutdown requested
2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.common.AbstractNettyClient : Shutting down
2023-06-26 13:50:47.247 INFO 1 --- [ Event Loop-1-2] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete
2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools
2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown
2023-06-26 13:50:47.247 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels
2023-06-26 13:50:47.248 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutting down 0 connection pools
2023-06-26 13:50:47.248 INFO 1 --- [ Event Loop-1-1] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete
2023-06-26 13:50:47.248 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shutdown
2023-06-26 13:50:47.248 INFO 1 --- [ionShutdownHook] c.l.r.t.h.c.c.ChannelPoolManagerImpl : All connection pools shut down, closing all channels
2023-06-26 13:50:47.248 INFO 1 --- [ Event Loop-1-2] c.l.r.t.h.c.c.ChannelPoolManagerImpl : Shutdown complete
2023-06-26 13:50:47.248 INFO 1 --- [ionShutdownHook] o.a.k.clients.producer.KafkaProducer : [Producer clientId=producer-1] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms.
ANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.7.2ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.7.2ANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.7.2ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.7.2