From 6c41fa792c0c089b252fd72db8701c5a8cf6a0a4 Mon Sep 17 00:00:00 2001 From: gulnaz Date: Thu, 19 Sep 2024 11:24:22 +0300 Subject: [PATCH] SUPPORT-8553: fix properties --- .../input/InputKafkaConsumerConfig.java | 6 ++-- .../output/OutputKafkaConsumerConfig.java | 6 ++-- .../output/OutputKafkaProducerConfig.java | 2 +- .../config/output/OutputKafkaTopicConfig.java | 4 +-- .../ervu/av/service/FileUploadService.java | 8 ++--- .../service/ReceiveScanReportRetryable.java | 2 +- src/main/resources/application.properties | 30 +++++++++---------- 7 files changed, 29 insertions(+), 29 deletions(-) diff --git a/src/main/java/ru/micord/ervu/av/kafka/config/input/InputKafkaConsumerConfig.java b/src/main/java/ru/micord/ervu/av/kafka/config/input/InputKafkaConsumerConfig.java index 7d1f8cc..7c8f597 100644 --- a/src/main/java/ru/micord/ervu/av/kafka/config/input/InputKafkaConsumerConfig.java +++ b/src/main/java/ru/micord/ervu/av/kafka/config/input/InputKafkaConsumerConfig.java @@ -23,7 +23,7 @@ import org.springframework.kafka.listener.ContainerProperties; @Configuration @EnableKafka public class InputKafkaConsumerConfig { - @Value("${spring.kafka.consumer.bootstrap-servers}") + @Value("${spring.kafka.consumer.bootstrap.servers}") private List bootstrapAddress; @Value("${spring.kafka.consumer.security.protocol}") private String securityProtocol; @@ -31,9 +31,9 @@ public class InputKafkaConsumerConfig { private String jaasConfig; @Value("${spring.kafka.consumer.properties.sasl.mechanism}") private String saslMechanism; - @Value("${spring.kafka.consumer.enable-auto-commit}") + @Value("${spring.kafka.consumer.enable.auto.commit}") private String enableAutoCommit; - @Value("${spring.kafka.listener.ack-mode}") + @Value("${spring.kafka.listener.ack.mode}") private String ackMode; @Bean diff --git a/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaConsumerConfig.java b/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaConsumerConfig.java index 397512c..bc5ac81 100644 --- a/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaConsumerConfig.java +++ b/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaConsumerConfig.java @@ -23,7 +23,7 @@ import org.springframework.kafka.listener.ContainerProperties; @Configuration @EnableKafka public class OutputKafkaConsumerConfig { - @Value("${spring.kafka.out.consumer.bootstrap-servers}") + @Value("${spring.kafka.out.consumer.bootstrap.servers}") private List bootstrapAddress; @Value("${spring.kafka.out.consumer.security.protocol}") private String securityProtocol; @@ -31,9 +31,9 @@ public class OutputKafkaConsumerConfig { private String jaasConfig; @Value("${spring.kafka.out.consumer.properties.sasl.mechanism}") private String saslMechanism; - @Value("${spring.kafka.out.consumer.enable-auto-commit}") + @Value("${spring.kafka.out.consumer.enable.auto.commit}") private String enableAutoCommit; - @Value("${spring.kafka.out.listener.ack-mode}") + @Value("${spring.kafka.out.listener.ack.mode}") private String ackMode; @Bean diff --git a/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaProducerConfig.java b/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaProducerConfig.java index 340fa3d..2a14b10 100644 --- a/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaProducerConfig.java +++ b/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaProducerConfig.java @@ -20,7 +20,7 @@ import org.springframework.kafka.core.ProducerFactory; */ @Configuration public class OutputKafkaProducerConfig { - @Value("${spring.kafka.producer.bootstrap-servers}") + @Value("${spring.kafka.producer.bootstrap.servers}") private List bootstrapAddress; @Value("${spring.kafka.producer.security.protocol}") private String securityProtocol; diff --git a/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaTopicConfig.java b/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaTopicConfig.java index 8fb5ef9..79f1ef8 100644 --- a/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaTopicConfig.java +++ b/src/main/java/ru/micord/ervu/av/kafka/config/output/OutputKafkaTopicConfig.java @@ -11,9 +11,9 @@ import org.springframework.kafka.config.TopicBuilder; */ @Configuration public class OutputKafkaTopicConfig { - @Value("${kafka-out.error.topic.name}") + @Value("${kafka.out.error.topic.name}") private String kafkaOutErrorTopicName; - @Value("${kafka-out.success.topic.name}") + @Value("${kafka.out.success.topic.name}") private String kafkaOutSuccessTopicName; @Bean diff --git a/src/main/java/ru/micord/ervu/av/service/FileUploadService.java b/src/main/java/ru/micord/ervu/av/service/FileUploadService.java index bd00a74..380cca8 100644 --- a/src/main/java/ru/micord/ervu/av/service/FileUploadService.java +++ b/src/main/java/ru/micord/ervu/av/service/FileUploadService.java @@ -73,7 +73,7 @@ public class FileUploadService { this.s3Service = s3Service; } - @KafkaListener(id = "${spring.kafka.consumer.group-id}", topics = "${kafka-in.topic.name}", + @KafkaListener(id = "${spring.kafka.consumer.group.id}", topics = "${kafka.in.topic.name}", containerFactory = "inputKafkaListenerContainerFactory") public void listenKafkaIn(String kafkaInMessage, Acknowledgment acknowledgment) { DownloadRequest downloadRequest = new Gson().fromJson(kafkaInMessage, DownloadRequest.class); @@ -128,8 +128,8 @@ public class FileUploadService { } } - @KafkaListener(id = "${spring.kafka.out.consumer.group-id}", - topics = "${kafka-out.response.topic.name}", + @KafkaListener(id = "${spring.kafka.out.consumer.group.id}", + topics = "${kafka.out.response.topic.name}", containerFactory = "outputKafkaListenerContainerFactory") public void listenKafkaOut(String kafkaOutResponseMessage, Acknowledgment acknowledgment) { DownloadResponse downloadResponse = new Gson().fromJson(kafkaOutResponseMessage, @@ -269,7 +269,7 @@ public class FileUploadService { } } catch (IOException e) { - // непредусмотренная ошибка доступа через http-клиент + // непредусмотренная ошибка доступа через http-клиент throw new FileUploadException(e); } } diff --git a/src/main/java/ru/micord/ervu/av/service/ReceiveScanReportRetryable.java b/src/main/java/ru/micord/ervu/av/service/ReceiveScanReportRetryable.java index 3c4f2cf..77d3838 100644 --- a/src/main/java/ru/micord/ervu/av/service/ReceiveScanReportRetryable.java +++ b/src/main/java/ru/micord/ervu/av/service/ReceiveScanReportRetryable.java @@ -22,7 +22,7 @@ import ru.micord.ervu.av.response.AvResponse; @Service public class ReceiveScanReportRetryable { @Retryable(retryFor = {RetryableException.class}, - maxAttemptsExpression = "${av.retry.max-attempts.count}", + maxAttemptsExpression = "${av.retry.max.attempts.count}", backoff = @Backoff(delayExpression = "${av.retry.delay.milliseconds}")) public AvResponse receiveScanReport(CloseableHttpClient client, HttpGet get) throws FileUploadException { diff --git a/src/main/resources/application.properties b/src/main/resources/application.properties index e411671..153cb49 100644 --- a/src/main/resources/application.properties +++ b/src/main/resources/application.properties @@ -6,24 +6,24 @@ spring.kafka.admin.properties.sasl.jaas.config=org.apache.kafka.common.security. spring.kafka.admin.properties.sasl.mechanism=SCRAM-SHA-256 # kafka out servers for admin bean #host1:port1, host2:port2 -spring.kafka.bootstrap-servers=${OUT_KAFKA_SERVERS} +spring.kafka.bootstrap.servers=${OUT_KAFKA_SERVERS} # # kafka in consumer (with possibility for default bean) #host1:port1, host2:port2 -spring.kafka.consumer.bootstrap-servers=${IN_KAFKA_SERVERS} +spring.kafka.consumer.bootstrap.servers=${IN_KAFKA_SERVERS} spring.kafka.consumer.security.protocol=SASL_PLAINTEXT #login password to set spring.kafka.consumer.properties.sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule required username="${IN_KAFKA_USERNAME}" password="${IN_KAFKA_PASSWORD}"; spring.kafka.consumer.properties.sasl.mechanism=SCRAM-SHA-256 # -spring.kafka.consumer.enable-auto-commit=false -spring.kafka.consumer.group-id=file-to-upload-consumers +spring.kafka.consumer.enable.auto.commit=false +spring.kafka.consumer.group.id=file-to-upload-consumers # kafka in listeners -spring.kafka.listener.ack-mode=MANUAL_IMMEDIATE +spring.kafka.listener.ack.mode=MANUAL_IMMEDIATE # # kafka out producer (with possibility for default bean) #host1:port1, host2:port2 -spring.kafka.producer.bootstrap-servers=${OUT_KAFKA_SERVERS} +spring.kafka.producer.bootstrap.servers=${OUT_KAFKA_SERVERS} spring.kafka.producer.security.protocol=SASL_PLAINTEXT #login password to set spring.kafka.producer.properties.sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule required username="${OUT_KAFKA_USERNAME}" password="${OUT_KAFKA_PASSWORD}"; @@ -38,26 +38,26 @@ spring.kafka.properties.sasl.mechanism=SCRAM-SHA-256 # # kafka out consumer (not for default bean creation by spring) #host1:port1, host2:port2 -spring.kafka.out.consumer.bootstrap-servers=${OUT_KAFKA_SERVERS} +spring.kafka.out.consumer.bootstrap.servers=${OUT_KAFKA_SERVERS} spring.kafka.out.consumer.security.protocol=SASL_PLAINTEXT #login password to set spring.kafka.out.consumer.properties.sasl.jaas.config=org.apache.kafka.common.security.scram.ScramLoginModule required username="${OUT_KAFKA_USERNAME}" password="${OUT_KAFKA_PASSWORD}"; spring.kafka.out.consumer.properties.sasl.mechanism=SCRAM-SHA-256 # -spring.kafka.out.consumer.enable-auto-commit=false -spring.kafka.out.consumer.group-id=response-consumers +spring.kafka.out.consumer.enable.auto.commit=false +spring.kafka.out.consumer.group.id=response-consumers # kafka out listeners -spring.kafka.out.listener.ack-mode=MANUAL_IMMEDIATE +spring.kafka.out.listener.ack.mode=MANUAL_IMMEDIATE # # -kafka-in.topic.name=${IN_KAFKA_TOPIC_NAME} -kafka-out.error.topic.name=${OUT_KAFKA_ERROR_TOPIC_NAME} -kafka-out.success.topic.name=${OUT_KAFKA_SUCCESS_TOPIC_NAME} -kafka-out.response.topic.name=${OUT_KAFKA_RESPONSE_TOPIC_NAME} +kafka.in.topic.name=${IN_KAFKA_TOPIC_NAME} +kafka.out.error.topic.name=${OUT_KAFKA_ERROR_TOPIC_NAME} +kafka.out.success.topic.name=${OUT_KAFKA_SUCCESS_TOPIC_NAME} +kafka.out.response.topic.name=${OUT_KAFKA_RESPONSE_TOPIC_NAME} # av.rest.address=${AV_REST_ADDRESS} av.first.timeout.milliseconds=${AV_FIRST_TIMEOUT_MILLISECONDS} -av.retry.max-attempts.count=${AV_RETRY_MAX_ATTEMPTS_COUNT} +av.retry.max.attempts.count=${AV_RETRY_MAX_ATTEMPTS_COUNT} av.retry.delay.milliseconds=${AV_RETRY_DELAY_MILLISECONDS} file.saving.path=${FILE_SAVING_PATH} #