From 1f072d4ca24a12d9166a78cb3223b98b0db56f7f Mon Sep 17 00:00:00 2001 From: Number7 <1845377266@qq.com> Date: Mon, 30 Sep 2024 10:29:58 +0800 Subject: [PATCH] =?UTF-8?q?fix():=E4=BF=AE=E6=94=B9=E6=8A=A5=E6=96=87?= =?UTF-8?q?=E6=8E=A5=E5=8F=A3=E5=AD=97=E6=AE=B5=E9=97=AE=E9=A2=98?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../common/domain/MessageTemplateType.java | 6 +- .../common/util/KafkaCommonProperties.java | 239 ------------------ .../com/muyu/common/util/KafkaConstants.java | 16 ++ .../muyu/common/util/KafkaConsumerConfig.java | 43 ++++ .../muyu/common/util/KafkaProducerTest.java | 64 ----- .../muyu/common/util/KafkaProviderConfig.java | 38 +++ .../java/com/muyu/server/SaasApplication.java | 1 + .../com/muyu/server/config/MqttConfigure.java | 5 +- .../server/controller/TemplateController.java | 5 +- .../com/muyu/server/mapper/KafkaMapper.java | 13 + .../impl/MessageTemplateTypeServiceImpl.java | 14 +- .../service/impl/TemplateServiceImpl.java | 16 +- 12 files changed, 131 insertions(+), 329 deletions(-) delete mode 100644 cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaCommonProperties.java create mode 100644 cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaConstants.java create mode 100644 cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaConsumerConfig.java delete mode 100644 cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaProducerTest.java create mode 100644 cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaProviderConfig.java create mode 100644 cloud-modules/saas/saas-server/src/main/java/com/muyu/server/mapper/KafkaMapper.java diff --git a/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/domain/MessageTemplateType.java b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/domain/MessageTemplateType.java index 3cdd2a5..dad9090 100644 --- a/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/domain/MessageTemplateType.java +++ b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/domain/MessageTemplateType.java @@ -1,5 +1,6 @@ package com.muyu.common.domain; import com.baomidou.mybatisplus.annotation.IdType; +import com.baomidou.mybatisplus.annotation.TableField; import com.baomidou.mybatisplus.annotation.TableId; import com.baomidou.mybatisplus.annotation.TableName; import io.swagger.v3.oas.annotations.tags.Tag; @@ -7,9 +8,7 @@ import lombok.AllArgsConstructor; import lombok.Data; import lombok.NoArgsConstructor; import lombok.experimental.SuperBuilder; - import java.io.Serializable; - /** * 模版对应的配置 * @Author:liuxinyue @@ -25,8 +24,6 @@ import java.io.Serializable; @Tag(name = "报文模版表") @TableName(value = "message_template_type",autoResultMap = true) public class MessageTemplateType implements Serializable { - - /** * 主键 */ @@ -59,6 +56,7 @@ public class MessageTemplateType implements Serializable { /** * 数据类型名称 */ + @TableField(exist = false) private String dataTypeName; /** *最小值 diff --git a/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaCommonProperties.java b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaCommonProperties.java deleted file mode 100644 index a713ae1..0000000 --- a/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaCommonProperties.java +++ /dev/null @@ -1,239 +0,0 @@ -package com.muyu.common.util; - -import java.util.Collection; -import java.util.Collections; - -/** - * kafka通用配置 - * @author liuxinyue - * @Package:com.muyu.common.util - * @name:KafkaCommonProperties - * @Date:2024/9/29 12:26 - */ -public class KafkaCommonProperties{ - - /** - * Kafka主机 - */ - private String kafkaHost = "47.101.53.251:9092"; - - /** - * 生产者:要求leader请求结束前收到的确认次数,来控制发送数据的持久化 - * 消息确认: - * 0:生产者不等待服务器确认,此时retry参数不生效 - * 1:leader写入记录到log,不会等待follower的确认即向生产者发送通知 - * all:leader等待所有副本通知,然后向生产者发送通知,保证所有数据落盘到所有副本,功能同设置为-1 - */ - private String ack = "all"; - - /** - * 生产者重试次数 - */ - private Integer retryTimes = 1; - - /** - * 生产者:向同一分区发送打包发送的数据量,单位:bytes,默认16384bytes=16K - */ - private Integer batchSize = 16384; - - /** - * 生产者:批量发送消息的间隔时间(延迟时间),单位:毫秒 - */ - private Integer lingerMs = 1; - - /** - * 生产者:可以使用的最大缓存空间,单位:bytes,默认33554432bytes=32M. - */ - private Integer bufferMemory = 33554432; - - /** - * 生产者:键编码器 - */ - private String keyEncoder = "org.apache.kafka.common.serialization.StringSerializer"; - - /** - * 生产者:值编码器 - */ - private String valueEncoder = "org.apache.kafka.common.serialization.StringSerializer"; - - /** - * 消费者:消费topic的组ID - */ - private String groupId = "my-group-id"; - - /** - * 消费者:后台定期提交offset - */ - private String autoCommit = "true"; - - /** - * 消费者提交offset的时间间隔:单位:毫秒,当enable.auto.commit为true时生效 - */ - private String autoCommitIntervalMs = "1000"; - - /** - * 消费者:键解码器 - */ - private String keyDecoder = "org.apache.kafka.common.serialization.StringDeserializer"; - - /** - * 消费者:值解码器 - */ - private String valueDecoder = "org.apache.kafka.common.serialization.StringDeserializer"; - - /** - * 消费者:重启后配置offset - * earliest:消费者恢复到当前topic最早的offset - * latest:消费者从最新的offset开始消费 - * none:如果消费者组没找到之前的offset抛出异常 - * 其他任何值都会抛出异常 - */ - private String autoOffsetReset = "latest"; - - /** - * TOPIC - */ - private Collection topic = Collections.singleton("my-topic"); - - public KafkaCommonProperties() { - - } - - public KafkaCommonProperties(String kafkaHost, String ack, Integer retryTimes, Integer batchSize, Integer lingerMs, Integer bufferMemory, String keyEncoder, String valueEncoder, String groupId, String autoCommit, String autoCommitIntervalMs, String keyDecoder, String valueDecoder, String autoOffsetReset, Collection topic) { - this.kafkaHost = kafkaHost; - this.ack = ack; - this.retryTimes = retryTimes; - this.batchSize = batchSize; - this.lingerMs = lingerMs; - this.bufferMemory = bufferMemory; - this.keyEncoder = keyEncoder; - this.valueEncoder = valueEncoder; - this.groupId = groupId; - this.autoCommit = autoCommit; - this.autoCommitIntervalMs = autoCommitIntervalMs; - this.keyDecoder = keyDecoder; - this.valueDecoder = valueDecoder; - this.autoOffsetReset = autoOffsetReset; - this.topic = topic; - } - - public String getKafkaHost() { - return kafkaHost; - } - - public void setKafkaHost(String kafkaHost) { - this.kafkaHost = kafkaHost; - } - - public String getAck() { - return ack; - } - - public void setAck(String ack) { - this.ack = ack; - } - - public Integer getRetryTimes() { - return retryTimes; - } - - public void setRetryTimes(Integer retryTimes) { - this.retryTimes = retryTimes; - } - - public Integer getBatchSize() { - return batchSize; - } - - public void setBatchSize(Integer batchSize) { - this.batchSize = batchSize; - } - - public Integer getLingerMs() { - return lingerMs; - } - - public void setLingerMs(Integer lingerMs) { - this.lingerMs = lingerMs; - } - - public Integer getBufferMemory() { - return bufferMemory; - } - - public void setBufferMemory(Integer bufferMemory) { - this.bufferMemory = bufferMemory; - } - - public String getKeyEncoder() { - return keyEncoder; - } - - public void setKeyEncoder(String keyEncoder) { - this.keyEncoder = keyEncoder; - } - - public String getValueEncoder() { - return valueEncoder; - } - - public void setValueEncoder(String valueEncoder) { - this.valueEncoder = valueEncoder; - } - - public String getGroupId() { - return groupId; - } - - public void setGroupId(String groupId) { - this.groupId = groupId; - } - - public String getAutoCommit() { - return autoCommit; - } - - public void setAutoCommit(String autoCommit) { - this.autoCommit = autoCommit; - } - - public String getAutoCommitIntervalMs() { - return autoCommitIntervalMs; - } - - public void setAutoCommitIntervalMs(String autoCommitIntervalMs) { - this.autoCommitIntervalMs = autoCommitIntervalMs; - } - - public String getKeyDecoder() { - return keyDecoder; - } - - public void setKeyDecoder(String keyDecoder) { - this.keyDecoder = keyDecoder; - } - - public String getValueDecoder() { - return valueDecoder; - } - - public void setValueDecoder(String valueDecoder) { - this.valueDecoder = valueDecoder; - } - - public String getAutoOffsetReset() { - return autoOffsetReset; - } - - public void setAutoOffsetReset(String autoOffsetReset) { - this.autoOffsetReset = autoOffsetReset; - } - - public Collection getTopic() { - return topic; - } - - public void setTopic(Collection topic) { - this.topic = topic; - } -} diff --git a/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaConstants.java b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaConstants.java new file mode 100644 index 0000000..064911a --- /dev/null +++ b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaConstants.java @@ -0,0 +1,16 @@ +package com.muyu.common.util; + +/** + * @author liuxinyue + * @Package:com.muyu.common.util + * @name:KafkaConstants + * @Date:2024/9/29 20:22 + */ +public class KafkaConstants { + + + public final static String KafkaTopic="kafka_topic_test"; + + public final static String KafkaGrop="kafka_group_test"; + +} diff --git a/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaConsumerConfig.java b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaConsumerConfig.java new file mode 100644 index 0000000..bdd2095 --- /dev/null +++ b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaConsumerConfig.java @@ -0,0 +1,43 @@ +package com.muyu.common.util; + +import org.apache.kafka.clients.consumer.KafkaConsumer; +import org.apache.kafka.common.serialization.Deserializer; +import org.apache.kafka.common.serialization.StringDeserializer; +import org.springframework.context.annotation.Bean; +import org.springframework.context.annotation.Configuration; + +import java.util.HashMap; + +/** + * @author liuxinyue + * @Package:com.muyu.common.util + * @name:KafkaConsumerConfig + * @Date:2024/9/29 20:19 + */ +@Configuration +public class KafkaConsumerConfig +{ + + @Bean + public KafkaConsumer kafkaConsumer(){ + HashMap map = new HashMap<>(); + map.put("bootstrap.servers", "47.101.53.251:9092"); + map.put("enable.auto.commit",true); + map.put("auto.commit.interval", 5000); + map.put("auto.offset.reset", "latest"); + map.put("fetch.max.wait", 500); + map.put("fetch.min.size", 1); + map.put("heartbeat-interval", 3000); + map.put("max.poll.records", 500); + map.put("group.id", KafkaConstants.KafkaGrop); + //指定key使用的反序列化类 + Deserializer keyDeserializer = new StringDeserializer(); + //指定value使用的反序列化类 + Deserializer valueDeserializer = new StringDeserializer(); + //创建Kafka消费者 + KafkaConsumer kafkaConsumer = new KafkaConsumer(map, keyDeserializer, valueDeserializer); + return kafkaConsumer; + } + + +} diff --git a/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaProducerTest.java b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaProducerTest.java deleted file mode 100644 index 004dff9..0000000 --- a/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaProducerTest.java +++ /dev/null @@ -1,64 +0,0 @@ -package com.muyu.common.util; - -import org.apache.kafka.clients.producer.*; -import org.slf4j.Logger; -import org.slf4j.LoggerFactory; - -import java.util.Objects; -import java.util.Properties; -/** - * kafka生产 - * @author liuxinyue - * @Package:com.muyu.common.util - * @name:KafkaProducerTest - * @Date:2024/9/29 12:27 - */ -public class KafkaProducerTest { - private static final Logger logger = LoggerFactory.getLogger(KafkaProducerTest.class); - - public static KafkaProducer getDefaultKafkaProducer(KafkaCommonProperties kafkaCommonProperties) { - Properties properties = new Properties(); - properties.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, kafkaCommonProperties.getKafkaHost()); - properties.put(ProducerConfig.ACKS_CONFIG, kafkaCommonProperties.getAck()); - properties.put(ProducerConfig.RETRIES_CONFIG, kafkaCommonProperties.getRetryTimes()); - properties.put(ProducerConfig.BATCH_SIZE_CONFIG, kafkaCommonProperties.getBatchSize()); - properties.put(ProducerConfig.LINGER_MS_CONFIG, kafkaCommonProperties.getLingerMs()); - properties.put(ProducerConfig.BUFFER_MEMORY_CONFIG, kafkaCommonProperties.getBufferMemory()); - properties.put(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, kafkaCommonProperties.getKeyEncoder()); - properties.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, kafkaCommonProperties.getValueEncoder()); - return new KafkaProducer<>(properties); - } - - static class MyProducerCallback implements Callback { - - @Override - public void onCompletion(RecordMetadata metadata, Exception exception) { - if (Objects.nonNull(exception)) { - logger.error(">>>>>>>>>>Producer生产消息异常:", exception); - } - if (Objects.nonNull(metadata)) { - logger.info(">>>>>>>>>>Producer生产消息:metadata:{},partition:{}, offset:{}", metadata, metadata.partition(), metadata.offset()); - } - } - } - - public static void main(String[] args) { - KafkaCommonProperties kafkaCommonProperties = new KafkaCommonProperties(); - KafkaProducer producer = getDefaultKafkaProducer(kafkaCommonProperties); - String message = "hello world "; - try { - for (int i = 0; i < 10; i++) { - // 异步写入数据 - String topic = kafkaCommonProperties.getTopic().toArray()[0].toString(); - ProducerRecord producerRecord = new ProducerRecord<>(topic, message + i); - producer.send(producerRecord, new MyProducerCallback()); - } - } catch (Exception ex) { - logger.error(">>>>>>>>生产数据异常:", ex); - throw new RuntimeException(ex); - } finally { - producer.close(); - } - } - -} diff --git a/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaProviderConfig.java b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaProviderConfig.java new file mode 100644 index 0000000..9ca8d40 --- /dev/null +++ b/cloud-modules/saas/saas-common/src/main/java/com/muyu/common/util/KafkaProviderConfig.java @@ -0,0 +1,38 @@ +package com.muyu.common.util; + +import org.apache.kafka.clients.producer.KafkaProducer; +import org.apache.kafka.common.serialization.Serializer; +import org.apache.kafka.common.serialization.StringSerializer; +import org.springframework.context.annotation.Bean; +import org.springframework.context.annotation.Configuration; + +import java.util.HashMap; + +/** + * @author liuxinyue + * @Package:com.muyu.common.util + * @name:KafkaProviderConfig + * @Date:2024/9/29 20:15 + */ +@Configuration +public class KafkaProviderConfig{ + + @Bean + public KafkaProducer KafkaProvider(){ + + HashMap map = new HashMap<>(); + map.put("bootstrap.servers", "47.101.53.251:9092"); + map.put("retries", "2"); + map.put("batch.size", 16384); + map.put("buffer-memory", 33554432); + map.put("acks", "-1"); + Serializer keySerializer = new StringSerializer(); + //指定value使用的序列化类 + Serializer valueSerializer = new StringSerializer(); + //创建Kafka生产者 + KafkaProducer kafkaProducer = new KafkaProducer(map, keySerializer, valueSerializer); + return kafkaProducer; + } + + +} diff --git a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/SaasApplication.java b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/SaasApplication.java index ce9a6ac..2d3639a 100644 --- a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/SaasApplication.java +++ b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/SaasApplication.java @@ -10,6 +10,7 @@ import org.springframework.boot.autoconfigure.SpringBootApplication; * @Version 1.0 * @Data 2024-09-28 17:34:31 */ + @SpringBootApplication @EnableMyFeignClients public class SaasApplication { diff --git a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/config/MqttConfigure.java b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/config/MqttConfigure.java index 1b2be06..41de3c7 100644 --- a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/config/MqttConfigure.java +++ b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/config/MqttConfigure.java @@ -10,8 +10,7 @@ import org.springframework.stereotype.Component; import javax.annotation.PostConstruct; /** - * - * @author liuxinyue + * @Author:liuxinyue * @Package:com.muyu.mqtt.configure * @Project:cloud-server * @name:MqttConfigure @@ -30,7 +29,7 @@ public class MqttConfigure { String topic = "vehicle"; int qos = 2; String broker = "tcp://47.101.53.251:1883"; - String clientId = "测试mqtt"; + String clientId = "hhhhhh"; try { MqttClient sampleClient = new MqttClient(broker, clientId); MqttConnectOptions connOpts = new MqttConnectOptions(); diff --git a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/controller/TemplateController.java b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/controller/TemplateController.java index c087753..36f5980 100644 --- a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/controller/TemplateController.java +++ b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/controller/TemplateController.java @@ -6,7 +6,6 @@ import com.muyu.server.service.TemplateService; import io.swagger.v3.oas.annotations.Operation; import io.swagger.v3.oas.annotations.tags.Tag; import lombok.AllArgsConstructor; -import lombok.extern.log4j.Log4j2; import org.apache.iotdb.rpc.IoTDBConnectionException; import org.apache.iotdb.rpc.StatementExecutionException; import org.springframework.beans.factory.annotation.Autowired; @@ -20,8 +19,7 @@ import java.util.List; import java.util.concurrent.ExecutionException; /** - * 报文模版管理 - * @author liuxinyue + * @Author:liuxinyue * @Package:com.template.controller * @Project:cloud-server-c * @name:TemplateController @@ -31,7 +29,6 @@ import java.util.concurrent.ExecutionException; @RequestMapping("/template") @AllArgsConstructor @Tag(name = "报文模版管理",description = "报文模版管理") -@Log4j2 public class TemplateController { @Autowired diff --git a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/mapper/KafkaMapper.java b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/mapper/KafkaMapper.java new file mode 100644 index 0000000..9b36ff6 --- /dev/null +++ b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/mapper/KafkaMapper.java @@ -0,0 +1,13 @@ +package com.muyu.server.mapper; + +import org.apache.ibatis.annotations.Mapper; + +/** + * @author liuxinyue + * @Package:com.muyu.server.mapper + * @name:KafkaMapper + * @Date:2024/9/29 20:53 + */ +@Mapper +public interface KafkaMapper { +} diff --git a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/service/impl/MessageTemplateTypeServiceImpl.java b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/service/impl/MessageTemplateTypeServiceImpl.java index 1212ad0..6f1883a 100644 --- a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/service/impl/MessageTemplateTypeServiceImpl.java +++ b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/service/impl/MessageTemplateTypeServiceImpl.java @@ -1,5 +1,4 @@ package com.muyu.server.service.impl; - import com.baomidou.mybatisplus.core.conditions.query.QueryWrapper; import com.baomidou.mybatisplus.extension.service.impl.ServiceImpl; import com.muyu.common.domain.MessageTemplateType; @@ -7,9 +6,7 @@ import com.muyu.server.mapper.MessageTemplateTypeMapper; import com.muyu.server.service.MessageTemplateTypeService; import org.springframework.beans.factory.annotation.Autowired; import org.springframework.stereotype.Service; - import java.util.List; - /** * @Author:liuxinyue * @Package:com.sheep.message.service.impl @@ -33,7 +30,7 @@ public class MessageTemplateTypeServiceImpl extends ServiceImpl findvehicleFoundationData(Integer templatedId, String code) { QueryWrapper messageTemplateTypeQueryWrapper = new QueryWrapper<>(); - messageTemplateTypeQueryWrapper.eq("templated_id",templatedId); + messageTemplateTypeQueryWrapper.eq("template_id",templatedId); messageTemplateTypeQueryWrapper.eq("message_class",code); List messageTemplateTypes = messageTemplateTypeMapper.selectList(messageTemplateTypeQueryWrapper); return messageTemplateTypes; @@ -42,7 +39,7 @@ public class MessageTemplateTypeServiceImpl extends ServiceImpl findvehicleData(Integer templatedId, String code) { QueryWrapper messageTemplateTypeQueryWrapper = new QueryWrapper<>(); - messageTemplateTypeQueryWrapper.eq("templated_id",templatedId); + messageTemplateTypeQueryWrapper.eq("template_id",templatedId); messageTemplateTypeQueryWrapper.eq("message_class",code); List messageTemplateTypes = messageTemplateTypeMapper.selectList(messageTemplateTypeQueryWrapper); return messageTemplateTypes; @@ -51,7 +48,7 @@ public class MessageTemplateTypeServiceImpl extends ServiceImpl finddeviceStatusData(Integer templatedId, String code) { QueryWrapper messageTemplateTypeQueryWrapper = new QueryWrapper<>(); - messageTemplateTypeQueryWrapper.eq("templated_id",templatedId); + messageTemplateTypeQueryWrapper.eq("template_id",templatedId); messageTemplateTypeQueryWrapper.eq("message_class",code); List messageTemplateTypes = messageTemplateTypeMapper.selectList(messageTemplateTypeQueryWrapper); return messageTemplateTypes; @@ -60,7 +57,7 @@ public class MessageTemplateTypeServiceImpl extends ServiceImpl findMessageByTemplateName(Integer templatedId) { QueryWrapper messageTemplateTypeQueryWrapper = new QueryWrapper<>(); - messageTemplateTypeQueryWrapper.eq("templated_id",templatedId); + messageTemplateTypeQueryWrapper.eq("template_id",templatedId); List messageTemplateTypes = messageTemplateTypeMapper.selectList(messageTemplateTypeQueryWrapper); return messageTemplateTypes; } @@ -69,6 +66,7 @@ public class MessageTemplateTypeServiceImpl extends ServiceImpl findTemplateById(Integer templateId) { QueryWrapper messageTemplateTypeQueryWrapper = new QueryWrapper<>(); messageTemplateTypeQueryWrapper.eq("template_id",templateId); - return List.of(); + List messageTemplateTypes = messageTemplateTypeMapper.selectList(messageTemplateTypeQueryWrapper); + return messageTemplateTypes; } } diff --git a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/service/impl/TemplateServiceImpl.java b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/service/impl/TemplateServiceImpl.java index 34b0814..b850308 100644 --- a/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/service/impl/TemplateServiceImpl.java +++ b/cloud-modules/saas/saas-server/src/main/java/com/muyu/server/service/impl/TemplateServiceImpl.java @@ -69,38 +69,40 @@ public class TemplateServiceImpl extends ServiceImpl i log.info("车辆信息为:" + carByVin); //对应车辆所对应的报文模版 Integer templateId = carByVin.getTemplateId(); - List templateTypeList; - //key String redisKey = "messageTemplateType" + templateId; - + log.info("key为:" + redisKey); //key存在 if (redisTemplate.hasKey(redisKey)) { List list = redisTemplate.opsForList().range(redisKey, 0, -1); - templateTypeList = list.stream().map(o -> JSON.parseObject(o.toString(), MessageTemplateType.class)) .toList(); - } else { List templateTypeList1 = messageTemplateTypeService.findTemplateById(templateId); + log.info("redis存入成功"); templateTypeList = templateTypeList1; - templateTypeList.forEach( + templateTypeList1.forEach( templateType -> redisTemplate.opsForList().rightPush( redisKey, com.alibaba.fastjson.JSON.toJSONString(templateType) ) ); + + } + log.info("哈哈哈哈哈哈哈"+templateTypeList); //将模版里面有的配置进行循环 for (MessageTemplateType messageTemplateType : templateTypeList) { //开始位置 Integer startIndex = messageTemplateType.getStartIndex() - 1; //结束位置 Integer endIndex = messageTemplateType.getEndIndex(); + String substring = result.substring(startIndex, endIndex); + log.info("截取后的字符1:" + substring); //将每个解析后的字段都存入到JSON对象中 - jsonObject.put(messageTemplateType.getMessageField(), result.substring(startIndex, endIndex)); + jsonObject.put(messageTemplateType.getMessageField(),substring ); } System.out.println("哈哈哈红红火火恍恍惚惚");