修改实体类规范

master
Number7 2024-09-29 15:11:57 +08:00
parent fd529f0cec
commit dfacea67f6
9 changed files with 332 additions and 0 deletions

View File

@ -6,6 +6,14 @@ import lombok.AllArgsConstructor;
import lombok.Data;
import lombok.NoArgsConstructor;
/**
*
* @Authorliuxinyue
* @Packagecom.sheep.message.domain
* @Projectcloud-server-c
* @nameMessageTemplateType
* @Date2024/9/18 21:01
*/
@Data
@AllArgsConstructor
@NoArgsConstructor

View File

@ -10,6 +10,7 @@ import lombok.NoArgsConstructor;
import lombok.experimental.SuperBuilder;
import java.io.Serializable;
/**
*
* @Authorliuxinyue
* @Packagecom.sheep.message.domain
* @Projectcloud-server-c

View File

@ -9,6 +9,7 @@ import lombok.experimental.SuperBuilder;
import java.sql.Date;
/**
*
* @Authorliuxinyue
* @Packagecom.template.domain
* @Projectcloud-server

View File

@ -11,6 +11,7 @@ import lombok.experimental.SuperBuilder;
import java.io.Serializable;
/**
*
* @Authorliuxinyue
* @Packagecom.sheep.message.domain
* @Projectcloud-server-c
@ -25,6 +26,7 @@ import java.io.Serializable;
@TableName(value = "message_template_type",autoResultMap = true)
public class MessageTemplateType implements Serializable {
/**
*
*/

View File

@ -9,6 +9,14 @@ import lombok.Data;
import lombok.EqualsAndHashCode;
import lombok.NoArgsConstructor;
/**
*
* @Authorliuxinyue
* @Packagecom.sheep.message.domain
* @Projectcloud-server-c
* @nameMessageTemplateType
* @Date2024/9/18 21:01
*/
@Data
@AllArgsConstructor
@NoArgsConstructor

View File

@ -10,6 +10,14 @@ import lombok.NoArgsConstructor;
import java.util.Date;
/**
*
* @Authorliuxinyue
* @Packagecom.sheep.message.domain
* @Projectcloud-server-c
* @nameMessageTemplateType
* @Date2024/9/18 21:01
*/
@Data
@AllArgsConstructor
@NoArgsConstructor

View File

@ -12,6 +12,7 @@ import lombok.experimental.SuperBuilder;
import java.util.Date;
/**
*
* @Authorliuxinyue
* @Packagecom.template.domain
* @Projectcloud-server-c

View File

@ -0,0 +1,239 @@
package com.muyu.common.util;
import java.util.Collection;
import java.util.Collections;
/**
* kafka
* @author liuxinyue
* @Packagecom.muyu.common.util
* @nameKafkaCommonProperties
* @Date2024/9/29 12:26
*/
public class KafkaCommonProperties{
/**
* Kafka
*/
private String kafkaHost = "47.101.53.251:9092";
/**
* leader
*
* 0retry
* 1leaderlogfollower
* allleader-1
*/
private String ack = "all";
/**
*
*/
private Integer retryTimes = 1;
/**
* bytes16384bytes=16K
*/
private Integer batchSize = 16384;
/**
*
*/
private Integer lingerMs = 1;
/**
* 使bytes33554432bytes=32M.
*/
private Integer bufferMemory = 33554432;
/**
*
*/
private String keyEncoder = "org.apache.kafka.common.serialization.StringSerializer";
/**
*
*/
private String valueEncoder = "org.apache.kafka.common.serialization.StringSerializer";
/**
* topicID
*/
private String groupId = "my-group-id";
/**
* offset
*/
private String autoCommit = "true";
/**
* offsetenable.auto.committrue
*/
private String autoCommitIntervalMs = "1000";
/**
*
*/
private String keyDecoder = "org.apache.kafka.common.serialization.StringDeserializer";
/**
*
*/
private String valueDecoder = "org.apache.kafka.common.serialization.StringDeserializer";
/**
* offset
* earliesttopicoffset
* latestoffset
* noneoffset
*
*/
private String autoOffsetReset = "latest";
/**
* TOPIC
*/
private Collection<String> topic = Collections.singleton("my-topic");
public KafkaCommonProperties() {
}
public KafkaCommonProperties(String kafkaHost, String ack, Integer retryTimes, Integer batchSize, Integer lingerMs, Integer bufferMemory, String keyEncoder, String valueEncoder, String groupId, String autoCommit, String autoCommitIntervalMs, String keyDecoder, String valueDecoder, String autoOffsetReset, Collection<String> topic) {
this.kafkaHost = kafkaHost;
this.ack = ack;
this.retryTimes = retryTimes;
this.batchSize = batchSize;
this.lingerMs = lingerMs;
this.bufferMemory = bufferMemory;
this.keyEncoder = keyEncoder;
this.valueEncoder = valueEncoder;
this.groupId = groupId;
this.autoCommit = autoCommit;
this.autoCommitIntervalMs = autoCommitIntervalMs;
this.keyDecoder = keyDecoder;
this.valueDecoder = valueDecoder;
this.autoOffsetReset = autoOffsetReset;
this.topic = topic;
}
public String getKafkaHost() {
return kafkaHost;
}
public void setKafkaHost(String kafkaHost) {
this.kafkaHost = kafkaHost;
}
public String getAck() {
return ack;
}
public void setAck(String ack) {
this.ack = ack;
}
public Integer getRetryTimes() {
return retryTimes;
}
public void setRetryTimes(Integer retryTimes) {
this.retryTimes = retryTimes;
}
public Integer getBatchSize() {
return batchSize;
}
public void setBatchSize(Integer batchSize) {
this.batchSize = batchSize;
}
public Integer getLingerMs() {
return lingerMs;
}
public void setLingerMs(Integer lingerMs) {
this.lingerMs = lingerMs;
}
public Integer getBufferMemory() {
return bufferMemory;
}
public void setBufferMemory(Integer bufferMemory) {
this.bufferMemory = bufferMemory;
}
public String getKeyEncoder() {
return keyEncoder;
}
public void setKeyEncoder(String keyEncoder) {
this.keyEncoder = keyEncoder;
}
public String getValueEncoder() {
return valueEncoder;
}
public void setValueEncoder(String valueEncoder) {
this.valueEncoder = valueEncoder;
}
public String getGroupId() {
return groupId;
}
public void setGroupId(String groupId) {
this.groupId = groupId;
}
public String getAutoCommit() {
return autoCommit;
}
public void setAutoCommit(String autoCommit) {
this.autoCommit = autoCommit;
}
public String getAutoCommitIntervalMs() {
return autoCommitIntervalMs;
}
public void setAutoCommitIntervalMs(String autoCommitIntervalMs) {
this.autoCommitIntervalMs = autoCommitIntervalMs;
}
public String getKeyDecoder() {
return keyDecoder;
}
public void setKeyDecoder(String keyDecoder) {
this.keyDecoder = keyDecoder;
}
public String getValueDecoder() {
return valueDecoder;
}
public void setValueDecoder(String valueDecoder) {
this.valueDecoder = valueDecoder;
}
public String getAutoOffsetReset() {
return autoOffsetReset;
}
public void setAutoOffsetReset(String autoOffsetReset) {
this.autoOffsetReset = autoOffsetReset;
}
public Collection<String> getTopic() {
return topic;
}
public void setTopic(Collection<String> topic) {
this.topic = topic;
}
}

View File

@ -0,0 +1,64 @@
package com.muyu.common.util;
import org.apache.kafka.clients.producer.*;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.util.Objects;
import java.util.Properties;
/**
* kafka
* @author liuxinyue
* @Packagecom.muyu.common.util
* @nameKafkaProducerTest
* @Date2024/9/29 12:27
*/
public class KafkaProducerTest {
private static final Logger logger = LoggerFactory.getLogger(KafkaProducerTest.class);
public static KafkaProducer<String, String> getDefaultKafkaProducer(KafkaCommonProperties kafkaCommonProperties) {
Properties properties = new Properties();
properties.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, kafkaCommonProperties.getKafkaHost());
properties.put(ProducerConfig.ACKS_CONFIG, kafkaCommonProperties.getAck());
properties.put(ProducerConfig.RETRIES_CONFIG, kafkaCommonProperties.getRetryTimes());
properties.put(ProducerConfig.BATCH_SIZE_CONFIG, kafkaCommonProperties.getBatchSize());
properties.put(ProducerConfig.LINGER_MS_CONFIG, kafkaCommonProperties.getLingerMs());
properties.put(ProducerConfig.BUFFER_MEMORY_CONFIG, kafkaCommonProperties.getBufferMemory());
properties.put(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, kafkaCommonProperties.getKeyEncoder());
properties.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, kafkaCommonProperties.getValueEncoder());
return new KafkaProducer<>(properties);
}
static class MyProducerCallback implements Callback {
@Override
public void onCompletion(RecordMetadata metadata, Exception exception) {
if (Objects.nonNull(exception)) {
logger.error(">>>>>>>>>>Producer生产消息异常", exception);
}
if (Objects.nonNull(metadata)) {
logger.info(">>>>>>>>>>Producer生产消息metadata{}partition:{}, offset{}", metadata, metadata.partition(), metadata.offset());
}
}
}
public static void main(String[] args) {
KafkaCommonProperties kafkaCommonProperties = new KafkaCommonProperties();
KafkaProducer<String, String> producer = getDefaultKafkaProducer(kafkaCommonProperties);
String message = "hello world ";
try {
for (int i = 0; i < 10; i++) {
// 异步写入数据
String topic = kafkaCommonProperties.getTopic().toArray()[0].toString();
ProducerRecord<String, String> producerRecord = new ProducerRecord<>(topic, message + i);
producer.send(producerRecord, new MyProducerCallback());
}
} catch (Exception ex) {
logger.error(">>>>>>>>生产数据异常:", ex);
throw new RuntimeException(ex);
} finally {
producer.close();
}
}
}