es数据源
commit
bf5cf89c1f
|
@ -72,9 +72,6 @@ public class DataDatabaseServiceImpl extends BaseServiceImpl<DataDatabaseDao, Da
|
||||||
private final DataAccessDao dataAccessDao;
|
private final DataAccessDao dataAccessDao;
|
||||||
private final DataAccessService dataAccessService;
|
private final DataAccessService dataAccessService;
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public PageResult<DataDatabaseVO> page(DataDatabaseQuery query) {
|
public PageResult<DataDatabaseVO> page(DataDatabaseQuery query) {
|
||||||
IPage<DataDatabaseEntity> page = baseMapper.selectPage(getPage(query), getWrapper(query));
|
IPage<DataDatabaseEntity> page = baseMapper.selectPage(getPage(query), getWrapper(query));
|
||||||
|
@ -194,12 +191,12 @@ public class DataDatabaseServiceImpl extends BaseServiceImpl<DataDatabaseDao, Da
|
||||||
if(vo.getDatabaseType().equals(Judgment.KAFKA.getIndex())){
|
if(vo.getDatabaseType().equals(Judgment.KAFKA.getIndex())){
|
||||||
// 配置生产者属性
|
// 配置生产者属性
|
||||||
Properties properties = new Properties();
|
Properties properties = new Properties();
|
||||||
String ip = "invalid_ip:invalid_port"; // 设置一个无效的IP和端口,以模拟连接失败的情况
|
String ip=vo.getDatabaseIp()+":"+vo.getDatabasePort();
|
||||||
properties.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, ip);
|
properties.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG, ip);
|
||||||
properties.put(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, org.apache.kafka.common.serialization.StringSerializer.class.getName());
|
properties.put(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, org.apache.kafka.common.serialization.StringSerializer.class.getName());
|
||||||
properties.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, org.apache.kafka.common.serialization.StringSerializer.class.getName());
|
properties.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG, org.apache.kafka.common.serialization.StringSerializer.class.getName());
|
||||||
|
|
||||||
// 创建生产者
|
// 创建生产者方娇大傻蛋蠢猪大傻春
|
||||||
Producer<String, String> producer = null;
|
Producer<String, String> producer = null;
|
||||||
try {
|
try {
|
||||||
producer = new org.apache.kafka.clients.producer.KafkaProducer<>(properties);
|
producer = new org.apache.kafka.clients.producer.KafkaProducer<>(properties);
|
||||||
|
@ -219,6 +216,7 @@ public class DataDatabaseServiceImpl extends BaseServiceImpl<DataDatabaseDao, Da
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if(vo.getDatabaseType().equals(Judgment.ES.getIndex())){
|
if(vo.getDatabaseType().equals(Judgment.ES.getIndex())){
|
||||||
// 调用方法连接 Elasticsearch 并执行操作
|
// 调用方法连接 Elasticsearch 并执行操作
|
||||||
|
@ -229,7 +227,6 @@ public class DataDatabaseServiceImpl extends BaseServiceImpl<DataDatabaseDao, Da
|
||||||
e.printStackTrace();
|
e.printStackTrace();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
if (vo.getId() != null) {
|
if (vo.getId() != null) {
|
||||||
//更新状态
|
//更新状态
|
||||||
baseMapper.changeStatusById(vo.getId(), YesOrNo.YES.getValue());
|
baseMapper.changeStatusById(vo.getId(), YesOrNo.YES.getValue());
|
||||||
|
|
|
@ -75,7 +75,7 @@ public class DatabaseRedisImpl extends DatabaseMysqlImpl implements IDatabaseInt
|
||||||
cd.setFieldTypeName(dataType);
|
cd.setFieldTypeName(dataType);
|
||||||
int csize = columnSize != null ? Integer.parseInt(columnSize) : 0;
|
int csize = columnSize != null ? Integer.parseInt(columnSize) : 0;
|
||||||
cd.setDisplaySize(csize);
|
cd.setDisplaySize(csize);
|
||||||
cd.setPrecisionSize(csize);
|
cd .setPrecisionSize(csize);
|
||||||
cd.setScaleSize(decimalDigits != null ? Integer.parseInt(decimalDigits) : 0);
|
cd.setScaleSize(decimalDigits != null ? Integer.parseInt(decimalDigits) : 0);
|
||||||
cd.setRemarks(columnComment);
|
cd.setRemarks(columnComment);
|
||||||
break;
|
break;
|
||||||
|
|
Loading…
Reference in New Issue